Llama-3-Taiwan-8B is a 8B parameter model finetuned on a large corpus of Traditional Mandarin and English data using the Llama-3 architecture. It demonstrates state-of-the-art performance on various Traditional Mandarin NLP benchmarks.

8B

382 Pulls Updated 2 months ago

c0aac7c7f00d · 128B
{ "stop": [ "<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>", "<|reserved_special_token" ] }