Llama-3-Taiwan-8B is a 8B parameter model finetuned on a large corpus of Traditional Mandarin and English data using the Llama-3 architecture. It demonstrates state-of-the-art performance on various Traditional Mandarin NLP benchmarks.
8B
382 Pulls Updated 2 months ago
4b3c4ede0a3f · 6.6GB
-
general.architecturellama
-
general.file_typeQ6_K
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length8192
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base500000
-
llama.vocab_size128258
-
tokenizer.ggml.bos_token_id128000
-
tokenizer.ggml.eos_token_id128009
-
tokenizer.ggml.merges[Ġ Ġ Ġ ĠĠĠ ĠĠ ĠĠ ĠĠĠ Ġ i n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.padding_token_id128001
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightQ6_K[4096 128258]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.ffn_gate.weightQ6_K[4096 14336]
-
blk.0.ffn_up.weightQ6_K[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightQ6_K[4096 1024]
-
blk.0.attn_output.weightQ6_K[4096 4096]
-
blk.0.attn_q.weightQ6_K[4096 4096]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.ffn_gate.weightQ6_K[4096 14336]
-
blk.1.ffn_up.weightQ6_K[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightQ6_K[4096 1024]
-
blk.1.attn_output.weightQ6_K[4096 4096]
-
blk.1.attn_q.weightQ6_K[4096 4096]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.ffn_gate.weightQ6_K[4096 14336]
-
blk.2.ffn_up.weightQ6_K[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightQ6_K[4096 1024]
-
blk.2.attn_output.weightQ6_K[4096 4096]
-
blk.2.attn_q.weightQ6_K[4096 4096]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.ffn_gate.weightQ6_K[4096 14336]
-
blk.3.ffn_up.weightQ6_K[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightQ6_K[4096 1024]
-
blk.3.attn_output.weightQ6_K[4096 4096]
-
blk.3.attn_q.weightQ6_K[4096 4096]
-
blk.3.attn_v.weightQ6_K[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightQ6_K[14336 4096]
-
blk.4.ffn_gate.weightQ6_K[4096 14336]
-
blk.4.ffn_up.weightQ6_K[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightQ6_K[4096 1024]
-
blk.4.attn_output.weightQ6_K[4096 4096]
-
blk.4.attn_q.weightQ6_K[4096 4096]
-
blk.4.attn_v.weightQ6_K[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightQ6_K[14336 4096]
-
blk.5.ffn_gate.weightQ6_K[4096 14336]
-
blk.5.ffn_up.weightQ6_K[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightQ6_K[4096 1024]
-
blk.5.attn_output.weightQ6_K[4096 4096]
-
blk.5.attn_q.weightQ6_K[4096 4096]
-
blk.5.attn_v.weightQ6_K[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.ffn_gate.weightQ6_K[4096 14336]
-
blk.6.ffn_up.weightQ6_K[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightQ6_K[4096 1024]
-
blk.6.attn_output.weightQ6_K[4096 4096]
-
blk.6.attn_q.weightQ6_K[4096 4096]
-
blk.6.attn_v.weightQ6_K[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightQ6_K[14336 4096]
-
blk.7.ffn_gate.weightQ6_K[4096 14336]
-
blk.7.ffn_up.weightQ6_K[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightQ6_K[4096 1024]
-
blk.7.attn_output.weightQ6_K[4096 4096]
-
blk.7.attn_q.weightQ6_K[4096 4096]
-
blk.7.attn_v.weightQ6_K[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightQ6_K[14336 4096]
-
blk.8.ffn_gate.weightQ6_K[4096 14336]
-
blk.8.ffn_up.weightQ6_K[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightQ6_K[4096 1024]
-
blk.8.attn_output.weightQ6_K[4096 4096]
-
blk.8.attn_q.weightQ6_K[4096 4096]
-
blk.8.attn_v.weightQ6_K[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightQ6_K[14336 4096]
-
blk.9.ffn_gate.weightQ6_K[4096 14336]
-
blk.9.ffn_up.weightQ6_K[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightQ6_K[4096 1024]
-
blk.9.attn_output.weightQ6_K[4096 4096]
-
blk.9.attn_q.weightQ6_K[4096 4096]
-
blk.9.attn_v.weightQ6_K[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightQ6_K[14336 4096]
-
blk.10.ffn_gate.weightQ6_K[4096 14336]
-
blk.10.ffn_up.weightQ6_K[4096 14336]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightQ6_K[4096 1024]
-
blk.10.attn_output.weightQ6_K[4096 4096]
-
blk.10.attn_q.weightQ6_K[4096 4096]
-
blk.10.attn_v.weightQ6_K[4096 1024]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightQ6_K[14336 4096]
-
blk.11.ffn_gate.weightQ6_K[4096 14336]
-
blk.11.ffn_up.weightQ6_K[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightQ6_K[4096 1024]
-
blk.11.attn_output.weightQ6_K[4096 4096]
-
blk.11.attn_q.weightQ6_K[4096 4096]
-
blk.11.attn_v.weightQ6_K[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightQ6_K[14336 4096]
-
blk.12.ffn_gate.weightQ6_K[4096 14336]
-
blk.12.ffn_up.weightQ6_K[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightQ6_K[4096 1024]
-
blk.12.attn_output.weightQ6_K[4096 4096]
-
blk.12.attn_q.weightQ6_K[4096 4096]
-
blk.12.attn_v.weightQ6_K[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightQ6_K[14336 4096]
-
blk.13.ffn_gate.weightQ6_K[4096 14336]
-
blk.13.ffn_up.weightQ6_K[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightQ6_K[4096 1024]
-
blk.13.attn_output.weightQ6_K[4096 4096]
-
blk.13.attn_q.weightQ6_K[4096 4096]
-
blk.13.attn_v.weightQ6_K[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightQ6_K[14336 4096]
-
blk.14.ffn_gate.weightQ6_K[4096 14336]
-
blk.14.ffn_up.weightQ6_K[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightQ6_K[4096 1024]
-
blk.14.attn_output.weightQ6_K[4096 4096]
-
blk.14.attn_q.weightQ6_K[4096 4096]
-
blk.14.attn_v.weightQ6_K[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightQ6_K[14336 4096]
-
blk.15.ffn_gate.weightQ6_K[4096 14336]
-
blk.15.ffn_up.weightQ6_K[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightQ6_K[4096 1024]
-
blk.15.attn_output.weightQ6_K[4096 4096]
-
blk.15.attn_q.weightQ6_K[4096 4096]
-
blk.15.attn_v.weightQ6_K[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightQ6_K[14336 4096]
-
blk.16.ffn_gate.weightQ6_K[4096 14336]
-
blk.16.ffn_up.weightQ6_K[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightQ6_K[4096 1024]
-
blk.16.attn_output.weightQ6_K[4096 4096]
-
blk.16.attn_q.weightQ6_K[4096 4096]
-
blk.16.attn_v.weightQ6_K[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightQ6_K[14336 4096]
-
blk.17.ffn_gate.weightQ6_K[4096 14336]
-
blk.17.ffn_up.weightQ6_K[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightQ6_K[4096 1024]
-
blk.17.attn_output.weightQ6_K[4096 4096]
-
blk.17.attn_q.weightQ6_K[4096 4096]
-
blk.17.attn_v.weightQ6_K[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightQ6_K[14336 4096]
-
blk.18.ffn_gate.weightQ6_K[4096 14336]
-
blk.18.ffn_up.weightQ6_K[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightQ6_K[4096 1024]
-
blk.18.attn_output.weightQ6_K[4096 4096]
-
blk.18.attn_q.weightQ6_K[4096 4096]
-
blk.18.attn_v.weightQ6_K[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightQ6_K[14336 4096]
-
blk.19.ffn_gate.weightQ6_K[4096 14336]
-
blk.19.ffn_up.weightQ6_K[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightQ6_K[4096 1024]
-
blk.19.attn_output.weightQ6_K[4096 4096]
-
blk.19.attn_q.weightQ6_K[4096 4096]
-
blk.19.attn_v.weightQ6_K[4096 1024]
-
blk.20.ffn_gate.weightQ6_K[4096 14336]
-
blk.20.attn_k.weightQ6_K[4096 1024]
-
blk.20.attn_output.weightQ6_K[4096 4096]
-
blk.20.attn_q.weightQ6_K[4096 4096]
-
blk.20.attn_v.weightQ6_K[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightQ6_K[14336 4096]
-
blk.20.ffn_up.weightQ6_K[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.ffn_gate.weightQ6_K[4096 14336]
-
blk.21.ffn_up.weightQ6_K[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightQ6_K[4096 1024]
-
blk.21.attn_output.weightQ6_K[4096 4096]
-
blk.21.attn_q.weightQ6_K[4096 4096]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightQ6_K[14336 4096]
-
blk.22.ffn_gate.weightQ6_K[4096 14336]
-
blk.22.ffn_up.weightQ6_K[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightQ6_K[4096 1024]
-
blk.22.attn_output.weightQ6_K[4096 4096]
-
blk.22.attn_q.weightQ6_K[4096 4096]
-
blk.22.attn_v.weightQ6_K[4096 1024]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightQ6_K[14336 4096]
-
blk.23.ffn_gate.weightQ6_K[4096 14336]
-
blk.23.ffn_up.weightQ6_K[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightQ6_K[4096 1024]
-
blk.23.attn_output.weightQ6_K[4096 4096]
-
blk.23.attn_q.weightQ6_K[4096 4096]
-
blk.23.attn_v.weightQ6_K[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.ffn_gate.weightQ6_K[4096 14336]
-
blk.24.ffn_up.weightQ6_K[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightQ6_K[4096 1024]
-
blk.24.attn_output.weightQ6_K[4096 4096]
-
blk.24.attn_q.weightQ6_K[4096 4096]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightQ6_K[14336 4096]
-
blk.25.ffn_gate.weightQ6_K[4096 14336]
-
blk.25.ffn_up.weightQ6_K[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightQ6_K[4096 1024]
-
blk.25.attn_output.weightQ6_K[4096 4096]
-
blk.25.attn_q.weightQ6_K[4096 4096]
-
blk.25.attn_v.weightQ6_K[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightQ6_K[14336 4096]
-
blk.26.ffn_gate.weightQ6_K[4096 14336]
-
blk.26.ffn_up.weightQ6_K[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightQ6_K[4096 1024]
-
blk.26.attn_output.weightQ6_K[4096 4096]
-
blk.26.attn_q.weightQ6_K[4096 4096]
-
blk.26.attn_v.weightQ6_K[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.ffn_gate.weightQ6_K[4096 14336]
-
blk.27.ffn_up.weightQ6_K[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightQ6_K[4096 1024]
-
blk.27.attn_output.weightQ6_K[4096 4096]
-
blk.27.attn_q.weightQ6_K[4096 4096]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.ffn_gate.weightQ6_K[4096 14336]
-
blk.28.ffn_up.weightQ6_K[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightQ6_K[4096 1024]
-
blk.28.attn_output.weightQ6_K[4096 4096]
-
blk.28.attn_q.weightQ6_K[4096 4096]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.ffn_gate.weightQ6_K[4096 14336]
-
blk.29.ffn_up.weightQ6_K[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightQ6_K[4096 1024]
-
blk.29.attn_output.weightQ6_K[4096 4096]
-
blk.29.attn_q.weightQ6_K[4096 4096]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.ffn_gate.weightQ6_K[4096 14336]
-
blk.30.ffn_up.weightQ6_K[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightQ6_K[4096 1024]
-
blk.30.attn_output.weightQ6_K[4096 4096]
-
blk.30.attn_q.weightQ6_K[4096 4096]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.31.ffn_gate.weightQ6_K[4096 14336]
-
blk.31.ffn_up.weightQ6_K[4096 14336]
-
blk.31.attn_k.weightQ6_K[4096 1024]
-
blk.31.attn_output.weightQ6_K[4096 4096]
-
blk.31.attn_q.weightQ6_K[4096 4096]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output.weightQ6_K[4096 128258]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31