latest
15GB
1 Pull Updated 3 months ago
b92fa4610462 · 15GB
-
general.architectureqwen
-
general.file_typeF16
-
qwen.attention.head_count32
-
qwen.attention.layer_norm_rms_epsilon1e-06
-
qwen.block_count32
-
qwen.context_length32768
-
qwen.embedding_length4096
-
qwen.feed_forward_length22016
-
qwen.rope.dimension_count128
-
qwen.rope.freq_base10000
-
tokenizer.ggml.bos_token_id151643
-
tokenizer.ggml.eos_token_id151643
-
tokenizer.ggml.merges[Ġ Ġ ĠĠ ĠĠ i n Ġ t ĠĠĠĠ ĠĠĠĠ ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.preqwen2
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
tokenizer.ggml.unknown_token_id151643
-
NameTypeShape
-
token_embd.weightF16[4096 151936]
-
blk.0.attn_qkv.biasF32[12288]
-
blk.0.attn_qkv.weightF16[4096 12288]
-
blk.0.attn_output.weightF16[4096 4096]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.ffn_down.weightF16[11008 4096]
-
blk.0.ffn_up.weightF16[4096 11008]
-
blk.0.ffn_gate.weightF16[4096 11008]
-
blk.1.attn_qkv.biasF32[12288]
-
blk.1.attn_qkv.weightF16[4096 12288]
-
blk.1.attn_output.weightF16[4096 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.ffn_down.weightF16[11008 4096]
-
blk.1.ffn_up.weightF16[4096 11008]
-
blk.1.ffn_gate.weightF16[4096 11008]
-
blk.2.attn_qkv.biasF32[12288]
-
blk.2.attn_qkv.weightF16[4096 12288]
-
blk.2.attn_output.weightF16[4096 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.ffn_down.weightF16[11008 4096]
-
blk.2.ffn_up.weightF16[4096 11008]
-
blk.2.ffn_gate.weightF16[4096 11008]
-
blk.3.attn_qkv.biasF32[12288]
-
blk.3.attn_qkv.weightF16[4096 12288]
-
blk.3.attn_output.weightF16[4096 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.ffn_down.weightF16[11008 4096]
-
blk.3.ffn_up.weightF16[4096 11008]
-
blk.3.ffn_gate.weightF16[4096 11008]
-
blk.4.attn_qkv.biasF32[12288]
-
blk.4.attn_qkv.weightF16[4096 12288]
-
blk.4.attn_output.weightF16[4096 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.ffn_down.weightF16[11008 4096]
-
blk.4.ffn_up.weightF16[4096 11008]
-
blk.4.ffn_gate.weightF16[4096 11008]
-
blk.5.attn_qkv.biasF32[12288]
-
blk.5.attn_qkv.weightF16[4096 12288]
-
blk.5.attn_output.weightF16[4096 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.ffn_down.weightF16[11008 4096]
-
blk.5.ffn_up.weightF16[4096 11008]
-
blk.5.ffn_gate.weightF16[4096 11008]
-
blk.6.attn_qkv.biasF32[12288]
-
blk.6.attn_qkv.weightF16[4096 12288]
-
blk.6.attn_output.weightF16[4096 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.ffn_down.weightF16[11008 4096]
-
blk.6.ffn_up.weightF16[4096 11008]
-
blk.6.ffn_gate.weightF16[4096 11008]
-
blk.7.attn_qkv.biasF32[12288]
-
blk.7.attn_qkv.weightF16[4096 12288]
-
blk.7.attn_output.weightF16[4096 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.ffn_down.weightF16[11008 4096]
-
blk.7.ffn_up.weightF16[4096 11008]
-
blk.7.ffn_gate.weightF16[4096 11008]
-
blk.8.attn_qkv.biasF32[12288]
-
blk.8.attn_qkv.weightF16[4096 12288]
-
blk.8.attn_output.weightF16[4096 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.ffn_down.weightF16[11008 4096]
-
blk.8.ffn_up.weightF16[4096 11008]
-
blk.8.ffn_gate.weightF16[4096 11008]
-
blk.9.attn_qkv.biasF32[12288]
-
blk.9.attn_qkv.weightF16[4096 12288]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.attn_output.weightF16[4096 4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.ffn_down.weightF16[11008 4096]
-
blk.9.ffn_up.weightF16[4096 11008]
-
blk.9.ffn_gate.weightF16[4096 11008]
-
blk.10.attn_qkv.biasF32[12288]
-
blk.10.attn_qkv.weightF16[4096 12288]
-
blk.10.attn_output.weightF16[4096 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.ffn_down.weightF16[11008 4096]
-
blk.10.ffn_up.weightF16[4096 11008]
-
blk.10.ffn_gate.weightF16[4096 11008]
-
blk.11.attn_qkv.biasF32[12288]
-
blk.11.attn_qkv.weightF16[4096 12288]
-
blk.11.attn_output.weightF16[4096 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.ffn_down.weightF16[11008 4096]
-
blk.11.ffn_up.weightF16[4096 11008]
-
blk.11.ffn_gate.weightF16[4096 11008]
-
blk.12.attn_qkv.biasF32[12288]
-
blk.12.attn_qkv.weightF16[4096 12288]
-
blk.12.attn_output.weightF16[4096 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.ffn_down.weightF16[11008 4096]
-
blk.12.ffn_up.weightF16[4096 11008]
-
blk.12.ffn_gate.weightF16[4096 11008]
-
blk.13.attn_qkv.biasF32[12288]
-
blk.13.attn_qkv.weightF16[4096 12288]
-
blk.13.attn_output.weightF16[4096 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.ffn_down.weightF16[11008 4096]
-
blk.13.ffn_up.weightF16[4096 11008]
-
blk.13.ffn_gate.weightF16[4096 11008]
-
blk.14.attn_qkv.biasF32[12288]
-
blk.14.attn_qkv.weightF16[4096 12288]
-
blk.14.attn_output.weightF16[4096 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.ffn_down.weightF16[11008 4096]
-
blk.14.ffn_up.weightF16[4096 11008]
-
blk.14.ffn_gate.weightF16[4096 11008]
-
blk.15.attn_qkv.biasF32[12288]
-
blk.15.attn_qkv.weightF16[4096 12288]
-
blk.15.attn_output.weightF16[4096 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.ffn_down.weightF16[11008 4096]
-
blk.15.ffn_up.weightF16[4096 11008]
-
blk.15.ffn_gate.weightF16[4096 11008]
-
blk.16.attn_qkv.biasF32[12288]
-
blk.16.attn_qkv.weightF16[4096 12288]
-
blk.16.attn_output.weightF16[4096 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.ffn_down.weightF16[11008 4096]
-
blk.16.ffn_up.weightF16[4096 11008]
-
blk.16.ffn_gate.weightF16[4096 11008]
-
blk.17.attn_qkv.biasF32[12288]
-
blk.17.attn_qkv.weightF16[4096 12288]
-
blk.17.attn_output.weightF16[4096 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.ffn_down.weightF16[11008 4096]
-
blk.17.ffn_up.weightF16[4096 11008]
-
blk.17.ffn_gate.weightF16[4096 11008]
-
blk.18.attn_qkv.biasF32[12288]
-
blk.18.attn_qkv.weightF16[4096 12288]
-
blk.18.attn_output.weightF16[4096 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.ffn_down.weightF16[11008 4096]
-
blk.18.ffn_up.weightF16[4096 11008]
-
blk.18.ffn_gate.weightF16[4096 11008]
-
blk.19.attn_qkv.biasF32[12288]
-
blk.19.attn_qkv.weightF16[4096 12288]
-
blk.19.attn_output.weightF16[4096 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.ffn_down.weightF16[11008 4096]
-
blk.19.ffn_up.weightF16[4096 11008]
-
blk.19.ffn_gate.weightF16[4096 11008]
-
blk.20.attn_qkv.biasF32[12288]
-
blk.20.attn_qkv.weightF16[4096 12288]
-
blk.20.attn_output.weightF16[4096 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.ffn_down.weightF16[11008 4096]
-
blk.20.ffn_up.weightF16[4096 11008]
-
blk.20.ffn_gate.weightF16[4096 11008]
-
blk.21.attn_qkv.biasF32[12288]
-
blk.21.attn_qkv.weightF16[4096 12288]
-
blk.21.attn_output.weightF16[4096 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.ffn_up.weightF16[4096 11008]
-
blk.21.ffn_down.weightF16[11008 4096]
-
blk.21.ffn_gate.weightF16[4096 11008]
-
blk.22.attn_qkv.biasF32[12288]
-
blk.22.attn_qkv.weightF16[4096 12288]
-
blk.22.attn_output.weightF16[4096 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.ffn_down.weightF16[11008 4096]
-
blk.22.ffn_up.weightF16[4096 11008]
-
blk.22.ffn_gate.weightF16[4096 11008]
-
blk.23.attn_qkv.biasF32[12288]
-
blk.23.attn_qkv.weightF16[4096 12288]
-
blk.23.attn_output.weightF16[4096 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.ffn_down.weightF16[11008 4096]
-
blk.23.ffn_up.weightF16[4096 11008]
-
blk.23.ffn_gate.weightF16[4096 11008]
-
blk.24.attn_qkv.biasF32[12288]
-
blk.24.attn_qkv.weightF16[4096 12288]
-
blk.24.attn_output.weightF16[4096 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.ffn_down.weightF16[11008 4096]
-
blk.24.ffn_up.weightF16[4096 11008]
-
blk.24.ffn_gate.weightF16[4096 11008]
-
blk.25.attn_qkv.biasF32[12288]
-
blk.25.attn_qkv.weightF16[4096 12288]
-
blk.25.attn_output.weightF16[4096 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.ffn_down.weightF16[11008 4096]
-
blk.25.ffn_up.weightF16[4096 11008]
-
blk.25.ffn_gate.weightF16[4096 11008]
-
blk.26.attn_qkv.biasF32[12288]
-
blk.26.attn_qkv.weightF16[4096 12288]
-
blk.26.attn_output.weightF16[4096 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.ffn_down.weightF16[11008 4096]
-
blk.26.ffn_up.weightF16[4096 11008]
-
blk.26.ffn_gate.weightF16[4096 11008]
-
blk.27.attn_qkv.biasF32[12288]
-
blk.27.attn_qkv.weightF16[4096 12288]
-
blk.27.attn_output.weightF16[4096 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.ffn_down.weightF16[11008 4096]
-
blk.27.ffn_up.weightF16[4096 11008]
-
blk.27.ffn_gate.weightF16[4096 11008]
-
blk.28.attn_qkv.biasF32[12288]
-
blk.28.attn_qkv.weightF16[4096 12288]
-
blk.28.attn_output.weightF16[4096 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.ffn_down.weightF16[11008 4096]
-
blk.28.ffn_up.weightF16[4096 11008]
-
blk.28.ffn_gate.weightF16[4096 11008]
-
blk.29.attn_qkv.biasF32[12288]
-
blk.29.attn_qkv.weightF16[4096 12288]
-
blk.29.attn_output.weightF16[4096 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.ffn_down.weightF16[11008 4096]
-
blk.29.ffn_up.weightF16[4096 11008]
-
blk.29.ffn_gate.weightF16[4096 11008]
-
blk.30.attn_qkv.biasF32[12288]
-
blk.30.attn_qkv.weightF16[4096 12288]
-
blk.30.attn_output.weightF16[4096 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.ffn_down.weightF16[11008 4096]
-
blk.30.ffn_up.weightF16[4096 11008]
-
blk.30.ffn_gate.weightF16[4096 11008]
-
blk.31.attn_qkv.biasF32[12288]
-
blk.31.attn_qkv.weightF16[4096 12288]
-
blk.31.attn_output.weightF16[4096 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.ffn_down.weightF16[11008 4096]
-
blk.31.ffn_up.weightF16[4096 11008]
-
blk.31.ffn_gate.weightF16[4096 11008]
-
output_norm.weightF32[4096]
-
output.weightF16[4096 151936]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31