State of the art large language model from Microsoft AI with improved performance on complex chat, multilingual, reasoning and agent use cases. More quantizations.
7B
46 Pulls Updated 5 months ago
0fd7238048a8 · 2.0GB
-
general.architecturellama
-
general.file_typeIQ2_XXS
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length32768
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
llama.vocab_size32000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id2
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
NameTypeShape
-
token_embd.weightQ2_K[4096 32000]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ2_K[14336 4096]
-
blk.0.ffn_gate.weightI8[4096 14336]
-
blk.0.ffn_up.weightI8[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightI8[4096 1024]
-
blk.0.attn_output.weightI8[4096 4096]
-
blk.0.attn_q.weightI8[4096 4096]
-
blk.0.attn_v.weightQ4_K[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ2_K[14336 4096]
-
blk.1.ffn_gate.weightI8[4096 14336]
-
blk.1.ffn_up.weightI8[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightI8[4096 1024]
-
blk.1.attn_output.weightI8[4096 4096]
-
blk.1.attn_q.weightI8[4096 4096]
-
blk.1.attn_v.weightQ4_K[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ2_K[14336 4096]
-
blk.2.ffn_gate.weightI8[4096 14336]
-
blk.2.ffn_up.weightI8[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightI8[4096 1024]
-
blk.2.attn_output.weightI8[4096 4096]
-
blk.2.attn_q.weightI8[4096 4096]
-
blk.2.attn_v.weightQ4_K[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ2_K[14336 4096]
-
blk.3.ffn_gate.weightI8[4096 14336]
-
blk.3.ffn_up.weightI8[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightI8[4096 1024]
-
blk.3.attn_output.weightI8[4096 4096]
-
blk.3.attn_q.weightI8[4096 4096]
-
blk.3.attn_v.weightQ4_K[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightI8[14336 4096]
-
blk.4.ffn_gate.weightI8[4096 14336]
-
blk.4.ffn_up.weightI8[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightI8[4096 1024]
-
blk.4.attn_output.weightI8[4096 4096]
-
blk.4.attn_q.weightI8[4096 4096]
-
blk.4.attn_v.weightQ4_K[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightI8[14336 4096]
-
blk.5.ffn_gate.weightI8[4096 14336]
-
blk.5.ffn_up.weightI8[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightI8[4096 1024]
-
blk.5.attn_output.weightI8[4096 4096]
-
blk.5.attn_q.weightI8[4096 4096]
-
blk.5.attn_v.weightQ4_K[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightI8[14336 4096]
-
blk.6.ffn_gate.weightI8[4096 14336]
-
blk.6.ffn_up.weightI8[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightI8[4096 1024]
-
blk.6.attn_output.weightI8[4096 4096]
-
blk.6.attn_q.weightI8[4096 4096]
-
blk.6.attn_v.weightQ4_K[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightI8[14336 4096]
-
blk.7.ffn_gate.weightI8[4096 14336]
-
blk.7.ffn_up.weightI8[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightI8[4096 1024]
-
blk.7.attn_output.weightI8[4096 4096]
-
blk.7.attn_q.weightI8[4096 4096]
-
blk.7.attn_v.weightQ4_K[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightI8[14336 4096]
-
blk.8.ffn_gate.weightI8[4096 14336]
-
blk.8.ffn_up.weightI8[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightI8[4096 1024]
-
blk.8.attn_output.weightI8[4096 4096]
-
blk.8.attn_q.weightI8[4096 4096]
-
blk.8.attn_v.weightQ4_K[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightI8[14336 4096]
-
blk.9.ffn_gate.weightI8[4096 14336]
-
blk.9.ffn_up.weightI8[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightI8[4096 1024]
-
blk.9.attn_output.weightI8[4096 4096]
-
blk.9.attn_q.weightI8[4096 4096]
-
blk.9.attn_v.weightQ4_K[4096 1024]
-
blk.10.ffn_gate.weightI8[4096 14336]
-
blk.10.ffn_up.weightI8[4096 14336]
-
blk.10.attn_k.weightI8[4096 1024]
-
blk.10.attn_output.weightI8[4096 4096]
-
blk.10.attn_q.weightI8[4096 4096]
-
blk.10.attn_v.weightQ4_K[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightI8[14336 4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightI8[14336 4096]
-
blk.11.ffn_gate.weightI8[4096 14336]
-
blk.11.ffn_up.weightI8[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightI8[4096 1024]
-
blk.11.attn_output.weightI8[4096 4096]
-
blk.11.attn_q.weightI8[4096 4096]
-
blk.11.attn_v.weightQ4_K[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightI8[14336 4096]
-
blk.12.ffn_gate.weightI8[4096 14336]
-
blk.12.ffn_up.weightI8[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightI8[4096 1024]
-
blk.12.attn_output.weightI8[4096 4096]
-
blk.12.attn_q.weightI8[4096 4096]
-
blk.12.attn_v.weightQ4_K[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightI8[14336 4096]
-
blk.13.ffn_gate.weightI8[4096 14336]
-
blk.13.ffn_up.weightI8[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightI8[4096 1024]
-
blk.13.attn_output.weightI8[4096 4096]
-
blk.13.attn_q.weightI8[4096 4096]
-
blk.13.attn_v.weightQ4_K[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightI8[14336 4096]
-
blk.14.ffn_gate.weightI8[4096 14336]
-
blk.14.ffn_up.weightI8[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightI8[4096 1024]
-
blk.14.attn_output.weightI8[4096 4096]
-
blk.14.attn_q.weightI8[4096 4096]
-
blk.14.attn_v.weightQ4_K[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightI8[14336 4096]
-
blk.15.ffn_gate.weightI8[4096 14336]
-
blk.15.ffn_up.weightI8[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightI8[4096 1024]
-
blk.15.attn_output.weightI8[4096 4096]
-
blk.15.attn_q.weightI8[4096 4096]
-
blk.15.attn_v.weightQ4_K[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightI8[14336 4096]
-
blk.16.ffn_gate.weightI8[4096 14336]
-
blk.16.ffn_up.weightI8[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightI8[4096 1024]
-
blk.16.attn_output.weightI8[4096 4096]
-
blk.16.attn_q.weightI8[4096 4096]
-
blk.16.attn_v.weightQ4_K[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightI8[14336 4096]
-
blk.17.ffn_gate.weightI8[4096 14336]
-
blk.17.ffn_up.weightI8[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightI8[4096 1024]
-
blk.17.attn_output.weightI8[4096 4096]
-
blk.17.attn_q.weightI8[4096 4096]
-
blk.17.attn_v.weightQ4_K[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightI8[14336 4096]
-
blk.18.ffn_gate.weightI8[4096 14336]
-
blk.18.ffn_up.weightI8[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightI8[4096 1024]
-
blk.18.attn_output.weightI8[4096 4096]
-
blk.18.attn_q.weightI8[4096 4096]
-
blk.18.attn_v.weightQ4_K[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightI8[14336 4096]
-
blk.19.ffn_gate.weightI8[4096 14336]
-
blk.19.ffn_up.weightI8[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightI8[4096 1024]
-
blk.19.attn_output.weightI8[4096 4096]
-
blk.19.attn_q.weightI8[4096 4096]
-
blk.19.attn_v.weightQ4_K[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightI8[14336 4096]
-
blk.20.ffn_gate.weightI8[4096 14336]
-
blk.20.ffn_up.weightI8[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightI8[4096 1024]
-
blk.20.attn_output.weightI8[4096 4096]
-
blk.20.attn_q.weightI8[4096 4096]
-
blk.20.attn_v.weightQ4_K[4096 1024]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightI8[14336 4096]
-
blk.21.ffn_gate.weightI8[4096 14336]
-
blk.21.ffn_up.weightI8[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightI8[4096 1024]
-
blk.21.attn_output.weightI8[4096 4096]
-
blk.21.attn_q.weightI8[4096 4096]
-
blk.21.attn_v.weightQ4_K[4096 1024]
-
blk.22.attn_k.weightI8[4096 1024]
-
blk.22.attn_output.weightI8[4096 4096]
-
blk.22.attn_q.weightI8[4096 4096]
-
blk.22.attn_v.weightQ4_K[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightI8[14336 4096]
-
blk.22.ffn_gate.weightI8[4096 14336]
-
blk.22.ffn_up.weightI8[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightI8[14336 4096]
-
blk.23.ffn_gate.weightI8[4096 14336]
-
blk.23.ffn_up.weightI8[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightI8[4096 1024]
-
blk.23.attn_output.weightI8[4096 4096]
-
blk.23.attn_q.weightI8[4096 4096]
-
blk.23.attn_v.weightQ4_K[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightI8[14336 4096]
-
blk.24.ffn_gate.weightI8[4096 14336]
-
blk.24.ffn_up.weightI8[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightI8[4096 1024]
-
blk.24.attn_output.weightI8[4096 4096]
-
blk.24.attn_q.weightI8[4096 4096]
-
blk.24.attn_v.weightQ4_K[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightI8[14336 4096]
-
blk.25.ffn_gate.weightI8[4096 14336]
-
blk.25.ffn_up.weightI8[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightI8[4096 1024]
-
blk.25.attn_output.weightI8[4096 4096]
-
blk.25.attn_q.weightI8[4096 4096]
-
blk.25.attn_v.weightQ4_K[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightI8[14336 4096]
-
blk.26.ffn_gate.weightI8[4096 14336]
-
blk.26.ffn_up.weightI8[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightI8[4096 1024]
-
blk.26.attn_output.weightI8[4096 4096]
-
blk.26.attn_q.weightI8[4096 4096]
-
blk.26.attn_v.weightQ4_K[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightI8[14336 4096]
-
blk.27.ffn_gate.weightI8[4096 14336]
-
blk.27.ffn_up.weightI8[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightI8[4096 1024]
-
blk.27.attn_output.weightI8[4096 4096]
-
blk.27.attn_q.weightI8[4096 4096]
-
blk.27.attn_v.weightQ4_K[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightI8[14336 4096]
-
blk.28.ffn_gate.weightI8[4096 14336]
-
blk.28.ffn_up.weightI8[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightI8[4096 1024]
-
blk.28.attn_output.weightI8[4096 4096]
-
blk.28.attn_q.weightI8[4096 4096]
-
blk.28.attn_v.weightQ4_K[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightI8[14336 4096]
-
blk.29.ffn_gate.weightI8[4096 14336]
-
blk.29.ffn_up.weightI8[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightI8[4096 1024]
-
blk.29.attn_output.weightI8[4096 4096]
-
blk.29.attn_q.weightI8[4096 4096]
-
blk.29.attn_v.weightQ4_K[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightI8[14336 4096]
-
blk.30.ffn_gate.weightI8[4096 14336]
-
blk.30.ffn_up.weightI8[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightI8[4096 1024]
-
blk.30.attn_output.weightI8[4096 4096]
-
blk.30.attn_q.weightI8[4096 4096]
-
blk.30.attn_v.weightQ4_K[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightI8[14336 4096]
-
blk.31.ffn_gate.weightI8[4096 14336]
-
blk.31.ffn_up.weightI8[4096 14336]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightI8[4096 1024]
-
blk.31.attn_output.weightI8[4096 4096]
-
blk.31.attn_q.weightI8[4096 4096]
-
blk.31.attn_v.weightQ4_K[4096 1024]
-
output.weightQ5_K[4096 32000]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31