latest
3.2GB
Ollama import of TheDrummer/Llama-3SOME-8B-v1-BETA-GGUF
8B
171 Pulls Updated 4 months ago
615ce8745117 · 3.2GB
-
general.architecturellama
-
general.file_typeQ2_K
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length24576
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base500000
-
llama.rope.scaling.factor3
-
llama.rope.scaling.typelinear
-
llama.vocab_size128256
-
tokenizer.ggml.bos_token_id128000
-
tokenizer.ggml.eos_token_id128001
-
tokenizer.ggml.merges[Ġ Ġ Ġ ĠĠĠ ĠĠ ĠĠ ĠĠĠ Ġ i n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.padding_token_id128001
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightQ2_K[4096 128256]
-
blk.0.attn_q.weightQ2_K[4096 4096]
-
blk.0.attn_k.weightQ2_K[4096 1024]
-
blk.0.attn_v.weightQ4_K[4096 1024]
-
blk.0.attn_output.weightQ3_K[4096 4096]
-
blk.0.ffn_gate.weightQ2_K[4096 14336]
-
blk.0.ffn_up.weightQ2_K[4096 14336]
-
blk.0.ffn_down.weightQ3_K[14336 4096]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.1.attn_q.weightQ2_K[4096 4096]
-
blk.1.attn_k.weightQ2_K[4096 1024]
-
blk.1.attn_v.weightQ4_K[4096 1024]
-
blk.1.attn_output.weightQ3_K[4096 4096]
-
blk.1.ffn_gate.weightQ2_K[4096 14336]
-
blk.1.ffn_up.weightQ2_K[4096 14336]
-
blk.1.ffn_down.weightQ3_K[14336 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.2.attn_q.weightQ2_K[4096 4096]
-
blk.2.attn_k.weightQ2_K[4096 1024]
-
blk.2.attn_v.weightQ4_K[4096 1024]
-
blk.2.attn_output.weightQ3_K[4096 4096]
-
blk.2.ffn_gate.weightQ2_K[4096 14336]
-
blk.2.ffn_up.weightQ2_K[4096 14336]
-
blk.2.ffn_down.weightQ3_K[14336 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.3.attn_q.weightQ2_K[4096 4096]
-
blk.3.attn_k.weightQ2_K[4096 1024]
-
blk.3.attn_v.weightQ4_K[4096 1024]
-
blk.3.attn_output.weightQ3_K[4096 4096]
-
blk.3.ffn_gate.weightQ2_K[4096 14336]
-
blk.3.ffn_up.weightQ2_K[4096 14336]
-
blk.3.ffn_down.weightQ3_K[14336 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.4.attn_q.weightQ2_K[4096 4096]
-
blk.4.attn_k.weightQ2_K[4096 1024]
-
blk.4.attn_v.weightQ4_K[4096 1024]
-
blk.4.attn_output.weightQ3_K[4096 4096]
-
blk.4.ffn_gate.weightQ2_K[4096 14336]
-
blk.4.ffn_up.weightQ2_K[4096 14336]
-
blk.4.ffn_down.weightQ3_K[14336 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.5.attn_q.weightQ2_K[4096 4096]
-
blk.5.attn_k.weightQ2_K[4096 1024]
-
blk.5.attn_v.weightQ4_K[4096 1024]
-
blk.5.attn_output.weightQ3_K[4096 4096]
-
blk.5.ffn_gate.weightQ2_K[4096 14336]
-
blk.5.ffn_up.weightQ2_K[4096 14336]
-
blk.5.ffn_down.weightQ3_K[14336 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.6.attn_q.weightQ2_K[4096 4096]
-
blk.6.attn_k.weightQ2_K[4096 1024]
-
blk.6.attn_v.weightQ4_K[4096 1024]
-
blk.6.attn_output.weightQ3_K[4096 4096]
-
blk.6.ffn_gate.weightQ2_K[4096 14336]
-
blk.6.ffn_up.weightQ2_K[4096 14336]
-
blk.6.ffn_down.weightQ3_K[14336 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.7.attn_q.weightQ2_K[4096 4096]
-
blk.7.attn_k.weightQ2_K[4096 1024]
-
blk.7.attn_v.weightQ4_K[4096 1024]
-
blk.7.attn_output.weightQ3_K[4096 4096]
-
blk.7.ffn_gate.weightQ2_K[4096 14336]
-
blk.7.ffn_up.weightQ2_K[4096 14336]
-
blk.7.ffn_down.weightQ3_K[14336 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.8.attn_q.weightQ2_K[4096 4096]
-
blk.8.attn_k.weightQ2_K[4096 1024]
-
blk.8.attn_v.weightQ4_K[4096 1024]
-
blk.8.attn_output.weightQ3_K[4096 4096]
-
blk.8.ffn_gate.weightQ2_K[4096 14336]
-
blk.8.ffn_up.weightQ2_K[4096 14336]
-
blk.8.ffn_down.weightQ3_K[14336 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.9.attn_q.weightQ2_K[4096 4096]
-
blk.9.attn_k.weightQ2_K[4096 1024]
-
blk.9.attn_v.weightQ4_K[4096 1024]
-
blk.9.attn_output.weightQ3_K[4096 4096]
-
blk.9.ffn_gate.weightQ2_K[4096 14336]
-
blk.9.ffn_up.weightQ2_K[4096 14336]
-
blk.9.ffn_down.weightQ3_K[14336 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.10.attn_q.weightQ2_K[4096 4096]
-
blk.10.attn_k.weightQ2_K[4096 1024]
-
blk.10.attn_v.weightQ4_K[4096 1024]
-
blk.10.attn_output.weightQ3_K[4096 4096]
-
blk.10.ffn_gate.weightQ2_K[4096 14336]
-
blk.10.ffn_up.weightQ2_K[4096 14336]
-
blk.10.ffn_down.weightQ3_K[14336 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_q.weightQ2_K[4096 4096]
-
blk.11.attn_k.weightQ2_K[4096 1024]
-
blk.11.attn_v.weightQ4_K[4096 1024]
-
blk.11.attn_output.weightQ3_K[4096 4096]
-
blk.11.ffn_gate.weightQ2_K[4096 14336]
-
blk.11.ffn_up.weightQ2_K[4096 14336]
-
blk.11.ffn_down.weightQ3_K[14336 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.12.attn_q.weightQ2_K[4096 4096]
-
blk.12.attn_k.weightQ2_K[4096 1024]
-
blk.12.attn_v.weightQ4_K[4096 1024]
-
blk.12.attn_output.weightQ3_K[4096 4096]
-
blk.12.ffn_gate.weightQ2_K[4096 14336]
-
blk.12.ffn_up.weightQ2_K[4096 14336]
-
blk.12.ffn_down.weightQ3_K[14336 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.13.attn_q.weightQ2_K[4096 4096]
-
blk.13.attn_k.weightQ2_K[4096 1024]
-
blk.13.attn_v.weightQ4_K[4096 1024]
-
blk.13.attn_output.weightQ3_K[4096 4096]
-
blk.13.ffn_gate.weightQ2_K[4096 14336]
-
blk.13.ffn_up.weightQ2_K[4096 14336]
-
blk.13.ffn_down.weightQ3_K[14336 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.14.attn_q.weightQ2_K[4096 4096]
-
blk.14.attn_k.weightQ2_K[4096 1024]
-
blk.14.attn_v.weightQ4_K[4096 1024]
-
blk.14.attn_output.weightQ3_K[4096 4096]
-
blk.14.ffn_gate.weightQ2_K[4096 14336]
-
blk.14.ffn_up.weightQ2_K[4096 14336]
-
blk.14.ffn_down.weightQ3_K[14336 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.15.attn_q.weightQ2_K[4096 4096]
-
blk.15.attn_k.weightQ2_K[4096 1024]
-
blk.15.attn_v.weightQ4_K[4096 1024]
-
blk.15.attn_output.weightQ3_K[4096 4096]
-
blk.15.ffn_gate.weightQ2_K[4096 14336]
-
blk.15.ffn_up.weightQ2_K[4096 14336]
-
blk.15.ffn_down.weightQ3_K[14336 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.16.attn_q.weightQ2_K[4096 4096]
-
blk.16.attn_k.weightQ2_K[4096 1024]
-
blk.16.attn_v.weightQ4_K[4096 1024]
-
blk.16.attn_output.weightQ3_K[4096 4096]
-
blk.16.ffn_gate.weightQ2_K[4096 14336]
-
blk.16.ffn_up.weightQ2_K[4096 14336]
-
blk.16.ffn_down.weightQ3_K[14336 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.17.attn_q.weightQ2_K[4096 4096]
-
blk.17.attn_k.weightQ2_K[4096 1024]
-
blk.17.attn_v.weightQ4_K[4096 1024]
-
blk.17.attn_output.weightQ3_K[4096 4096]
-
blk.17.ffn_gate.weightQ2_K[4096 14336]
-
blk.17.ffn_up.weightQ2_K[4096 14336]
-
blk.17.ffn_down.weightQ3_K[14336 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.18.attn_q.weightQ2_K[4096 4096]
-
blk.18.attn_k.weightQ2_K[4096 1024]
-
blk.18.attn_v.weightQ4_K[4096 1024]
-
blk.18.attn_output.weightQ3_K[4096 4096]
-
blk.18.ffn_gate.weightQ2_K[4096 14336]
-
blk.18.ffn_up.weightQ2_K[4096 14336]
-
blk.18.ffn_down.weightQ3_K[14336 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.19.attn_q.weightQ2_K[4096 4096]
-
blk.19.attn_k.weightQ2_K[4096 1024]
-
blk.19.attn_v.weightQ4_K[4096 1024]
-
blk.19.attn_output.weightQ3_K[4096 4096]
-
blk.19.ffn_gate.weightQ2_K[4096 14336]
-
blk.19.ffn_up.weightQ2_K[4096 14336]
-
blk.19.ffn_down.weightQ3_K[14336 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.20.attn_q.weightQ2_K[4096 4096]
-
blk.20.attn_k.weightQ2_K[4096 1024]
-
blk.20.attn_v.weightQ4_K[4096 1024]
-
blk.20.attn_output.weightQ3_K[4096 4096]
-
blk.20.ffn_gate.weightQ2_K[4096 14336]
-
blk.20.ffn_up.weightQ2_K[4096 14336]
-
blk.20.ffn_down.weightQ3_K[14336 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_q.weightQ2_K[4096 4096]
-
blk.21.attn_k.weightQ2_K[4096 1024]
-
blk.21.attn_v.weightQ4_K[4096 1024]
-
blk.21.attn_output.weightQ3_K[4096 4096]
-
blk.21.ffn_gate.weightQ2_K[4096 14336]
-
blk.21.ffn_up.weightQ2_K[4096 14336]
-
blk.21.ffn_down.weightQ3_K[14336 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.22.attn_q.weightQ2_K[4096 4096]
-
blk.22.attn_k.weightQ2_K[4096 1024]
-
blk.22.attn_v.weightQ4_K[4096 1024]
-
blk.22.attn_output.weightQ3_K[4096 4096]
-
blk.22.ffn_gate.weightQ2_K[4096 14336]
-
blk.22.ffn_up.weightQ2_K[4096 14336]
-
blk.22.ffn_down.weightQ3_K[14336 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_q.weightQ2_K[4096 4096]
-
blk.23.attn_k.weightQ2_K[4096 1024]
-
blk.23.attn_v.weightQ4_K[4096 1024]
-
blk.23.attn_output.weightQ3_K[4096 4096]
-
blk.23.ffn_gate.weightQ2_K[4096 14336]
-
blk.23.ffn_up.weightQ2_K[4096 14336]
-
blk.23.ffn_down.weightQ3_K[14336 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_q.weightQ2_K[4096 4096]
-
blk.24.attn_k.weightQ2_K[4096 1024]
-
blk.24.attn_v.weightQ4_K[4096 1024]
-
blk.24.attn_output.weightQ3_K[4096 4096]
-
blk.24.ffn_gate.weightQ2_K[4096 14336]
-
blk.24.ffn_up.weightQ2_K[4096 14336]
-
blk.24.ffn_down.weightQ3_K[14336 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.25.attn_q.weightQ2_K[4096 4096]
-
blk.25.attn_k.weightQ2_K[4096 1024]
-
blk.25.attn_v.weightQ4_K[4096 1024]
-
blk.25.attn_output.weightQ3_K[4096 4096]
-
blk.25.ffn_gate.weightQ2_K[4096 14336]
-
blk.25.ffn_up.weightQ2_K[4096 14336]
-
blk.25.ffn_down.weightQ3_K[14336 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.26.attn_q.weightQ2_K[4096 4096]
-
blk.26.attn_k.weightQ2_K[4096 1024]
-
blk.26.attn_v.weightQ4_K[4096 1024]
-
blk.26.attn_output.weightQ3_K[4096 4096]
-
blk.26.ffn_gate.weightQ2_K[4096 14336]
-
blk.26.ffn_up.weightQ2_K[4096 14336]
-
blk.26.ffn_down.weightQ3_K[14336 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.27.attn_q.weightQ2_K[4096 4096]
-
blk.27.attn_k.weightQ2_K[4096 1024]
-
blk.27.attn_v.weightQ4_K[4096 1024]
-
blk.27.attn_output.weightQ3_K[4096 4096]
-
blk.27.ffn_gate.weightQ2_K[4096 14336]
-
blk.27.ffn_up.weightQ2_K[4096 14336]
-
blk.27.ffn_down.weightQ3_K[14336 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.28.attn_q.weightQ2_K[4096 4096]
-
blk.28.attn_k.weightQ2_K[4096 1024]
-
blk.28.attn_v.weightQ4_K[4096 1024]
-
blk.28.attn_output.weightQ3_K[4096 4096]
-
blk.28.ffn_gate.weightQ2_K[4096 14336]
-
blk.28.ffn_up.weightQ2_K[4096 14336]
-
blk.28.ffn_down.weightQ3_K[14336 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.29.attn_q.weightQ2_K[4096 4096]
-
blk.29.attn_k.weightQ2_K[4096 1024]
-
blk.29.attn_v.weightQ4_K[4096 1024]
-
blk.29.attn_output.weightQ3_K[4096 4096]
-
blk.29.ffn_gate.weightQ2_K[4096 14336]
-
blk.29.ffn_up.weightQ2_K[4096 14336]
-
blk.29.ffn_down.weightQ3_K[14336 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.30.attn_q.weightQ2_K[4096 4096]
-
blk.30.attn_k.weightQ2_K[4096 1024]
-
blk.30.attn_v.weightQ4_K[4096 1024]
-
blk.30.attn_output.weightQ3_K[4096 4096]
-
blk.30.ffn_gate.weightQ2_K[4096 14336]
-
blk.30.ffn_up.weightQ2_K[4096 14336]
-
blk.30.ffn_down.weightQ3_K[14336 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.31.attn_q.weightQ2_K[4096 4096]
-
blk.31.attn_k.weightQ2_K[4096 1024]
-
blk.31.attn_v.weightQ4_K[4096 1024]
-
blk.31.attn_output.weightQ3_K[4096 4096]
-
blk.31.ffn_gate.weightQ2_K[4096 14336]
-
blk.31.ffn_up.weightQ2_K[4096 14336]
-
blk.31.ffn_down.weightQ3_K[14336 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output_norm.weightF32[4096]
-
output.weightQ6_K[4096 128256]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31