Quantized version of cognitivecomputations/dolphin-2.8-experiment26-7b
7B
49 Pulls Updated 6 months ago
d1ef248c2d70 · 5.1GB
-
general.architecturellama
-
general.file_typeQ5_K_M
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length32768
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id32000
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
NameTypeShape
-
token_embd.weightQ5_K[4096 32002]
-
blk.0.attn_q.weightQ5_K[4096 4096]
-
blk.0.attn_k.weightQ5_K[4096 1024]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.0.attn_output.weightQ5_K[4096 4096]
-
blk.0.ffn_gate.weightQ5_K[4096 14336]
-
blk.0.ffn_up.weightQ5_K[4096 14336]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.1.attn_q.weightQ5_K[4096 4096]
-
blk.1.attn_k.weightQ5_K[4096 1024]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_output.weightQ5_K[4096 4096]
-
blk.1.ffn_gate.weightQ5_K[4096 14336]
-
blk.1.ffn_up.weightQ5_K[4096 14336]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.2.attn_q.weightQ5_K[4096 4096]
-
blk.2.attn_k.weightQ5_K[4096 1024]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_output.weightQ5_K[4096 4096]
-
blk.2.ffn_gate.weightQ5_K[4096 14336]
-
blk.2.ffn_up.weightQ5_K[4096 14336]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.3.attn_q.weightQ5_K[4096 4096]
-
blk.3.attn_k.weightQ5_K[4096 1024]
-
blk.3.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_output.weightQ5_K[4096 4096]
-
blk.3.ffn_gate.weightQ5_K[4096 14336]
-
blk.3.ffn_up.weightQ5_K[4096 14336]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.4.attn_q.weightQ5_K[4096 4096]
-
blk.4.attn_k.weightQ5_K[4096 1024]
-
blk.4.attn_v.weightQ5_K[4096 1024]
-
blk.4.attn_output.weightQ5_K[4096 4096]
-
blk.4.ffn_gate.weightQ5_K[4096 14336]
-
blk.4.ffn_up.weightQ5_K[4096 14336]
-
blk.4.ffn_down.weightQ5_K[14336 4096]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.5.attn_q.weightQ5_K[4096 4096]
-
blk.5.attn_k.weightQ5_K[4096 1024]
-
blk.5.attn_v.weightQ5_K[4096 1024]
-
blk.5.attn_output.weightQ5_K[4096 4096]
-
blk.5.ffn_gate.weightQ5_K[4096 14336]
-
blk.5.ffn_up.weightQ5_K[4096 14336]
-
blk.5.ffn_down.weightQ5_K[14336 4096]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.6.attn_q.weightQ5_K[4096 4096]
-
blk.6.attn_k.weightQ5_K[4096 1024]
-
blk.6.attn_v.weightQ6_K[4096 1024]
-
blk.6.attn_output.weightQ5_K[4096 4096]
-
blk.6.ffn_gate.weightQ5_K[4096 14336]
-
blk.6.ffn_up.weightQ5_K[4096 14336]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.7.attn_q.weightQ5_K[4096 4096]
-
blk.7.attn_k.weightQ5_K[4096 1024]
-
blk.7.attn_v.weightQ5_K[4096 1024]
-
blk.7.attn_output.weightQ5_K[4096 4096]
-
blk.7.ffn_gate.weightQ5_K[4096 14336]
-
blk.7.ffn_up.weightQ5_K[4096 14336]
-
blk.7.ffn_down.weightQ5_K[14336 4096]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.8.attn_q.weightQ5_K[4096 4096]
-
blk.8.attn_k.weightQ5_K[4096 1024]
-
blk.8.attn_v.weightQ5_K[4096 1024]
-
blk.8.attn_output.weightQ5_K[4096 4096]
-
blk.8.ffn_gate.weightQ5_K[4096 14336]
-
blk.8.ffn_up.weightQ5_K[4096 14336]
-
blk.8.ffn_down.weightQ5_K[14336 4096]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.9.attn_q.weightQ5_K[4096 4096]
-
blk.9.attn_k.weightQ5_K[4096 1024]
-
blk.9.attn_v.weightQ6_K[4096 1024]
-
blk.9.attn_output.weightQ5_K[4096 4096]
-
blk.9.ffn_gate.weightQ5_K[4096 14336]
-
blk.9.ffn_up.weightQ5_K[4096 14336]
-
blk.9.ffn_down.weightQ6_K[14336 4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.10.attn_q.weightQ5_K[4096 4096]
-
blk.10.attn_k.weightQ5_K[4096 1024]
-
blk.10.attn_v.weightQ5_K[4096 1024]
-
blk.10.attn_output.weightQ5_K[4096 4096]
-
blk.10.ffn_gate.weightQ5_K[4096 14336]
-
blk.10.ffn_up.weightQ5_K[4096 14336]
-
blk.10.ffn_down.weightQ5_K[14336 4096]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.11.attn_q.weightQ5_K[4096 4096]
-
blk.11.attn_k.weightQ5_K[4096 1024]
-
blk.11.attn_v.weightQ5_K[4096 1024]
-
blk.11.attn_output.weightQ5_K[4096 4096]
-
blk.11.ffn_gate.weightQ5_K[4096 14336]
-
blk.11.ffn_up.weightQ5_K[4096 14336]
-
blk.11.ffn_down.weightQ5_K[14336 4096]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.12.attn_q.weightQ5_K[4096 4096]
-
blk.12.attn_k.weightQ5_K[4096 1024]
-
blk.12.attn_v.weightQ6_K[4096 1024]
-
blk.12.attn_output.weightQ5_K[4096 4096]
-
blk.12.ffn_gate.weightQ5_K[4096 14336]
-
blk.12.ffn_up.weightQ5_K[4096 14336]
-
blk.12.ffn_down.weightQ6_K[14336 4096]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.13.attn_q.weightQ5_K[4096 4096]
-
blk.13.attn_k.weightQ5_K[4096 1024]
-
blk.13.attn_v.weightQ5_K[4096 1024]
-
blk.13.attn_output.weightQ5_K[4096 4096]
-
blk.13.ffn_gate.weightQ5_K[4096 14336]
-
blk.13.ffn_up.weightQ5_K[4096 14336]
-
blk.13.ffn_down.weightQ5_K[14336 4096]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.14.attn_q.weightQ5_K[4096 4096]
-
blk.14.attn_k.weightQ5_K[4096 1024]
-
blk.14.attn_v.weightQ5_K[4096 1024]
-
blk.14.attn_output.weightQ5_K[4096 4096]
-
blk.14.ffn_gate.weightQ5_K[4096 14336]
-
blk.14.ffn_up.weightQ5_K[4096 14336]
-
blk.14.ffn_down.weightQ5_K[14336 4096]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.15.attn_q.weightQ5_K[4096 4096]
-
blk.15.attn_k.weightQ5_K[4096 1024]
-
blk.15.attn_v.weightQ6_K[4096 1024]
-
blk.15.attn_output.weightQ5_K[4096 4096]
-
blk.15.ffn_gate.weightQ5_K[4096 14336]
-
blk.15.ffn_up.weightQ5_K[4096 14336]
-
blk.15.ffn_down.weightQ6_K[14336 4096]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.16.attn_q.weightQ5_K[4096 4096]
-
blk.16.attn_k.weightQ5_K[4096 1024]
-
blk.16.attn_v.weightQ5_K[4096 1024]
-
blk.16.attn_output.weightQ5_K[4096 4096]
-
blk.16.ffn_gate.weightQ5_K[4096 14336]
-
blk.16.ffn_up.weightQ5_K[4096 14336]
-
blk.16.ffn_down.weightQ5_K[14336 4096]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.17.attn_q.weightQ5_K[4096 4096]
-
blk.17.attn_k.weightQ5_K[4096 1024]
-
blk.17.attn_v.weightQ5_K[4096 1024]
-
blk.17.attn_output.weightQ5_K[4096 4096]
-
blk.17.ffn_gate.weightQ5_K[4096 14336]
-
blk.17.ffn_up.weightQ5_K[4096 14336]
-
blk.17.ffn_down.weightQ5_K[14336 4096]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.18.attn_q.weightQ5_K[4096 4096]
-
blk.18.attn_k.weightQ5_K[4096 1024]
-
blk.18.attn_v.weightQ6_K[4096 1024]
-
blk.18.attn_output.weightQ5_K[4096 4096]
-
blk.18.ffn_gate.weightQ5_K[4096 14336]
-
blk.18.ffn_up.weightQ5_K[4096 14336]
-
blk.18.ffn_down.weightQ6_K[14336 4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.19.attn_q.weightQ5_K[4096 4096]
-
blk.19.attn_k.weightQ5_K[4096 1024]
-
blk.19.attn_v.weightQ5_K[4096 1024]
-
blk.19.attn_output.weightQ5_K[4096 4096]
-
blk.19.ffn_gate.weightQ5_K[4096 14336]
-
blk.19.ffn_up.weightQ5_K[4096 14336]
-
blk.19.ffn_down.weightQ5_K[14336 4096]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.20.attn_q.weightQ5_K[4096 4096]
-
blk.20.attn_k.weightQ5_K[4096 1024]
-
blk.20.attn_v.weightQ5_K[4096 1024]
-
blk.20.attn_output.weightQ5_K[4096 4096]
-
blk.20.ffn_gate.weightQ5_K[4096 14336]
-
blk.20.ffn_up.weightQ5_K[4096 14336]
-
blk.20.ffn_down.weightQ5_K[14336 4096]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_q.weightQ5_K[4096 4096]
-
blk.21.attn_k.weightQ5_K[4096 1024]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.21.attn_output.weightQ5_K[4096 4096]
-
blk.21.ffn_gate.weightQ5_K[4096 14336]
-
blk.21.ffn_up.weightQ5_K[4096 14336]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.22.attn_q.weightQ5_K[4096 4096]
-
blk.22.attn_k.weightQ5_K[4096 1024]
-
blk.22.attn_v.weightQ5_K[4096 1024]
-
blk.22.attn_output.weightQ5_K[4096 4096]
-
blk.22.ffn_gate.weightQ5_K[4096 14336]
-
blk.22.ffn_up.weightQ5_K[4096 14336]
-
blk.22.ffn_down.weightQ5_K[14336 4096]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.23.attn_q.weightQ5_K[4096 4096]
-
blk.23.attn_k.weightQ5_K[4096 1024]
-
blk.23.attn_v.weightQ5_K[4096 1024]
-
blk.23.attn_output.weightQ5_K[4096 4096]
-
blk.23.ffn_gate.weightQ5_K[4096 14336]
-
blk.23.ffn_up.weightQ5_K[4096 14336]
-
blk.23.ffn_down.weightQ5_K[14336 4096]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.24.attn_q.weightQ5_K[4096 4096]
-
blk.24.attn_k.weightQ5_K[4096 1024]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.24.attn_output.weightQ5_K[4096 4096]
-
blk.24.ffn_gate.weightQ5_K[4096 14336]
-
blk.24.ffn_up.weightQ5_K[4096 14336]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.25.attn_q.weightQ5_K[4096 4096]
-
blk.25.attn_k.weightQ5_K[4096 1024]
-
blk.25.attn_v.weightQ5_K[4096 1024]
-
blk.25.attn_output.weightQ5_K[4096 4096]
-
blk.25.ffn_gate.weightQ5_K[4096 14336]
-
blk.25.ffn_up.weightQ5_K[4096 14336]
-
blk.25.ffn_down.weightQ5_K[14336 4096]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.26.attn_q.weightQ5_K[4096 4096]
-
blk.26.attn_k.weightQ5_K[4096 1024]
-
blk.26.attn_v.weightQ5_K[4096 1024]
-
blk.26.attn_output.weightQ5_K[4096 4096]
-
blk.26.ffn_gate.weightQ5_K[4096 14336]
-
blk.26.ffn_up.weightQ5_K[4096 14336]
-
blk.26.ffn_down.weightQ5_K[14336 4096]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.27.attn_q.weightQ5_K[4096 4096]
-
blk.27.attn_k.weightQ5_K[4096 1024]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.27.attn_output.weightQ5_K[4096 4096]
-
blk.27.ffn_gate.weightQ5_K[4096 14336]
-
blk.27.ffn_up.weightQ5_K[4096 14336]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.28.attn_q.weightQ5_K[4096 4096]
-
blk.28.attn_k.weightQ5_K[4096 1024]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_output.weightQ5_K[4096 4096]
-
blk.28.ffn_gate.weightQ5_K[4096 14336]
-
blk.28.ffn_up.weightQ5_K[4096 14336]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.29.attn_q.weightQ5_K[4096 4096]
-
blk.29.attn_k.weightQ5_K[4096 1024]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_output.weightQ5_K[4096 4096]
-
blk.29.ffn_gate.weightQ5_K[4096 14336]
-
blk.29.ffn_up.weightQ5_K[4096 14336]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.30.attn_q.weightQ5_K[4096 4096]
-
blk.30.attn_k.weightQ5_K[4096 1024]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_output.weightQ5_K[4096 4096]
-
blk.30.ffn_gate.weightQ5_K[4096 14336]
-
blk.30.ffn_up.weightQ5_K[4096 14336]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.31.attn_q.weightQ5_K[4096 4096]
-
blk.31.attn_k.weightQ5_K[4096 1024]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_output.weightQ5_K[4096 4096]
-
blk.31.ffn_gate.weightQ5_K[4096 14336]
-
blk.31.ffn_up.weightQ5_K[4096 14336]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output_norm.weightF32[4096]
-
output.weightQ6_K[4096 32002]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31