latest
4.4GB
Hermes 2 Pro on Mistral 7B is the new flagship 7B Hermes!
7B
679 Pulls Updated 6 months ago
02eb767cc232 · 5.0GB
-
general.architecturellama
-
general.file_typeQ5_K_S
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length32768
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base10000
-
tokenizer.ggml.add_bos_tokentrue
-
tokenizer.ggml.add_eos_tokenfalse
-
tokenizer.ggml.bos_token_id1
-
tokenizer.ggml.eos_token_id32000
-
tokenizer.ggml.modelllama
-
tokenizer.ggml.scores[0 0 0 0 0 ...]
-
tokenizer.ggml.token_type[2 3 3 6 6 ...]
-
tokenizer.ggml.tokens[<unk> <s> </s> <0x00> <0x01> ...]
-
NameTypeShape
-
token_embd.weightQ5_K[4096 32032]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ5_K[14336 4096]
-
blk.0.ffn_gate.weightQ5_K[4096 14336]
-
blk.0.ffn_up.weightQ5_K[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightQ5_K[4096 1024]
-
blk.0.attn_output.weightQ5_K[4096 4096]
-
blk.0.attn_q.weightQ5_K[4096 4096]
-
blk.0.attn_v.weightQ5_K[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ5_K[14336 4096]
-
blk.1.ffn_gate.weightQ5_K[4096 14336]
-
blk.1.ffn_up.weightQ5_K[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightQ5_K[4096 1024]
-
blk.1.attn_output.weightQ5_K[4096 4096]
-
blk.1.attn_q.weightQ5_K[4096 4096]
-
blk.1.attn_v.weightQ5_K[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ5_K[14336 4096]
-
blk.2.ffn_gate.weightQ5_K[4096 14336]
-
blk.2.ffn_up.weightQ5_K[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightQ5_K[4096 1024]
-
blk.2.attn_output.weightQ5_K[4096 4096]
-
blk.2.attn_q.weightQ5_K[4096 4096]
-
blk.2.attn_v.weightQ5_K[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ5_K[14336 4096]
-
blk.3.ffn_gate.weightQ5_K[4096 14336]
-
blk.3.ffn_up.weightQ5_K[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightQ5_K[4096 1024]
-
blk.3.attn_output.weightQ5_K[4096 4096]
-
blk.3.attn_q.weightQ5_K[4096 4096]
-
blk.3.attn_v.weightQ5_K[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightQ5_K[14336 4096]
-
blk.4.ffn_gate.weightQ5_K[4096 14336]
-
blk.4.ffn_up.weightQ5_K[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightQ5_K[4096 1024]
-
blk.4.attn_output.weightQ5_K[4096 4096]
-
blk.4.attn_q.weightQ5_K[4096 4096]
-
blk.4.attn_v.weightQ5_K[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightQ5_K[14336 4096]
-
blk.5.ffn_gate.weightQ5_K[4096 14336]
-
blk.5.ffn_up.weightQ5_K[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightQ5_K[4096 1024]
-
blk.5.attn_output.weightQ5_K[4096 4096]
-
blk.5.attn_q.weightQ5_K[4096 4096]
-
blk.5.attn_v.weightQ5_K[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightQ5_K[14336 4096]
-
blk.6.ffn_gate.weightQ5_K[4096 14336]
-
blk.6.ffn_up.weightQ5_K[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightQ5_K[4096 1024]
-
blk.6.attn_output.weightQ5_K[4096 4096]
-
blk.6.attn_q.weightQ5_K[4096 4096]
-
blk.6.attn_v.weightQ5_K[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightQ5_K[14336 4096]
-
blk.7.ffn_gate.weightQ5_K[4096 14336]
-
blk.7.ffn_up.weightQ5_K[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightQ5_K[4096 1024]
-
blk.7.attn_output.weightQ5_K[4096 4096]
-
blk.7.attn_q.weightQ5_K[4096 4096]
-
blk.7.attn_v.weightQ5_K[4096 1024]
-
blk.8.ffn_gate.weightQ5_K[4096 14336]
-
blk.8.attn_k.weightQ5_K[4096 1024]
-
blk.8.attn_output.weightQ5_K[4096 4096]
-
blk.8.attn_q.weightQ5_K[4096 4096]
-
blk.8.attn_v.weightQ5_K[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightQ5_K[14336 4096]
-
blk.8.ffn_up.weightQ5_K[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightQ5_K[14336 4096]
-
blk.9.ffn_gate.weightQ5_K[4096 14336]
-
blk.9.ffn_up.weightQ5_K[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightQ5_K[4096 1024]
-
blk.9.attn_output.weightQ5_K[4096 4096]
-
blk.9.attn_q.weightQ5_K[4096 4096]
-
blk.9.attn_v.weightQ5_K[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightQ5_K[14336 4096]
-
blk.10.ffn_gate.weightQ5_K[4096 14336]
-
blk.10.ffn_up.weightQ5_K[4096 14336]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightQ5_K[4096 1024]
-
blk.10.attn_output.weightQ5_K[4096 4096]
-
blk.10.attn_q.weightQ5_K[4096 4096]
-
blk.10.attn_v.weightQ5_K[4096 1024]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightQ5_K[14336 4096]
-
blk.11.ffn_gate.weightQ5_K[4096 14336]
-
blk.11.ffn_up.weightQ5_K[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightQ5_K[4096 1024]
-
blk.11.attn_output.weightQ5_K[4096 4096]
-
blk.11.attn_q.weightQ5_K[4096 4096]
-
blk.11.attn_v.weightQ5_K[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightQ5_K[14336 4096]
-
blk.12.ffn_gate.weightQ5_K[4096 14336]
-
blk.12.ffn_up.weightQ5_K[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightQ5_K[4096 1024]
-
blk.12.attn_output.weightQ5_K[4096 4096]
-
blk.12.attn_q.weightQ5_K[4096 4096]
-
blk.12.attn_v.weightQ5_K[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightQ5_K[14336 4096]
-
blk.13.ffn_gate.weightQ5_K[4096 14336]
-
blk.13.ffn_up.weightQ5_K[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightQ5_K[4096 1024]
-
blk.13.attn_output.weightQ5_K[4096 4096]
-
blk.13.attn_q.weightQ5_K[4096 4096]
-
blk.13.attn_v.weightQ5_K[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightQ5_K[14336 4096]
-
blk.14.ffn_gate.weightQ5_K[4096 14336]
-
blk.14.ffn_up.weightQ5_K[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightQ5_K[4096 1024]
-
blk.14.attn_output.weightQ5_K[4096 4096]
-
blk.14.attn_q.weightQ5_K[4096 4096]
-
blk.14.attn_v.weightQ5_K[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightQ5_K[14336 4096]
-
blk.15.ffn_gate.weightQ5_K[4096 14336]
-
blk.15.ffn_up.weightQ5_K[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightQ5_K[4096 1024]
-
blk.15.attn_output.weightQ5_K[4096 4096]
-
blk.15.attn_q.weightQ5_K[4096 4096]
-
blk.15.attn_v.weightQ5_K[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightQ5_K[14336 4096]
-
blk.16.ffn_gate.weightQ5_K[4096 14336]
-
blk.16.ffn_up.weightQ5_K[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightQ5_K[4096 1024]
-
blk.16.attn_output.weightQ5_K[4096 4096]
-
blk.16.attn_q.weightQ5_K[4096 4096]
-
blk.16.attn_v.weightQ5_K[4096 1024]
-
blk.17.ffn_gate.weightQ5_K[4096 14336]
-
blk.17.attn_k.weightQ5_K[4096 1024]
-
blk.17.attn_output.weightQ5_K[4096 4096]
-
blk.17.attn_q.weightQ5_K[4096 4096]
-
blk.17.attn_v.weightQ5_K[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightQ5_K[14336 4096]
-
blk.17.ffn_up.weightQ5_K[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightQ5_K[14336 4096]
-
blk.18.ffn_gate.weightQ5_K[4096 14336]
-
blk.18.ffn_up.weightQ5_K[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightQ5_K[4096 1024]
-
blk.18.attn_output.weightQ5_K[4096 4096]
-
blk.18.attn_q.weightQ5_K[4096 4096]
-
blk.18.attn_v.weightQ5_K[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightQ5_K[14336 4096]
-
blk.19.ffn_gate.weightQ5_K[4096 14336]
-
blk.19.ffn_up.weightQ5_K[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightQ5_K[4096 1024]
-
blk.19.attn_output.weightQ5_K[4096 4096]
-
blk.19.attn_q.weightQ5_K[4096 4096]
-
blk.19.attn_v.weightQ5_K[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightQ5_K[14336 4096]
-
blk.20.ffn_gate.weightQ5_K[4096 14336]
-
blk.20.ffn_up.weightQ5_K[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.20.attn_k.weightQ5_K[4096 1024]
-
blk.20.attn_output.weightQ5_K[4096 4096]
-
blk.20.attn_q.weightQ5_K[4096 4096]
-
blk.20.attn_v.weightQ5_K[4096 1024]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightQ5_K[14336 4096]
-
blk.21.ffn_gate.weightQ5_K[4096 14336]
-
blk.21.ffn_up.weightQ5_K[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightQ5_K[4096 1024]
-
blk.21.attn_output.weightQ5_K[4096 4096]
-
blk.21.attn_q.weightQ5_K[4096 4096]
-
blk.21.attn_v.weightQ5_K[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightQ5_K[14336 4096]
-
blk.22.ffn_gate.weightQ5_K[4096 14336]
-
blk.22.ffn_up.weightQ5_K[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightQ5_K[4096 1024]
-
blk.22.attn_output.weightQ5_K[4096 4096]
-
blk.22.attn_q.weightQ5_K[4096 4096]
-
blk.22.attn_v.weightQ5_K[4096 1024]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightQ5_K[14336 4096]
-
blk.23.ffn_gate.weightQ5_K[4096 14336]
-
blk.23.ffn_up.weightQ5_K[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightQ5_K[4096 1024]
-
blk.23.attn_output.weightQ5_K[4096 4096]
-
blk.23.attn_q.weightQ5_K[4096 4096]
-
blk.23.attn_v.weightQ5_K[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightQ5_K[14336 4096]
-
blk.24.ffn_gate.weightQ5_K[4096 14336]
-
blk.24.ffn_up.weightQ5_K[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightQ5_K[4096 1024]
-
blk.24.attn_output.weightQ5_K[4096 4096]
-
blk.24.attn_q.weightQ5_K[4096 4096]
-
blk.24.attn_v.weightQ5_K[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightQ5_K[14336 4096]
-
blk.25.ffn_gate.weightQ5_K[4096 14336]
-
blk.25.ffn_up.weightQ5_K[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightQ5_K[4096 1024]
-
blk.25.attn_output.weightQ5_K[4096 4096]
-
blk.25.attn_q.weightQ5_K[4096 4096]
-
blk.25.attn_v.weightQ5_K[4096 1024]
-
blk.26.ffn_gate.weightQ5_K[4096 14336]
-
blk.26.attn_k.weightQ5_K[4096 1024]
-
blk.26.attn_output.weightQ5_K[4096 4096]
-
blk.26.attn_q.weightQ5_K[4096 4096]
-
blk.26.attn_v.weightQ5_K[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightQ5_K[14336 4096]
-
blk.26.ffn_up.weightQ5_K[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightQ5_K[14336 4096]
-
blk.27.ffn_gate.weightQ5_K[4096 14336]
-
blk.27.ffn_up.weightQ5_K[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightQ5_K[4096 1024]
-
blk.27.attn_output.weightQ5_K[4096 4096]
-
blk.27.attn_q.weightQ5_K[4096 4096]
-
blk.27.attn_v.weightQ5_K[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightQ5_K[14336 4096]
-
blk.28.ffn_gate.weightQ5_K[4096 14336]
-
blk.28.ffn_up.weightQ5_K[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightQ5_K[4096 1024]
-
blk.28.attn_output.weightQ5_K[4096 4096]
-
blk.28.attn_q.weightQ5_K[4096 4096]
-
blk.28.attn_v.weightQ5_K[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightQ5_K[14336 4096]
-
blk.29.ffn_gate.weightQ5_K[4096 14336]
-
blk.29.ffn_up.weightQ5_K[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightQ5_K[4096 1024]
-
blk.29.attn_output.weightQ5_K[4096 4096]
-
blk.29.attn_q.weightQ5_K[4096 4096]
-
blk.29.attn_v.weightQ5_K[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightQ5_K[14336 4096]
-
blk.30.ffn_gate.weightQ5_K[4096 14336]
-
blk.30.ffn_up.weightQ5_K[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightQ5_K[4096 1024]
-
blk.30.attn_output.weightQ5_K[4096 4096]
-
blk.30.attn_q.weightQ5_K[4096 4096]
-
blk.30.attn_v.weightQ5_K[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightQ5_K[14336 4096]
-
blk.31.ffn_gate.weightQ5_K[4096 14336]
-
blk.31.ffn_up.weightQ5_K[4096 14336]
-
blk.31.ffn_norm.weightF32[4096]
-
blk.31.attn_k.weightQ5_K[4096 1024]
-
blk.31.attn_output.weightQ5_K[4096 4096]
-
blk.31.attn_q.weightQ5_K[4096 4096]
-
blk.31.attn_v.weightQ5_K[4096 1024]
-
output.weightQ6_K[4096 32032]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31