latest
5.6GB
NeuralDaredevil-8B-abliterated with more quantization
8B
901 Pulls Updated 3 months ago
a81b3d5a20cd · 6.6GB
-
general.architecturellama
-
general.file_typeQ6_K
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length8192
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base500000
-
llama.vocab_size128256
-
tokenizer.ggml.bos_token_id128000
-
tokenizer.ggml.eos_token_id128001
-
tokenizer.ggml.merges[Ġ Ġ Ġ ĠĠĠ ĠĠ ĠĠ ĠĠĠ Ġ i n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.prellama-bpe
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightQ6_K[4096 128256]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ6_K[14336 4096]
-
blk.0.ffn_gate.weightQ6_K[4096 14336]
-
blk.0.ffn_up.weightQ6_K[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightQ6_K[4096 1024]
-
blk.0.attn_output.weightQ6_K[4096 4096]
-
blk.0.attn_q.weightQ6_K[4096 4096]
-
blk.0.attn_v.weightQ6_K[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ6_K[14336 4096]
-
blk.1.ffn_gate.weightQ6_K[4096 14336]
-
blk.1.ffn_up.weightQ6_K[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightQ6_K[4096 1024]
-
blk.1.attn_output.weightQ6_K[4096 4096]
-
blk.1.attn_q.weightQ6_K[4096 4096]
-
blk.1.attn_v.weightQ6_K[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ6_K[14336 4096]
-
blk.2.ffn_gate.weightQ6_K[4096 14336]
-
blk.2.ffn_up.weightQ6_K[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightQ6_K[4096 1024]
-
blk.2.attn_output.weightQ6_K[4096 4096]
-
blk.2.attn_q.weightQ6_K[4096 4096]
-
blk.2.attn_v.weightQ6_K[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ6_K[14336 4096]
-
blk.3.ffn_gate.weightQ6_K[4096 14336]
-
blk.3.ffn_up.weightQ6_K[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightQ6_K[4096 1024]
-
blk.3.attn_output.weightQ6_K[4096 4096]
-
blk.3.attn_q.weightQ6_K[4096 4096]
-
blk.3.attn_v.weightQ6_K[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightQ6_K[14336 4096]
-
blk.4.ffn_gate.weightQ6_K[4096 14336]
-
blk.4.ffn_up.weightQ6_K[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightQ6_K[4096 1024]
-
blk.4.attn_output.weightQ6_K[4096 4096]
-
blk.4.attn_q.weightQ6_K[4096 4096]
-
blk.4.attn_v.weightQ6_K[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightQ6_K[14336 4096]
-
blk.5.ffn_gate.weightQ6_K[4096 14336]
-
blk.5.ffn_up.weightQ6_K[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightQ6_K[4096 1024]
-
blk.5.attn_output.weightQ6_K[4096 4096]
-
blk.5.attn_q.weightQ6_K[4096 4096]
-
blk.5.attn_v.weightQ6_K[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightQ6_K[14336 4096]
-
blk.6.ffn_gate.weightQ6_K[4096 14336]
-
blk.6.ffn_up.weightQ6_K[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightQ6_K[4096 1024]
-
blk.6.attn_output.weightQ6_K[4096 4096]
-
blk.6.attn_q.weightQ6_K[4096 4096]
-
blk.6.attn_v.weightQ6_K[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightQ6_K[14336 4096]
-
blk.7.ffn_gate.weightQ6_K[4096 14336]
-
blk.7.ffn_up.weightQ6_K[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightQ6_K[4096 1024]
-
blk.7.attn_output.weightQ6_K[4096 4096]
-
blk.7.attn_q.weightQ6_K[4096 4096]
-
blk.7.attn_v.weightQ6_K[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightQ6_K[14336 4096]
-
blk.8.ffn_gate.weightQ6_K[4096 14336]
-
blk.8.ffn_up.weightQ6_K[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightQ6_K[4096 1024]
-
blk.8.attn_output.weightQ6_K[4096 4096]
-
blk.8.attn_q.weightQ6_K[4096 4096]
-
blk.8.attn_v.weightQ6_K[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightQ6_K[14336 4096]
-
blk.9.ffn_gate.weightQ6_K[4096 14336]
-
blk.9.ffn_up.weightQ6_K[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightQ6_K[4096 1024]
-
blk.9.attn_output.weightQ6_K[4096 4096]
-
blk.9.attn_q.weightQ6_K[4096 4096]
-
blk.9.attn_v.weightQ6_K[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightQ6_K[14336 4096]
-
blk.10.ffn_gate.weightQ6_K[4096 14336]
-
blk.10.ffn_up.weightQ6_K[4096 14336]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightQ6_K[4096 1024]
-
blk.10.attn_output.weightQ6_K[4096 4096]
-
blk.10.attn_q.weightQ6_K[4096 4096]
-
blk.10.attn_v.weightQ6_K[4096 1024]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightQ6_K[14336 4096]
-
blk.11.ffn_gate.weightQ6_K[4096 14336]
-
blk.11.ffn_up.weightQ6_K[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightQ6_K[4096 1024]
-
blk.11.attn_output.weightQ6_K[4096 4096]
-
blk.11.attn_q.weightQ6_K[4096 4096]
-
blk.11.attn_v.weightQ6_K[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightQ6_K[14336 4096]
-
blk.12.ffn_gate.weightQ6_K[4096 14336]
-
blk.12.ffn_up.weightQ6_K[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightQ6_K[4096 1024]
-
blk.12.attn_output.weightQ6_K[4096 4096]
-
blk.12.attn_q.weightQ6_K[4096 4096]
-
blk.12.attn_v.weightQ6_K[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightQ6_K[14336 4096]
-
blk.13.ffn_gate.weightQ6_K[4096 14336]
-
blk.13.ffn_up.weightQ6_K[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightQ6_K[4096 1024]
-
blk.13.attn_output.weightQ6_K[4096 4096]
-
blk.13.attn_q.weightQ6_K[4096 4096]
-
blk.13.attn_v.weightQ6_K[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightQ6_K[14336 4096]
-
blk.14.ffn_gate.weightQ6_K[4096 14336]
-
blk.14.ffn_up.weightQ6_K[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightQ6_K[4096 1024]
-
blk.14.attn_output.weightQ6_K[4096 4096]
-
blk.14.attn_q.weightQ6_K[4096 4096]
-
blk.14.attn_v.weightQ6_K[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightQ6_K[14336 4096]
-
blk.15.ffn_gate.weightQ6_K[4096 14336]
-
blk.15.ffn_up.weightQ6_K[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightQ6_K[4096 1024]
-
blk.15.attn_output.weightQ6_K[4096 4096]
-
blk.15.attn_q.weightQ6_K[4096 4096]
-
blk.15.attn_v.weightQ6_K[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightQ6_K[14336 4096]
-
blk.16.ffn_gate.weightQ6_K[4096 14336]
-
blk.16.ffn_up.weightQ6_K[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightQ6_K[4096 1024]
-
blk.16.attn_output.weightQ6_K[4096 4096]
-
blk.16.attn_q.weightQ6_K[4096 4096]
-
blk.16.attn_v.weightQ6_K[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightQ6_K[14336 4096]
-
blk.17.ffn_gate.weightQ6_K[4096 14336]
-
blk.17.ffn_up.weightQ6_K[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightQ6_K[4096 1024]
-
blk.17.attn_output.weightQ6_K[4096 4096]
-
blk.17.attn_q.weightQ6_K[4096 4096]
-
blk.17.attn_v.weightQ6_K[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightQ6_K[14336 4096]
-
blk.18.ffn_gate.weightQ6_K[4096 14336]
-
blk.18.ffn_up.weightQ6_K[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightQ6_K[4096 1024]
-
blk.18.attn_output.weightQ6_K[4096 4096]
-
blk.18.attn_q.weightQ6_K[4096 4096]
-
blk.18.attn_v.weightQ6_K[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightQ6_K[14336 4096]
-
blk.19.ffn_gate.weightQ6_K[4096 14336]
-
blk.19.ffn_up.weightQ6_K[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightQ6_K[4096 1024]
-
blk.19.attn_output.weightQ6_K[4096 4096]
-
blk.19.attn_q.weightQ6_K[4096 4096]
-
blk.19.attn_v.weightQ6_K[4096 1024]
-
blk.20.ffn_gate.weightQ6_K[4096 14336]
-
blk.20.attn_k.weightQ6_K[4096 1024]
-
blk.20.attn_output.weightQ6_K[4096 4096]
-
blk.20.attn_q.weightQ6_K[4096 4096]
-
blk.20.attn_v.weightQ6_K[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightQ6_K[14336 4096]
-
blk.20.ffn_up.weightQ6_K[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightQ6_K[14336 4096]
-
blk.21.ffn_gate.weightQ6_K[4096 14336]
-
blk.21.ffn_up.weightQ6_K[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightQ6_K[4096 1024]
-
blk.21.attn_output.weightQ6_K[4096 4096]
-
blk.21.attn_q.weightQ6_K[4096 4096]
-
blk.21.attn_v.weightQ6_K[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightQ6_K[14336 4096]
-
blk.22.ffn_gate.weightQ6_K[4096 14336]
-
blk.22.ffn_up.weightQ6_K[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightQ6_K[4096 1024]
-
blk.22.attn_output.weightQ6_K[4096 4096]
-
blk.22.attn_q.weightQ6_K[4096 4096]
-
blk.22.attn_v.weightQ6_K[4096 1024]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightQ6_K[14336 4096]
-
blk.23.ffn_gate.weightQ6_K[4096 14336]
-
blk.23.ffn_up.weightQ6_K[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightQ6_K[4096 1024]
-
blk.23.attn_output.weightQ6_K[4096 4096]
-
blk.23.attn_q.weightQ6_K[4096 4096]
-
blk.23.attn_v.weightQ6_K[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightQ6_K[14336 4096]
-
blk.24.ffn_gate.weightQ6_K[4096 14336]
-
blk.24.ffn_up.weightQ6_K[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightQ6_K[4096 1024]
-
blk.24.attn_output.weightQ6_K[4096 4096]
-
blk.24.attn_q.weightQ6_K[4096 4096]
-
blk.24.attn_v.weightQ6_K[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightQ6_K[14336 4096]
-
blk.25.ffn_gate.weightQ6_K[4096 14336]
-
blk.25.ffn_up.weightQ6_K[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightQ6_K[4096 1024]
-
blk.25.attn_output.weightQ6_K[4096 4096]
-
blk.25.attn_q.weightQ6_K[4096 4096]
-
blk.25.attn_v.weightQ6_K[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightQ6_K[14336 4096]
-
blk.26.ffn_gate.weightQ6_K[4096 14336]
-
blk.26.ffn_up.weightQ6_K[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightQ6_K[4096 1024]
-
blk.26.attn_output.weightQ6_K[4096 4096]
-
blk.26.attn_q.weightQ6_K[4096 4096]
-
blk.26.attn_v.weightQ6_K[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightQ6_K[14336 4096]
-
blk.27.ffn_gate.weightQ6_K[4096 14336]
-
blk.27.ffn_up.weightQ6_K[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightQ6_K[4096 1024]
-
blk.27.attn_output.weightQ6_K[4096 4096]
-
blk.27.attn_q.weightQ6_K[4096 4096]
-
blk.27.attn_v.weightQ6_K[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightQ6_K[14336 4096]
-
blk.28.ffn_gate.weightQ6_K[4096 14336]
-
blk.28.ffn_up.weightQ6_K[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightQ6_K[4096 1024]
-
blk.28.attn_output.weightQ6_K[4096 4096]
-
blk.28.attn_q.weightQ6_K[4096 4096]
-
blk.28.attn_v.weightQ6_K[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightQ6_K[14336 4096]
-
blk.29.ffn_gate.weightQ6_K[4096 14336]
-
blk.29.ffn_up.weightQ6_K[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightQ6_K[4096 1024]
-
blk.29.attn_output.weightQ6_K[4096 4096]
-
blk.29.attn_q.weightQ6_K[4096 4096]
-
blk.29.attn_v.weightQ6_K[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightQ6_K[14336 4096]
-
blk.30.ffn_gate.weightQ6_K[4096 14336]
-
blk.30.ffn_up.weightQ6_K[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightQ6_K[4096 1024]
-
blk.30.attn_output.weightQ6_K[4096 4096]
-
blk.30.attn_q.weightQ6_K[4096 4096]
-
blk.30.attn_v.weightQ6_K[4096 1024]
-
blk.31.ffn_gate.weightQ6_K[4096 14336]
-
blk.31.ffn_up.weightQ6_K[4096 14336]
-
blk.31.attn_k.weightQ6_K[4096 1024]
-
blk.31.attn_output.weightQ6_K[4096 4096]
-
blk.31.attn_q.weightQ6_K[4096 4096]
-
blk.31.attn_v.weightQ6_K[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightQ6_K[14336 4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output.weightQ6_K[4096 128256]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31