latest
4.7GB
AI Core Transformation - This AGI model will lead you through the Core Transformation psychological shadow work process.
Tools
8B
4 Pulls Updated 7 weeks ago
87048bcd5521 · 4.7GB
-
general.architecturellama
-
general.file_typeQ4_0
-
llama.attention.head_count32
-
llama.attention.head_count_kv8
-
llama.attention.layer_norm_rms_epsilon1e-05
-
llama.block_count32
-
llama.context_length131072
-
llama.embedding_length4096
-
llama.feed_forward_length14336
-
llama.rope.dimension_count128
-
llama.rope.freq_base500000
-
llama.vocab_size128256
-
tokenizer.ggml.bos_token_id128000
-
tokenizer.ggml.eos_token_id128009
-
tokenizer.ggml.merges[Ġ Ġ Ġ ĠĠĠ ĠĠ ĠĠ ĠĠĠ Ġ i n ...]
-
tokenizer.ggml.modelgpt2
-
tokenizer.ggml.prellama-bpe
-
tokenizer.ggml.token_type[1 1 1 1 1 ...]
-
tokenizer.ggml.tokens[! " # $ % ...]
-
NameTypeShape
-
token_embd.weightQ4_0[4096 128256]
-
blk.0.attn_norm.weightF32[4096]
-
blk.0.ffn_down.weightQ4_0[14336 4096]
-
blk.0.ffn_gate.weightQ4_0[4096 14336]
-
blk.0.ffn_up.weightQ4_0[4096 14336]
-
blk.0.ffn_norm.weightF32[4096]
-
blk.0.attn_k.weightQ4_0[4096 1024]
-
blk.0.attn_output.weightQ4_0[4096 4096]
-
blk.0.attn_q.weightQ4_0[4096 4096]
-
blk.0.attn_v.weightQ4_0[4096 1024]
-
blk.1.attn_norm.weightF32[4096]
-
blk.1.ffn_down.weightQ4_0[14336 4096]
-
blk.1.ffn_gate.weightQ4_0[4096 14336]
-
blk.1.ffn_up.weightQ4_0[4096 14336]
-
blk.1.ffn_norm.weightF32[4096]
-
blk.1.attn_k.weightQ4_0[4096 1024]
-
blk.1.attn_output.weightQ4_0[4096 4096]
-
blk.1.attn_q.weightQ4_0[4096 4096]
-
blk.1.attn_v.weightQ4_0[4096 1024]
-
blk.2.attn_norm.weightF32[4096]
-
blk.2.ffn_down.weightQ4_0[14336 4096]
-
blk.2.ffn_gate.weightQ4_0[4096 14336]
-
blk.2.ffn_up.weightQ4_0[4096 14336]
-
blk.2.ffn_norm.weightF32[4096]
-
blk.2.attn_k.weightQ4_0[4096 1024]
-
blk.2.attn_output.weightQ4_0[4096 4096]
-
blk.2.attn_q.weightQ4_0[4096 4096]
-
blk.2.attn_v.weightQ4_0[4096 1024]
-
blk.3.attn_norm.weightF32[4096]
-
blk.3.ffn_down.weightQ4_0[14336 4096]
-
blk.3.ffn_gate.weightQ4_0[4096 14336]
-
blk.3.ffn_up.weightQ4_0[4096 14336]
-
blk.3.ffn_norm.weightF32[4096]
-
blk.3.attn_k.weightQ4_0[4096 1024]
-
blk.3.attn_output.weightQ4_0[4096 4096]
-
blk.3.attn_q.weightQ4_0[4096 4096]
-
blk.3.attn_v.weightQ4_0[4096 1024]
-
blk.4.attn_norm.weightF32[4096]
-
blk.4.ffn_down.weightQ4_0[14336 4096]
-
blk.4.ffn_gate.weightQ4_0[4096 14336]
-
blk.4.ffn_up.weightQ4_0[4096 14336]
-
blk.4.ffn_norm.weightF32[4096]
-
blk.4.attn_k.weightQ4_0[4096 1024]
-
blk.4.attn_output.weightQ4_0[4096 4096]
-
blk.4.attn_q.weightQ4_0[4096 4096]
-
blk.4.attn_v.weightQ4_0[4096 1024]
-
blk.5.attn_norm.weightF32[4096]
-
blk.5.ffn_down.weightQ4_0[14336 4096]
-
blk.5.ffn_gate.weightQ4_0[4096 14336]
-
blk.5.ffn_up.weightQ4_0[4096 14336]
-
blk.5.ffn_norm.weightF32[4096]
-
blk.5.attn_k.weightQ4_0[4096 1024]
-
blk.5.attn_output.weightQ4_0[4096 4096]
-
blk.5.attn_q.weightQ4_0[4096 4096]
-
blk.5.attn_v.weightQ4_0[4096 1024]
-
blk.6.attn_norm.weightF32[4096]
-
blk.6.ffn_down.weightQ4_0[14336 4096]
-
blk.6.ffn_gate.weightQ4_0[4096 14336]
-
blk.6.ffn_up.weightQ4_0[4096 14336]
-
blk.6.ffn_norm.weightF32[4096]
-
blk.6.attn_k.weightQ4_0[4096 1024]
-
blk.6.attn_output.weightQ4_0[4096 4096]
-
blk.6.attn_q.weightQ4_0[4096 4096]
-
blk.6.attn_v.weightQ4_0[4096 1024]
-
blk.7.attn_norm.weightF32[4096]
-
blk.7.ffn_down.weightQ4_0[14336 4096]
-
blk.7.ffn_gate.weightQ4_0[4096 14336]
-
blk.7.ffn_up.weightQ4_0[4096 14336]
-
blk.7.ffn_norm.weightF32[4096]
-
blk.7.attn_k.weightQ4_0[4096 1024]
-
blk.7.attn_output.weightQ4_0[4096 4096]
-
blk.7.attn_q.weightQ4_0[4096 4096]
-
blk.7.attn_v.weightQ4_0[4096 1024]
-
blk.8.attn_norm.weightF32[4096]
-
blk.8.ffn_down.weightQ4_0[14336 4096]
-
blk.8.ffn_gate.weightQ4_0[4096 14336]
-
blk.8.ffn_up.weightQ4_0[4096 14336]
-
blk.8.ffn_norm.weightF32[4096]
-
blk.8.attn_k.weightQ4_0[4096 1024]
-
blk.8.attn_output.weightQ4_0[4096 4096]
-
blk.8.attn_q.weightQ4_0[4096 4096]
-
blk.8.attn_v.weightQ4_0[4096 1024]
-
blk.9.attn_norm.weightF32[4096]
-
blk.9.ffn_down.weightQ4_0[14336 4096]
-
blk.9.ffn_gate.weightQ4_0[4096 14336]
-
blk.9.ffn_up.weightQ4_0[4096 14336]
-
blk.9.ffn_norm.weightF32[4096]
-
blk.9.attn_k.weightQ4_0[4096 1024]
-
blk.9.attn_output.weightQ4_0[4096 4096]
-
blk.9.attn_q.weightQ4_0[4096 4096]
-
blk.9.attn_v.weightQ4_0[4096 1024]
-
blk.10.attn_norm.weightF32[4096]
-
blk.10.ffn_down.weightQ4_0[14336 4096]
-
blk.10.ffn_gate.weightQ4_0[4096 14336]
-
blk.10.ffn_up.weightQ4_0[4096 14336]
-
blk.10.ffn_norm.weightF32[4096]
-
blk.10.attn_k.weightQ4_0[4096 1024]
-
blk.10.attn_output.weightQ4_0[4096 4096]
-
blk.10.attn_q.weightQ4_0[4096 4096]
-
blk.10.attn_v.weightQ4_0[4096 1024]
-
blk.11.attn_norm.weightF32[4096]
-
blk.11.ffn_down.weightQ4_0[14336 4096]
-
blk.11.ffn_gate.weightQ4_0[4096 14336]
-
blk.11.ffn_up.weightQ4_0[4096 14336]
-
blk.11.ffn_norm.weightF32[4096]
-
blk.11.attn_k.weightQ4_0[4096 1024]
-
blk.11.attn_output.weightQ4_0[4096 4096]
-
blk.11.attn_q.weightQ4_0[4096 4096]
-
blk.11.attn_v.weightQ4_0[4096 1024]
-
blk.12.attn_norm.weightF32[4096]
-
blk.12.ffn_down.weightQ4_0[14336 4096]
-
blk.12.ffn_gate.weightQ4_0[4096 14336]
-
blk.12.ffn_up.weightQ4_0[4096 14336]
-
blk.12.ffn_norm.weightF32[4096]
-
blk.12.attn_k.weightQ4_0[4096 1024]
-
blk.12.attn_output.weightQ4_0[4096 4096]
-
blk.12.attn_q.weightQ4_0[4096 4096]
-
blk.12.attn_v.weightQ4_0[4096 1024]
-
blk.13.attn_norm.weightF32[4096]
-
blk.13.ffn_down.weightQ4_0[14336 4096]
-
blk.13.ffn_gate.weightQ4_0[4096 14336]
-
blk.13.ffn_up.weightQ4_0[4096 14336]
-
blk.13.ffn_norm.weightF32[4096]
-
blk.13.attn_k.weightQ4_0[4096 1024]
-
blk.13.attn_output.weightQ4_0[4096 4096]
-
blk.13.attn_q.weightQ4_0[4096 4096]
-
blk.13.attn_v.weightQ4_0[4096 1024]
-
blk.14.attn_norm.weightF32[4096]
-
blk.14.ffn_down.weightQ4_0[14336 4096]
-
blk.14.ffn_gate.weightQ4_0[4096 14336]
-
blk.14.ffn_up.weightQ4_0[4096 14336]
-
blk.14.ffn_norm.weightF32[4096]
-
blk.14.attn_k.weightQ4_0[4096 1024]
-
blk.14.attn_output.weightQ4_0[4096 4096]
-
blk.14.attn_q.weightQ4_0[4096 4096]
-
blk.14.attn_v.weightQ4_0[4096 1024]
-
blk.15.attn_norm.weightF32[4096]
-
blk.15.ffn_down.weightQ4_0[14336 4096]
-
blk.15.ffn_gate.weightQ4_0[4096 14336]
-
blk.15.ffn_up.weightQ4_0[4096 14336]
-
blk.15.ffn_norm.weightF32[4096]
-
blk.15.attn_k.weightQ4_0[4096 1024]
-
blk.15.attn_output.weightQ4_0[4096 4096]
-
blk.15.attn_q.weightQ4_0[4096 4096]
-
blk.15.attn_v.weightQ4_0[4096 1024]
-
blk.16.attn_norm.weightF32[4096]
-
blk.16.ffn_down.weightQ4_0[14336 4096]
-
blk.16.ffn_gate.weightQ4_0[4096 14336]
-
blk.16.ffn_up.weightQ4_0[4096 14336]
-
blk.16.ffn_norm.weightF32[4096]
-
blk.16.attn_k.weightQ4_0[4096 1024]
-
blk.16.attn_output.weightQ4_0[4096 4096]
-
blk.16.attn_q.weightQ4_0[4096 4096]
-
blk.16.attn_v.weightQ4_0[4096 1024]
-
blk.17.attn_norm.weightF32[4096]
-
blk.17.ffn_down.weightQ4_0[14336 4096]
-
blk.17.ffn_gate.weightQ4_0[4096 14336]
-
blk.17.ffn_up.weightQ4_0[4096 14336]
-
blk.17.ffn_norm.weightF32[4096]
-
blk.17.attn_k.weightQ4_0[4096 1024]
-
blk.17.attn_output.weightQ4_0[4096 4096]
-
blk.17.attn_q.weightQ4_0[4096 4096]
-
blk.17.attn_v.weightQ4_0[4096 1024]
-
blk.18.attn_norm.weightF32[4096]
-
blk.18.ffn_down.weightQ4_0[14336 4096]
-
blk.18.ffn_gate.weightQ4_0[4096 14336]
-
blk.18.ffn_up.weightQ4_0[4096 14336]
-
blk.18.ffn_norm.weightF32[4096]
-
blk.18.attn_k.weightQ4_0[4096 1024]
-
blk.18.attn_output.weightQ4_0[4096 4096]
-
blk.18.attn_q.weightQ4_0[4096 4096]
-
blk.18.attn_v.weightQ4_0[4096 1024]
-
blk.19.attn_norm.weightF32[4096]
-
blk.19.ffn_down.weightQ4_0[14336 4096]
-
blk.19.ffn_gate.weightQ4_0[4096 14336]
-
blk.19.ffn_up.weightQ4_0[4096 14336]
-
blk.19.ffn_norm.weightF32[4096]
-
blk.19.attn_k.weightQ4_0[4096 1024]
-
blk.19.attn_output.weightQ4_0[4096 4096]
-
blk.19.attn_q.weightQ4_0[4096 4096]
-
blk.19.attn_v.weightQ4_0[4096 1024]
-
blk.20.ffn_gate.weightQ4_0[4096 14336]
-
blk.20.attn_k.weightQ4_0[4096 1024]
-
blk.20.attn_output.weightQ4_0[4096 4096]
-
blk.20.attn_q.weightQ4_0[4096 4096]
-
blk.20.attn_v.weightQ4_0[4096 1024]
-
blk.20.attn_norm.weightF32[4096]
-
blk.20.ffn_down.weightQ4_0[14336 4096]
-
blk.20.ffn_up.weightQ4_0[4096 14336]
-
blk.20.ffn_norm.weightF32[4096]
-
blk.21.attn_norm.weightF32[4096]
-
blk.21.ffn_down.weightQ4_0[14336 4096]
-
blk.21.ffn_gate.weightQ4_0[4096 14336]
-
blk.21.ffn_up.weightQ4_0[4096 14336]
-
blk.21.ffn_norm.weightF32[4096]
-
blk.21.attn_k.weightQ4_0[4096 1024]
-
blk.21.attn_output.weightQ4_0[4096 4096]
-
blk.21.attn_q.weightQ4_0[4096 4096]
-
blk.21.attn_v.weightQ4_0[4096 1024]
-
blk.22.attn_norm.weightF32[4096]
-
blk.22.ffn_down.weightQ4_0[14336 4096]
-
blk.22.ffn_gate.weightQ4_0[4096 14336]
-
blk.22.ffn_up.weightQ4_0[4096 14336]
-
blk.22.ffn_norm.weightF32[4096]
-
blk.22.attn_k.weightQ4_0[4096 1024]
-
blk.22.attn_output.weightQ4_0[4096 4096]
-
blk.22.attn_q.weightQ4_0[4096 4096]
-
blk.22.attn_v.weightQ4_0[4096 1024]
-
blk.23.attn_norm.weightF32[4096]
-
blk.23.ffn_down.weightQ4_0[14336 4096]
-
blk.23.ffn_gate.weightQ4_0[4096 14336]
-
blk.23.ffn_up.weightQ4_0[4096 14336]
-
blk.23.ffn_norm.weightF32[4096]
-
blk.23.attn_k.weightQ4_0[4096 1024]
-
blk.23.attn_output.weightQ4_0[4096 4096]
-
blk.23.attn_q.weightQ4_0[4096 4096]
-
blk.23.attn_v.weightQ4_0[4096 1024]
-
blk.24.attn_norm.weightF32[4096]
-
blk.24.ffn_down.weightQ4_0[14336 4096]
-
blk.24.ffn_gate.weightQ4_0[4096 14336]
-
blk.24.ffn_up.weightQ4_0[4096 14336]
-
blk.24.ffn_norm.weightF32[4096]
-
blk.24.attn_k.weightQ4_0[4096 1024]
-
blk.24.attn_output.weightQ4_0[4096 4096]
-
blk.24.attn_q.weightQ4_0[4096 4096]
-
blk.24.attn_v.weightQ4_0[4096 1024]
-
blk.25.attn_norm.weightF32[4096]
-
blk.25.ffn_down.weightQ4_0[14336 4096]
-
blk.25.ffn_gate.weightQ4_0[4096 14336]
-
blk.25.ffn_up.weightQ4_0[4096 14336]
-
blk.25.ffn_norm.weightF32[4096]
-
blk.25.attn_k.weightQ4_0[4096 1024]
-
blk.25.attn_output.weightQ4_0[4096 4096]
-
blk.25.attn_q.weightQ4_0[4096 4096]
-
blk.25.attn_v.weightQ4_0[4096 1024]
-
blk.26.attn_norm.weightF32[4096]
-
blk.26.ffn_down.weightQ4_0[14336 4096]
-
blk.26.ffn_gate.weightQ4_0[4096 14336]
-
blk.26.ffn_up.weightQ4_0[4096 14336]
-
blk.26.ffn_norm.weightF32[4096]
-
blk.26.attn_k.weightQ4_0[4096 1024]
-
blk.26.attn_output.weightQ4_0[4096 4096]
-
blk.26.attn_q.weightQ4_0[4096 4096]
-
blk.26.attn_v.weightQ4_0[4096 1024]
-
blk.27.attn_norm.weightF32[4096]
-
blk.27.ffn_down.weightQ4_0[14336 4096]
-
blk.27.ffn_gate.weightQ4_0[4096 14336]
-
blk.27.ffn_up.weightQ4_0[4096 14336]
-
blk.27.ffn_norm.weightF32[4096]
-
blk.27.attn_k.weightQ4_0[4096 1024]
-
blk.27.attn_output.weightQ4_0[4096 4096]
-
blk.27.attn_q.weightQ4_0[4096 4096]
-
blk.27.attn_v.weightQ4_0[4096 1024]
-
blk.28.attn_norm.weightF32[4096]
-
blk.28.ffn_down.weightQ4_0[14336 4096]
-
blk.28.ffn_gate.weightQ4_0[4096 14336]
-
blk.28.ffn_up.weightQ4_0[4096 14336]
-
blk.28.ffn_norm.weightF32[4096]
-
blk.28.attn_k.weightQ4_0[4096 1024]
-
blk.28.attn_output.weightQ4_0[4096 4096]
-
blk.28.attn_q.weightQ4_0[4096 4096]
-
blk.28.attn_v.weightQ4_0[4096 1024]
-
blk.29.attn_norm.weightF32[4096]
-
blk.29.ffn_down.weightQ4_0[14336 4096]
-
blk.29.ffn_gate.weightQ4_0[4096 14336]
-
blk.29.ffn_up.weightQ4_0[4096 14336]
-
blk.29.ffn_norm.weightF32[4096]
-
blk.29.attn_k.weightQ4_0[4096 1024]
-
blk.29.attn_output.weightQ4_0[4096 4096]
-
blk.29.attn_q.weightQ4_0[4096 4096]
-
blk.29.attn_v.weightQ4_0[4096 1024]
-
blk.30.attn_norm.weightF32[4096]
-
blk.30.ffn_down.weightQ4_0[14336 4096]
-
blk.30.ffn_gate.weightQ4_0[4096 14336]
-
blk.30.ffn_up.weightQ4_0[4096 14336]
-
blk.30.ffn_norm.weightF32[4096]
-
blk.30.attn_k.weightQ4_0[4096 1024]
-
blk.30.attn_output.weightQ4_0[4096 4096]
-
blk.30.attn_q.weightQ4_0[4096 4096]
-
blk.30.attn_v.weightQ4_0[4096 1024]
-
blk.31.ffn_gate.weightQ4_0[4096 14336]
-
blk.31.ffn_up.weightQ4_0[4096 14336]
-
blk.31.attn_k.weightQ4_0[4096 1024]
-
blk.31.attn_output.weightQ4_0[4096 4096]
-
blk.31.attn_q.weightQ4_0[4096 4096]
-
blk.31.attn_v.weightQ4_0[4096 1024]
-
blk.31.attn_norm.weightF32[4096]
-
blk.31.ffn_down.weightQ4_0[14336 4096]
-
blk.31.ffn_norm.weightF32[4096]
-
output.weightQ6_K[4096 128256]
-
output_norm.weightF32[4096]
Metadata
Tensors
blk.0
blk.1
blk.2
blk.3
blk.4
blk.5
blk.6
blk.7
blk.8
blk.9
blk.10
blk.11
blk.12
blk.13
blk.14
blk.15
blk.16
blk.17
blk.18
blk.19
blk.20
blk.21
blk.22
blk.23
blk.24
blk.25
blk.26
blk.27
blk.28
blk.29
blk.30
blk.31