latest
4.4GB
from mlabonne/NeuralMonarch-7B-GGUF
7B
49 Pulls Updated 7 months ago
Updated 7 months ago
7 months ago
d9d31d179466 · 4.4GB
model
archllama
·
parameters7.24B
·
quantizationQ4_K_M
4.4GB
params
{"num_ctx":8192,"stop":["[INST]","[/INST]"]}
45B
template
[INST] {{ .System }} {{ .Prompt }} [/INST]
43B
Readme
No readme