latest
7.3GB
quantization of perlthoughts/Mistral-7B-Instruct-v0.2-2x7B-MoE
13B
254 Pulls Updated 4 months ago
Updated 7 months ago
7 months ago
ad95f022607a · 7.3GB
model
archllama
·
parameters12.9B
·
quantizationQ4_K_M
7.3GB
template
[INST] {{ .System }} {{ .Prompt }} [/INST]
42B
params
{"num_ctx":32000,"stop":["[INST]","[/INST]"]}
46B
Readme
No readme