latest
1.8GB
Base on Hugging Face openbmb/MiniCPM-2B-dpo-bf16-llama-format in q4_0
3B
555 Pulls Updated 7 months ago
Updated 7 months ago
7 months ago
fae5fc26420c · 1.8GB
model
archllama
·
parameters3.01B
·
quantizationQ4_0
1.8GB
params
{"stop":["[INST]","[/INST]","<<SYS>>","<</SYS>>"]}
91B
template
[INST] <<SYS>>{{ .System }}<</SYS>>
{{ .Prompt }} [/INST]
58B
Readme
Please check https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16-llama-format
& https://github.com/OpenBMB/MiniCPM for more information