https://huggingface.co/wolfram/miquliz-120b-v2.0-GGUF
101 Pulls Updated 6 months ago
Updated 6 months ago
6 months ago
d252392a0348 · 44GB
model
archllama
·
parameters120B
·
quantizationQ2_K
44GB
params
{"stop":["<|im_start|>","<|im_end|>"]}
59B
template
{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
155B