latest
5.1GB
quantized DRXD1000/Phoenix - which was trained with german dpo (ultrachat_200k & ultrafeedback_binarized transl. by haoranxu/ALMA-13B) based on LeoLM/leo-mistral-hessianai-7b
7B
294 Pulls Updated 4 months ago
Updated 6 months ago
6 months ago
b4f17b74652b · 5.1GB
model
archllama
·
parameters7.24B
·
quantizationQ5_K_M
5.1GB
template
{{- if .System }}<|system|>
{{ .System }}</s>{{- end }}
<|user|>
{{ .Prompt }}</s>
<|assistant|>
{{ .Response }}
112B
params
{"stop":["<|assistant|>","</s>","<|system|>","<|user|>"]}
98B
Readme
No readme