latest
7.7GB
Experimental model doing a DPO training on top of Kunoichi-DPO-v2-7b, i.e. double-DPO.
7B
152 Pulls Updated 3 months ago
Updated 3 months ago
3 months ago
4742d588810c · 7.7GB
model
archllama
·
parameters7.24B
·
quantizationQ8_0
7.7GB
template
{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>
254B
params
{"stop":["<|start_header_id|>","<|end_header_id|>","<|eot_id|>","<|reserved_special_token"],"temperature":0.6}
146B
system
You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.
128B
Readme
Source: https://huggingface.co/crestf411/daybreak-kunoichi-2dpo-7b
Experimental model doing a DPO training on top of Kunoichi-DPO-v2-7b, i.e. double-DPO.