The model used is a quantized version of `Llama-3-Taiwan-8B-Instruct-DPO`. More details can be found on the website (https://huggingface.co/yentinglin/Llama-3-Taiwan-8B-Instruct-DPO)
8B
41 Pulls Updated 2 months ago
Updated 2 months ago
2 months ago
0e79d6c5222f · 3.7GB
model
archllama
·
parameters8.03B
·
quantizationQ3_K_S
3.7GB
template
"{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>""
257B
params
{"num_ctx":8192,"stop":["<|start_header_id|>","<|end_header_id|>","<|end_of_text|>","<|eot_id|>","<|reserved_special_token"]}
171B
Readme
No readme