16B
63 Pulls Updated 7 weeks ago
Updated 7 weeks ago
7 weeks ago
f1a7478de823 · 9.5GB
model
archdeepseek2
·
parameters15.7B
·
quantizationQ4_K_S
9.5GB
template
{{ if .System }}System: {{ .System }}
{{ end }}{{ if .Prompt }}User: {{ .Prompt }}
{{ end }}Assistant: {{ .Response }}
122B
params
{"stop":["System:","User:","Assistant:","<|begin_of_text|>"]}
72B
Readme
Models:
DeepSeek Coder V2 Instruct Lite Quantized with IQ4_XS (https://ollama.com/akuldatta/deepseek-coder-v2-lite) and Q5_K_S (https://ollama.com/akuldatta/deepseek-coder-v2-lite:q5ks) quants.