16B
63 Pulls Updated 7 weeks ago
Updated 2 months ago
2 months ago
d333eb998ea0 · 11GB
model
archdeepseek2
·
parameters15.7B
·
quantizationQ5_K_S
11GB
template
{{ if .System }}System: {{ .System }}
{{ end }}{{ if .Prompt }}User: {{ .Prompt }}
{{ end }}Assistant: <|begin_of_text|>{{ .Response }}
137B
Readme
Models:
DeepSeek Coder V2 Instruct Lite Quantized with IQ4_XS (https://ollama.com/akuldatta/deepseek-coder-v2-lite) and Q5_K_S (https://ollama.com/akuldatta/deepseek-coder-v2-lite:q5ks) quants.