latest
16GB
minicpm-llama3-2.5-8b-16-v With only 8B parameters, it surpasses widely used proprietary models like GPT-4V-1106, Gemini Pro, Claude 3 and Qwen-VL-Max and greatly outperforms other Llama 3-based MLLMs
8B
351 Pulls Updated 3 months ago
Updated 3 months ago
3 months ago
4afe8c21f45c · 16GB
model
archllama
·
parameters8.03B
·
quantizationF16
16GB
params
{"num_gpu":12,"num_keep":32,"num_predict":-2,"stop":["[<|eot_id|>"],"top_k":33,"top_p":0.66}
103B
template
{{ if .System }}<|start_header_id|>system<|end_header_id|>{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>{{ .Response }}<|eot_id|>
248B
Readme
No readme