Embedding
8B
115 Pulls Updated 8 weeks ago
embedding-bge-m3-f16
1.1GB
embedding-bge-m3-f16-bert-cpp
1.1GB
embedding-bge-m3-q8-0
618MB
embedding-bge-m3-q8-0-bert-cpp
616MB
llama3_alpha_ko_instruct_fp16
16GB
llama3_alpha_ko_instruct_q2
3.2GB
llama3_alpha_ko_instruct_q3
4.0GB
llama3_alpha_ko_instruct_q4
4.9GB
llama3_open_ko_instruct_fp16
16GB
llama3_open_ko_instruct_q2
3.2GB
llama3_open_ko_instruct_q3
4.0GB
llama3_open_ko_instruct_q4
4.9GB
8b-llama3-komultigen-general-fp16-v1
16GB
8b-llama3-komultigen-general-q3-v1
4.0GB
8b-llama3-komultigen-general-q4-v1
4.9GB
8b-llama3-komultigen-general-q8-v1
8.5GB
10.8b-korean-eeve-fp16-v1
22GB
10.8b-korean-eeve-q3-v1
5.2GB
10.8b-korean-eeve-q4-v1
6.5GB
10.8b-korean-eeve-q8-v1
11GB
10.8b-korean-instruct-eeve-fp16-v1
22GB
10.8b-korean-instruct-eeve-q3-v1
5.2GB
10.8b-korean-instruct-eeve-q4-v1
6.5GB
10.8b-korean-instruct-eeve-q8-v1
11GB
10.8b-korean-instruct-it119-fp16-v1
22GB
10.8b-korean-instruct-it119-q3-v1
5.2GB
10.8b-korean-instruct-it119-q4-v1
6.5GB
10.8b-korean-instruct-it119-q8-v1
11GB
10.8b-korean-instruct-reg-fp16-v1
22GB
10.8b-korean-instruct-reg-fp16-v2
22GB
10.8b-korean-instruct-reg-q3-v1
5.2GB
10.8b-korean-instruct-reg-q3-v2
5.2GB
10.8b-korean-instruct-reg-q4-v1
6.5GB
10.8b-korean-instruct-reg-q4-v2
6.5GB
10.8b-korean-instruct-reg-q8-v1
11GB
10.8b-korean-instruct-reg-q8-v2
11GB
36952e7fe300 · 120B
{
"stop": [
"<|begin_of_text|>",
"<|end_of_text|>",
"<|eot_id|>",
"<|end_of_text|>"
]
}