-
deepseek-coder
DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.
Code 1B 7B 33B289.9K Pulls 102 Tags Updated 8 months ago
-
mxbai-embed-large
State-of-the-art large embedding model from mixedbread.ai
Embedding288.4K Pulls 4 Tags Updated 5 months ago
-
starcoder2
StarCoder2 is the next generation of transparently trained open code LLMs that comes in three sizes: 3B, 7B and 15B parameters.
Code 3B 7B217.4K Pulls 67 Tags Updated 2 weeks ago
-
dolphin-mistral
The uncensored Dolphin model based on Mistral that excels at coding tasks. Updated to version 2.8.
7B217.1K Pulls 120 Tags Updated 5 months ago
-
zephyr
Zephyr is a series of fine-tuned versions of the Mistral and Mixtral models that are trained to act as helpful assistants.
7B 8x22B209.3K Pulls 40 Tags Updated 5 months ago
-
dolphin-llama3
Dolphin 2.9 is a new model with 8B and 70B sizes by Eric Hartford based on Llama 3 that has a variety of instruction, conversational, and coding skills.
8B 70B200.5K Pulls 53 Tags Updated 2 weeks ago
-
orca-mini
A general-purpose model ranging from 3 billion parameters to 70 billion, suitable for entry-level hardware.
3B 7B 13B197.8K Pulls 119 Tags Updated 10 months ago
-
yi
Yi 1.5 is a high-performing, bilingual language model.
6B 9B 34B177.1K Pulls 174 Tags Updated 4 months ago
-
llava-llama3
A LLaVA model fine-tuned from Llama 3 Instruct with better scores in several benchmarks.
Vision 8B156.2K Pulls 4 Tags Updated 4 months ago
-
mistral-openorca
Mistral OpenOrca is a 7 billion parameter model, fine-tuned on top of the Mistral 7B model using the OpenOrca dataset.
7B151.7K Pulls 17 Tags Updated 11 months ago
-
starcoder
StarCoder is a code generation model trained on 80+ programming languages.
Code 1B 3B 7B 15B137K Pulls 100 Tags Updated 11 months ago
-
tinyllama
The TinyLlama project is an open endeavor to train a compact 1.1B Llama model on 3 trillion tokens.
1B133.6K Pulls 36 Tags Updated 8 months ago
-
vicuna
General use chat model based on Llama and Llama 2 with 2K to 16K context sizes.
7B 13B 30B129.5K Pulls 111 Tags Updated 10 months ago
-
codestral
Codestral is Mistral AI’s first-ever code model designed for code generation tasks.
Code 22B128.3K Pulls 17 Tags Updated 2 weeks ago
-
llama2-chinese
Llama 2 based model fine tuned to improve Chinese dialogue ability.
7B 13B125.9K Pulls 35 Tags Updated 11 months ago
-
wizard-vicuna-uncensored
Wizard Vicuna Uncensored is a 7B, 13B, and 30B parameter model based on Llama 2 uncensored by Eric Hartford.
7B 13B 30B116.4K Pulls 49 Tags Updated 10 months ago
-
codegeex4
A versatile model for AI software development scenarios, including code completion.
Code 9B113K Pulls 17 Tags Updated 2 months ago
-
nous-hermes2
The powerful family of models by Nous Research that excels at scientific discussion and coding tasks.
34B106K Pulls 33 Tags Updated 8 months ago
-
granite-code
A family of open foundation models by IBM for Code Intelligence
Code 3B 8B 20B 34B102.1K Pulls 162 Tags Updated 2 weeks ago
-
openchat
A family of open-source models trained on a wide variety of data, surpassing ChatGPT on various benchmarks. Updated to version 3.5-0106.
7B98.9K Pulls 50 Tags Updated 8 months ago
-
aya
Aya 23, released by Cohere, is a new family of state-of-the-art, multilingual models that support 23 languages.
8B 35B98.2K Pulls 33 Tags Updated 2 weeks ago
-
wizardlm2
State of the art large language model from Microsoft AI with improved performance on complex chat, multilingual, reasoning and agent use cases.
7B 8x22B94K Pulls 22 Tags Updated 5 months ago
-
tinydolphin
An experimental 1.1B parameter model trained on the new Dolphin 2.8 dataset by Eric Hartford and based on TinyLlama.
1B90.5K Pulls 18 Tags Updated 8 months ago
-
codeqwen
CodeQwen1.5 is a large language model pretrained on a large amount of code data.
Code 7B90.4K Pulls 30 Tags Updated 5 months ago
-
all-minilm
Embedding models on very large sentence level datasets.
Embedding 22M 33M89.8K Pulls 10 Tags Updated 7 months ago