latest
8.8GB
🤗 w4r10ck/SOLAR-10.7B-Instruct-v1.0-uncensored
958 Pulls Updated 5 months ago
Updated 5 months ago
5 months ago
125db36e66cb · 8.8GB
model
archllama
·
parameters10.7B
·
quantizationQ6_K
8.8GB
template
### System:
{{ .System }}
### User:
{{ .Prompt }}
### Assistant:
67B
params
{"num_ctx":4096,"stop":["</s8>","### System:","### User:","### Assistant:"]}
87B
Readme
Solar 10.7B uncensored
Original model
SOLAR-10.7B is an advanced large language model (LLM) with 10.7 billion parameters, demonstrating superior performance in various natural language processing (NLP) tasks. It’s compact, yet remarkably powerful, and demonstrates unparalleled state-of-the-art performance in models with parameters under 30B.
Uncensored training model
Trained using Toxic-DPO v0.1 to uncensor the model.
GGUF
All credits to the authors and contributors above.