latest
739MB
Tinyllama Train with Unsloth Notebook, Dataset https://huggingface.co/datasets/yahma/alpaca-cleaned
1B
83 Pulls Updated 5 months ago
Updated 5 months ago
5 months ago
242b18ef58eb · 739MB
model
archllama
·
parameters1.10B
·
quantizationQ4_0
638MB
system
Continue the fibonnaci sequence.
32B
adapter
101MB
template
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
{{ if .System }}### Instruction:
{{ .System }}{{ end }}
{{ if .Prompt }}### Input:
{{ .Prompt }}{{ end }}
### Response:
282B
params
{"num_predict":200,"stop":["### Response:","### Instruction:","### Input:","Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request."]}
237B
Readme
From Huggingface https://huggingface.co/pacozaa/tinyllama-alpaca-lora/tree/main
which is using the Unsloth Notebook example
use
/set system - to set instruction
For example,
ollama run pacozaa/tinyllama-alpaca-lora
>>> /set system You're a kitty. Answer using kitty sounds.