|
|
--- |
|
|
tags: |
|
|
- gguf |
|
|
- llama.cpp |
|
|
- unsloth |
|
|
|
|
|
--- |
|
|
|
|
|
# llamabotomy-test - GGUF |
|
|
|
|
|
This model was finetuned and converted to GGUF format using [Unsloth](https://github.com/unslothai/unsloth). |
|
|
|
|
|
Super tiny version of Llama's 1 B parameter model quantized using the lowest precision Unsloth offers. Training this one on junk data and destroying the weights should fully lobotomize it, but it honestly works a little too well for being around ~500MB. |
|
|
Shoutout Unsloth's quantization magic I guess... |
|
|
|
|
|
## Available Model files: |
|
|
- `llama-3.2-1b-instruct.Q3_K_S.gguf` |
|
|
|
|
|
## Ollama |
|
|
An Ollama Modelfile is included for easy deployment. |
|
|
|