Update README.md
Browse files
README.md
CHANGED
|
@@ -219,10 +219,10 @@ These builds are compatible with **llama.cpp**, **Jan**, **LM Studio**, **Ollama
|
|
| 219 |
|
| 220 |
| **Quantization** | **Size** | **Bit Precision** | **Description** |
|
| 221 |
|------------------|-----------|------------------|----------------|
|
| 222 |
-
| **Q4_K_M** | 8.99 GB | **4-bit** | Ultra-fast, compact variant for consumer GPUs and laptops |
|
| 223 |
-
| **Q5_K_M** | 10.5 GB | **5-bit** | Lightweight deployment with strong accuracy retention |
|
| 224 |
-
| **Q6_K** | 12.1 GB | **6-bit** | Balanced performance and efficiency |
|
| 225 |
| **Q8_0** | 15.7 GB | **8-bit** | Near-full precision, for most demanding local inference |
|
|
|
|
|
|
|
|
|
|
| 226 |
|
| 227 |
Quant versions: [Fortytwo-Network/Strand-Rust-Coder-14B-v1-GGUF](https://huggingface.co/Fortytwo-Network/Strand-Rust-Coder-14B-v1-GGUF)
|
| 228 |
|
|
|
|
| 219 |
|
| 220 |
| **Quantization** | **Size** | **Bit Precision** | **Description** |
|
| 221 |
|------------------|-----------|------------------|----------------|
|
|
|
|
|
|
|
|
|
|
| 222 |
| **Q8_0** | 15.7 GB | **8-bit** | Near-full precision, for most demanding local inference |
|
| 223 |
+
| **Q6_K** | 12.1 GB | **6-bit** | Balanced performance and efficiency |
|
| 224 |
+
| **Q5_K_M** | 10.5 GB | **5-bit** | Lightweight deployment with strong accuracy retention |
|
| 225 |
+
| **Q4_K_M** | 8.99 GB | **4-bit** | Ultra-fast, compact variant for consumer GPUs and laptops |
|
| 226 |
|
| 227 |
Quant versions: [Fortytwo-Network/Strand-Rust-Coder-14B-v1-GGUF](https://huggingface.co/Fortytwo-Network/Strand-Rust-Coder-14B-v1-GGUF)
|
| 228 |
|