Q4_K_M
13 GBMin VRAM: 15 GB
Recommended VRAM: 20 GB
Min RAM: 20 GB
Context: 8K / 32K
Loading model details...
Fetching variants, compatibility details, and metadata.
New to local models? Smaller quantization variants are easier to run, while larger ones can improve quality at the cost of more memory.
Q4_K_M
13 GBMin VRAM: 15 GB
Recommended VRAM: 20 GB
Min RAM: 20 GB
Context: 8K / 32K
Q8_0
23 GBMin VRAM: 25 GB
Recommended VRAM: 28 GB
Min RAM: 28 GB
Context: 8K / 32K
| Quantization | File Size | Min VRAM | Recommended VRAM | Min RAM | Context |
|---|---|---|---|---|---|
| Q4_K_M | 13 GB | 15 GB | 20 GB | 20 GB | 8K / 32K |
| Q8_0 | 23 GB | 25 GB | 28 GB | 28 GB | 8K / 32K |