ik_llama.cpp
quantizations of DeepSeek-R1-0528
Quantized using ik_llama.cpp
build = 3788 (4622fadc)
NOTE: These quants MUST be run using the llama.cpp
fork, ik_llama.cpp
Credits to @ubergarm for his DeepSeek quant recipes for which these quants were based on.
name | file size | quant type | bpw |
---|---|---|---|
DeepSeek-R1-0528-IQ4_KT | 322.355 GiB | IQ4_KT (97.5%) / Q8_0 (2.5%) |
4.127 |
- Downloads last month
- 15
Hardware compatibility
Log In
to view the estimation
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for tachyphylaxis/DeepSeek-R1-0528-IK_GGUF
Base model
deepseek-ai/DeepSeek-R1-0528