ik_llama.cpp quantizations of DeepSeek-R1-0528

Quantized using ik_llama.cpp build = 3788 (4622fadc)

NOTE: These quants MUST be run using the llama.cpp fork, ik_llama.cpp

Credits to @ubergarm for his DeepSeek quant recipes for which these quants were based on.

name file size quant type bpw
DeepSeek-R1-0528-IQ4_KT 322.355 GiB IQ4_KT (97.5%) / Q8_0 (2.5%) 4.127
Downloads last month
15
GGUF
Model size
671B params
Architecture
deepseek2
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for tachyphylaxis/DeepSeek-R1-0528-IK_GGUF

Quantized
(36)
this model