This model was converted using the OpenVINO Model Server export_model.py script and contains metadata for running pipeline paralell on two GPUs.

Learn more about pipeline paralell with OVMS here

Downloads last month
6
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Echo9Zulu/DeepSeek-R1-Distill-Qwen-32B_INT4-OVMS

Finetuned
(85)
this model

Collection including Echo9Zulu/DeepSeek-R1-Distill-Qwen-32B_INT4-OVMS