This model was converted using the OpenVINO Model Server export_model.py script and contains metadata for running pipeline paralell on two GPUs.

Learn more about pipeline paralell with OVMS here

Downloads last month
10
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Echo9Zulu/DeepSeek-R1-Distill-Qwen-32B_INT4-OVMS

Finetuned
(85)
this model

Collection including Echo9Zulu/DeepSeek-R1-Distill-Qwen-32B_INT4-OVMS