base_model: - Wan-AI/Wan2.1-VACE-14B - Wan-AI/Wan2.1-VACE-1.3B
Better fp8 scaled models (when measured against fp16) based on quantization code from https://github.com/Tencent-Hunyuan/HunyuanVideo/blob/main/hyvideo/modules/fp8_optimization.py
Can be used with: https://github.com/kijai/ComfyUI-WanVideoWrapper (latest version) and ComfyUI native WanVideo nodes.
14B-T2V comparison test without LoRAs, 25 steps, 832x480x81
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support