Text-to-Image
Diffusers
English
SVDQuant
SANA
Diffusion
Quantization
ICLR2025

This repository has been migrated to https://huggingface.co/nunchaku-tech/nunchaku-sana and will be hidden in December 2025.

Nunchaku Logo

Model Card for nunchaku-sana

visual This repository contains Nunchaku-quantized versions of SANA-1.6B, designed to generate high-quality images from text prompts. It is optimized for efficient inference while maintaining minimal loss in performance.

Model Details

Model Description

Model Files

Model Sources

Usage

See sana1.6b.py.

Performance

performance

Citation

@inproceedings{
  li2024svdquant,
  title={SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models},
  author={Li*, Muyang and Lin*, Yujun and Zhang*, Zhekai and Cai, Tianle and Li, Xiuyu and Guo, Junxian and Xie, Enze and Meng, Chenlin and Zhu, Jun-Yan and Han, Song},
  booktitle={The Thirteenth International Conference on Learning Representations},
  year={2025}
}
@article{
  xie2024sana,
  title={Sana: Efficient high-resolution image synthesis with linear diffusion transformers},
  author={Xie, Enze and Chen, Junsong and Chen, Junyu and Cai, Han and Tang, Haotian and Lin, Yujun and Zhang, Zhekai and Li, Muyang and Zhu, Ligeng and Lu, Yao and others},
  journal={arXiv preprint arXiv:2410.10629},
  year={2024}
}
Downloads last month
471
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for mit-han-lab/nunchaku-sana

Unable to build the model tree, the base model loops to the model itself. Learn more.

Dataset used to train mit-han-lab/nunchaku-sana

Collection including mit-han-lab/nunchaku-sana