DDT: Decoupled Diffusion Transformer
Collection
5 items
โข
Updated
We decouple diffusion transformer into encoder-decoder design, and surpresingly that a more substantial encoder yields performance improvements as model size increases.
We take the off-shelf VAE to encode image into latent space, and train the decoder with DDT.
Dataset | Model | Params | FID | HuggingFace |
---|---|---|---|---|
ImageNet256 | DDT-XL/2(22en6de) | 675M | 1.26 | ๐ค |
ImageNet512 | DDT-XL/2(22en6de) | 675M | 1.28 | ๐ค |
Coming soon.
We use ADM evaluation suite to report FID.
# for installation
pip install -r requirements.txt
# for inference
python main.py predict -c configs/repa_improved_ddt_xlen22de6_256.yaml --ckpt_path=XXX.ckpt
# for training
# extract image latent (optional)
python3 tools/cache_imlatent4.py
# train
python main.py fit -c configs/repa_improved_ddt_xlen22de6_256.yaml
@ARTICLE{ddt,
title = "DDT: Decoupled Diffusion Transformer",
author = "Wang, Shuai and Tian, Zhi and Huang, Weilin and Wang, Limin",
month = apr,
year = 2025,
archivePrefix = "arXiv",
primaryClass = "cs.CV",
eprint = "2504.05741"
}