MADFormer
Collection
Models from the paper: MADFormer: Mixed Autoregressive and Diffusion Transformers for Continuous Image Generation.
β’
3 items
β’
Updated
β’
1
This repository provides checkpoints for MADFormer trained on Imagenet-256, combining autoregressive global conditioning and diffusion-based local refinement for high-resolution image synthesis.
MADFormer: Mixed Autoregressive & Diffusion Transformers for Continuous Image Generation
ckpts.pt
# TODO
π‘ MADFormer supports flexible ARβDiff trade-offs. On ImageNet-256, increasing AR layer allocation yields up to 60% FID improvements under low NFE settings.
If you find our work useful, please cite:
@misc{chen2025madformermixedautoregressivediffusion,
title={MADFormer: Mixed Autoregressive and Diffusion Transformers for Continuous Image Generation},
author={Junhao Chen and Yulia Tsvetkov and Xiaochuang Han},
year={2025},
eprint={2506.07999},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2506.07999},
}