SoftCFG: Uncertainty-guided Stable Guidance for Visual Autoregressive Model
Abstract
SoftCFG, an uncertainty-guided inference method, enhances autoregressive image generation by distributing adaptive perturbations and stabilizing long-sequence generation, improving image quality and achieving state-of-the-art FID.
Autoregressive (AR) models have emerged as powerful tools for image generation by modeling images as sequences of discrete tokens. While Classifier-Free Guidance (CFG) has been adopted to improve conditional generation, its application in AR models faces two key issues: guidance diminishing, where the conditional-unconditional gap quickly vanishes as decoding progresses, and over-guidance, where strong conditions distort visual coherence. To address these challenges, we propose SoftCFG, an uncertainty-guided inference method that distributes adaptive perturbations across all tokens in the sequence. The key idea behind SoftCFG is to let each generated token contribute certainty-weighted guidance, ensuring that the signal persists across steps while resolving conflicts between text guidance and visual context. To further stabilize long-sequence generation, we introduce Step Normalization, which bounds cumulative perturbations of SoftCFG. Our method is training-free, model-agnostic, and seamlessly integrates with existing AR pipelines. Experiments show that SoftCFG significantly improves image quality over standard CFG and achieves state-of-the-art FID on ImageNet 256*256 among autoregressive models.
Community
Official Code for SoftCFG
We are excited to release the official implementation of SoftCFG: Uncertainty-Guided Stable Guidance for Visual Autoregressive Models.
- GitHub Repository: https://github.com/Xudangliatiger/SoftCFG
- Key Features: Uncertainty-guided perturbation, Step Normalization, compatible with AliTok and RAR.
- Installation: Follow the README for setup.
For questions, open an issue in the repo.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper