Scaling Text-to-Image Diffusion Transformers with Representation Autoencoders

Image generated by Gemini AI
Research on Representation Autoencoders (RAEs) indicates they excel in large-scale text-to-image (T2I) generation, outperforming state-of-the-art Variational Autoencoders (VAEs) across model scales. RAEs show faster convergence, superior generation quality, and stability during finetuning. This suggests RAEs could streamline T2I frameworks, enhancing multimodal models that integrate visual understanding and generation.
Advancements in Text-to-Image Generation with Representation Autoencoders
Recent research demonstrates that Representation Autoencoders (RAEs) significantly enhance text-to-image (T2I) generation by scaling diffusion models beyond traditional datasets like ImageNet. The study highlights the efficacy of RAEs in high-dimensional semantic latent spaces, indicating robust performance in generating images from freeform text.
By leveraging a frozen representation encoder, SigLIP-2, the research team expanded RAE decoders' capabilities by incorporating diverse datasets, including web, synthetic, and text-rendering data. The findings suggest that while increasing model scale enhances fidelity, the composition of the training data is crucial for optimizing performance in specific domains.
Key Findings from RAE Scaling
The investigation revealed that scaling simplifies the framework. Key insights include:
- Dimension-dependent noise scheduling is vital for effective performance.
- Architectural enhancements, such as wide diffusion heads, provide minimal advantages at larger scales.
RAEs were benchmarked against the state-of-the-art FLUX Variational Autoencoder (VAE) across a range of diffusion transformer scales, from 0.5 billion to 9.8 billion parameters. Results consistently showed that RAEs surpassed VAEs in pretraining across all scales, reflecting superior capability in T2I tasks.
During finetuning on high-quality datasets, RAE models exhibited notable stability, maintaining performance after 256 epochs, while VAE-based models tended to overfit after just 64 epochs. This stability indicates the RAE's robustness in handling large-scale data.
Enhanced Performance Metrics
RAE-based diffusion models converge faster and yield better image generation quality compared to VAE counterparts, positioning RAEs as a powerful foundation for large-scale T2I generation.
Related Topics:
📰 Original Source: https://arxiv.org/abs/2601.16208v1
All rights and credit belong to the original publisher.