Deconstructing Generative Diversity: An Information Bottleneck Analysis of Discrete Latent Generative Models
Wu, Yudi, Zhao, Wenhao, Liu, Dianbo
–arXiv.org Artificial Intelligence
Generative diversity varies significantly across discrete latent generative models such as AR, MIM, and Diffusion. We propose a diagnostic framework, grounded in Information Bottleneck (IB) theory, to analyze the underlying strategies resolving this behavior. The framework models generation as a conflict between a 'Compression Pressure' - a drive to minimize overall codebook entropy - and a 'Diversity Pressure' - a drive to maximize conditional entropy given an input. We further decompose this diversity into two primary sources: 'Path Diversity', representing the choice of high-level generative strategies, and 'Execution Diversity', the randomness in executing a chosen strategy. To make this decomposition operational, we introduce three zero-shot, inference-time interventions that directly perturb the latent generative process and reveal how models allocate and express diversity. Application of this probe-based framework to representative AR, MIM, and Diffusion systems reveals three distinct strategies: "Diversity-Prioritized" (MIM), "Compression-Prioritized" (AR), and "Decoupled" (Diffusion). Our analysis provides a principled explanation for their behavioral differences and informs a novel inference-time diversity enhancement technique.
arXiv.org Artificial Intelligence
Dec-2-2025
- Country:
- Europe > Austria
- Vienna (0.15)
- North America > United States (0.49)
- Europe > Austria
- Genre:
- Research Report (0.82)
- Technology: