Semantic Bottleneck Scene Generation
Azadi, Samaneh, Tschannen, Michael, Tzeng, Eric, Gelly, Sylvain, Darrell, Trevor, Lucic, Mario
Coupling the high-fidelity generation capabilities of label-conditional image synthesis methods with the flexibility of unconditional generative models, we propose a semantic bottleneck GAN model for unconditional synthesis of complex scenes. W e assume pixel-wise segmentation labels are available during training and use them to learn the scene structure. During inference, our model first synthesizes a realistic segmentation layout from scratch, then synthesizes a realistic scene conditioned on that layout. F or the former, we use an unconditional progressive segmentation generation network that captures the distribution of realistic semantic scene layouts. F or the latter, we use a conditional segmentation-to-image synthesis network that captures the distribution of photo-realistic images conditioned on the semantic layout. When trained end-to-end, the resulting model outperforms state-of-the-art generative models in unsupervised image synthesis on two challenging domains in terms of the Fr echet Inception Distance and user-study evaluations. Moreover, we demonstrate the generated segmentation maps can be used as additional training data to strongly improve recent segmentation-to-image synthesis networks.
Nov-26-2019
- Country:
- Europe
- Germany (0.04)
- Italy > Calabria
- Catanzaro Province > Catanzaro (0.04)
- North America > United States
- California > Alameda County > Berkeley (0.04)
- Europe
- Genre:
- Research Report (0.64)
- Industry:
- Information Technology (0.47)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (0.71)
- Natural Language > Generation (0.55)
- Vision (1.00)
- Information Technology > Artificial Intelligence