Improving User Interface Generation Models from Designer Feedback
Wu, Jason, Swearngin, Amanda, Vajjala, Arun Krishna, Leung, Alan, Nichols, Jeffrey, Barik, Titus
–arXiv.org Artificial Intelligence
Despite being trained on vast amounts of data, most LLMs are unable to reliably generate well-designed UIs. Designer feedback is essential to improving performance on UI generation; however, we find that existing RLHF methods based on ratings or rankings are not well-aligned with designers' workflows and ignore the rich rationale used to critique and improve UI designs. In this paper, we investigate several approaches for designers to give feedback to UI generation models, using familiar interactions such as commenting, sketching and direct manipulation. We first perform a study with 21 designers where they gave feedback using these interactions, which resulted in ~1500 design annotations. We then use this data to finetune a series of LLMs to generate higher quality UIs. Finally, we evaluate these models with human judges, and we find that our designer-aligned approaches outperform models trained with traditional ranking feedback and all tested baselines, including GPT-5.
arXiv.org Artificial Intelligence
Sep-23-2025
- Country:
- Asia
- Japan > Honshū
- Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Middle East > Jordan (0.04)
- Japan > Honshū
- Europe
- North America > United States
- Louisiana > Orleans Parish
- New Orleans (0.04)
- New York > New York County
- New York City (0.04)
- Louisiana > Orleans Parish
- Asia
- Genre:
- Instructional Material (0.92)
- Research Report
- Experimental Study (0.67)
- New Finding (1.00)
- Technology: