Multimodal graph representation learning for website generation based on visual sketch

Vu, Tung D., Hoang, Chung, Hy, Truong-Son

arXiv.org Artificial Intelligence 

The Design2Code problem, which involves converting digital designs into functional source code, is a significant challenge in software development due to its complexity and time-consuming nature. In this paper, we propose a novel method that leverages multimodal graph representation learning to address these challenges. By integrating both visual and structural information from design sketches, our approach enhances the accuracy and efficiency of code generation, particularly in producing semantically correct and structurally sound HTML code. We present a comprehensive evaluation of our method, demonstrating significant improvements in both accuracy and efficiency compared to existing techniques. Extensive evaluation demonstrates significant improvements of multimodal graph learning over existing techniques, highlighting the potential of our method to revolutionize design-to-code automation. The Design2Code problem, which involves converting UI designs into functional source code, is a pivotal challenge in software development that lies at the intersection of computer vision, natural language processing, and programming. This task is particularly demanding when generating HTML code from webpage designs, as it requires not only the interpretation of visual elements but also an understanding of their spatial arrangements and hierarchical relationships.