Efficient Multi-Task Learning via Generalist Recommender
Wang, Luyang, Tang, Cangcheng, Zhang, Chongyang, Ruan, Jun, Huang, Kai, Dai, Jason
–arXiv.org Artificial Intelligence
Multi-task learning (MTL) is a common machine learning technique that allows the model to share information across different tasks and improve the accuracy of recommendations for all of them. Many existing MTL implementations suffer from scalability issues as the training and inference performance can degrade with the increasing number of tasks, which can limit production use case scenarios for MTL-based recommender systems. Inspired by the recent advances of large language models, we developed an end-to-end efficient and scalable Generalist Recommender (GRec). GRec takes comprehensive data signals by utilizing NLP heads, parallel Transformers, as well as a wide and deep structure to process multi-modal inputs. These inputs are then combined and fed through a newly proposed task-sentence level routing mechanism to scale the model capabilities on multiple tasks without compromising performance. Offline evaluations and online experiments show that GRec significantly outperforms our previous recommender solutions. GRec has been successfully deployed on one of the largest telecom websites and apps, effectively managing high volumes of online traffic every day.
arXiv.org Artificial Intelligence
Apr-9-2025
- Country:
- Asia
- Middle East > Jordan (0.04)
- Russia (0.04)
- Europe
- Ireland (0.04)
- Netherlands (0.04)
- Russia (0.04)
- Spain (0.04)
- United Kingdom > England
- West Midlands > Birmingham (0.05)
- North America > United States
- California > Santa Clara County
- Santa Clara (0.05)
- Georgia > Fulton County
- Alpharetta (0.04)
- Massachusetts > Suffolk County
- Boston (0.04)
- New York > New York County
- New York City (0.04)
- California > Santa Clara County
- Asia
- Genre:
- Research Report (0.64)
- Industry:
- Information Technology (0.47)
- Telecommunications (0.69)
- Technology: