Tian, Jun
Yi-Lightning Technical Report
Wake, Alan, Chen, Bei, Lv, C. X., Li, Chao, Huang, Chengen, Cai, Chenglin, Zheng, Chujie, Cooper, Daniel, Zhou, Fan, Hu, Feng, Wang, Guoyin, Ji, Heng, Qiu, Howard, Zhu, Jiangcheng, Tian, Jun, Su, Katherine, Zhang, Lihuan, Li, Liying, Song, Ming, Li, Mou, Liu, Peng, Hu, Qicheng, Wang, Shawn, Zhou, Shijun, Yang, Shiming, Li, Shiyong, Zhu, Tianhang, Xie, Wen, He, Xiang, Chen, Xiaobo, Hu, Xiaohui, Ren, Xiaoyi, Niu, Xinyao, Li, Yanpeng, Zhao, Yongke, Luo, Yongzhen, Xu, Yuchi, Sha, Yuxuan, Yan, Zhaodong, Liu, Zhiyuan, Zhang, Zirui, Dai, Zonghong
This technical report presents Yi-Lightning, our latest flagship large language model (LLM). It achieves exceptional performance, ranking 6th overall on Chatbot Arena, with particularly strong results (2nd to 4th place) in specialized categories including Chinese, Math, Coding, and Hard Prompts. Yi-Lightning leverages an enhanced Mixture-of-Experts (MoE) architecture, featuring advanced expert segmentation and routing mechanisms coupled with optimized KV-caching techniques. Our development process encompasses comprehensive pre-training, supervised fine-tuning (SFT), and reinforcement learning from human feedback (RLHF), where we devise deliberate strategies for multi-stage training, synthetic data construction, and reward modeling. Furthermore, we implement RAISE (Responsible AI Safety Engine), a four-component framework to address safety issues across pre-training, post-training, and serving phases. Empowered by our scalable super-computing infrastructure, all these innovations substantially reduce training, deployment and inference costs while maintaining high-performance standards. With further evaluations on public academic benchmarks, Yi-Lightning demonstrates competitive performance against top-tier LLMs, while we observe a notable disparity between traditional, static benchmark results and real-world, dynamic human preferences. This observation prompts a critical reassessment of conventional benchmarks' utility in guiding the development of more intelligent and powerful AI systems for practical applications. Yi-Lightning is now available through our developer platform at https://platform.lingyiwanwu.com.
Integrating Pre-trained Model into Rule-based Dialogue Management
Quan, Jun, Yang, Meng, Gan, Qiang, Xiong, Deyi, Liu, Yiming, Dong, Yuchen, Ouyang, Fangxin, Tian, Jun, Deng, Ruiling, Li, Yongzhi, Yang, Yang, Jiang, Daxin
Rule-based dialogue management is still the most popular solution for industrial task-oriented dialogue systems for their interpretablility. However, it is hard for developers to maintain the dialogue logic when the scenarios get more and more complex. On the other hand, data-driven dialogue systems, usually with end-to-end structures, are popular in academic research and easier to deal with complex conversations, but such methods require plenty of training data and the behaviors are less interpretable. In this paper, we propose a method to leverages the strength of both rule-based and data-driven dialogue managers (DM). We firstly introduce the DM of Carina Dialog System (CDS, an advanced industrial dialogue system built by Microsoft). Then we propose the "model-trigger" design to make the DM trainable thus scalable to scenario changes. Furthermore, we integrate pre-trained models and empower the DM with few-shot capability. The experimental results demonstrate the effectiveness and strong few-shot capability of our method.