daDPO: Distribution-Aware DPO for Distilling Conversational Abilities
Zhang, Zhengze, Wang, Shiqi, Shen, Yiqun, Guo, Simin, Lin, Dahua, Wang, Xiaoliang, Cam-Tu, Nguyen, Tan, Fei
–arXiv.org Artificial Intelligence
Large language models (LLMs) have demonstrated exceptional performance across various applications, but their conversational abilities decline sharply as model size decreases, presenting a barrier to their deployment in resource-constrained environments. Knowledge distillation with Direct Preference Optimization (dDPO) has emerged as a promising approach to enhancing the conversational abilities of smaller models using a larger teacher model. However, current methods primarily focus on 'black-box' KD, which only uses the teacher's responses, overlooking the output distribution offered by the teacher. This paper addresses this gap by introducing daDPO (Distribution-Aware DPO), a unified method for preference optimization and distribution-based distillation. We provide rigorous theoretical analysis and empirical validation, showing that daDPO outperforms existing methods in restoring performance for pruned models and enhancing smaller LLM models. Notably, in in-domain evaluation, our method enables a 20% pruned Vicuna1.5-7B to achieve near-teacher performance (-7.3% preference rate compared to that of dDPO's -31%), and allows Qwen2.5-1.5B to occasionally outperform its 7B teacher model (14.0% win rate).
arXiv.org Artificial Intelligence
Jun-23-2025
- Country:
- Asia
- China
- Hong Kong (0.04)
- Jiangsu Province > Nanjing (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- China
- North America > United States
- Illinois > Cook County > Chicago (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Banking & Finance > Economy (0.94)
- Technology: