Co-Supervised Learning: Improving Weak-to-Strong Generalization with Hierarchical Mixture of Experts
Liu, Yuejiang, Alahi, Alexandre
–arXiv.org Artificial Intelligence
Steering the behavior of a strong model pre-trained on internet-scale data can be difficult due to the scarcity of competent supervisors. Recent studies reveal that, despite supervisory noises, a strong student model may surpass its weak teacher when fine-tuned on specific objectives. Yet, the effectiveness of such weak-to-strong generalization remains limited, especially in the presence of large capability gaps. In this paper, we propose to address this challenge by harnessing a diverse set of specialized teachers, instead of a single generalist one, that collectively supervises the strong student. Our approach resembles the classical hierarchical mixture of experts, with two components tailored for co-supervision: (i) we progressively alternate student training and teacher assignment, leveraging the growth of the strong student to identify plausible supervisions; (ii) we conservatively enforce teacher-student and local-global consistency, leveraging their dependencies to reject potential annotation noises. We validate the proposed method through visual recognition tasks on the OpenAI weak-to-strong benchmark and additional multi-domain datasets. Our code is available at \url{https://github.com/yuejiangliu/csl}.
arXiv.org Artificial Intelligence
Feb-23-2024
- Country:
- Asia
- Japan > Honshū
- Chūbu > Aichi Prefecture > Nagoya (0.04)
- Middle East > Jordan (0.04)
- Japan > Honshū
- Europe > Switzerland
- North America > United States
- California
- San Francisco County > San Francisco (0.14)
- Santa Clara County > Palo Alto (0.04)
- California
- Asia
- Genre:
- Instructional Material > Course Syllabus & Notes (0.46)
- Research Report > New Finding (0.46)
- Industry:
- Education (0.55)
- Technology: