METEOR: Evolutionary Journey of Large Language Models from Guidance to Self-Growth
Li, Jiawei, Xu, Xiaoang, Gao, Yang
–arXiv.org Artificial Intelligence
Model evolution enables learning from feedback to refine experiences and update skills, transforming models from having no domain knowledge to becoming domain experts. However, there is currently no unified and effective method for guiding this evolutionary process. To address this gap, we propose the Meteor method, which includes three training phases: weak-to-strong data distillation, iterative training, and self-evolution strategies. Each phase maximizes the model's inherent domain capabilities, allowing it to autonomously refine its domain knowledge and enhance performance. Experiments demonstrate that our approach significantly improves accuracy, completeness, relevance, coherence, and reliability across domain-specific tasks.
arXiv.org Artificial Intelligence
Nov-29-2024
- Country:
- Asia
- British Indian Ocean Territory > Diego Garcia (0.04)
- China > Beijing
- Beijing (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- Singapore (0.04)
- Taiwan > Taiwan Province
- Taipei (0.04)
- Thailand > Bangkok
- Bangkok (0.04)
- Europe
- Austria > Vienna (0.14)
- United Kingdom (0.04)
- North America
- Canada
- British Columbia > Metro Vancouver Regional District
- Vancouver (0.04)
- Ontario > Toronto (0.04)
- British Columbia > Metro Vancouver Regional District
- United States > Louisiana
- Orleans Parish > New Orleans (0.05)
- Canada
- Asia
- Genre:
- Overview (0.93)
- Research Report > New Finding (0.46)
- Industry:
- Education > Educational Setting > Online (0.68)
- Technology: