From Quantity to Quality: Boosting LLM Performance with Self-Guided Data Selection for Instruction Tuning
Li, Ming, Zhang, Yong, Li, Zhitao, Chen, Jiuhai, Chen, Lichang, Cheng, Ning, Wang, Jianzong, Zhou, Tianyi, Xiao, Jing
–arXiv.org Artificial Intelligence
In the realm of Large Language Models, the balance between instruction data quality and quantity has become a focal point. Recognizing this, we introduce a self-guided methodology for LLMs to autonomously discern and select cherry samples from vast open-source datasets, effectively minimizing manual curation and potential cost for instruction tuning an LLM. Our key innovation, the Instruction-Following Difficulty (IFD) metric, emerges as a pivotal tool to identify discrepancies between a model's expected responses and its autonomous generation prowess. Through the adept application of IFD, cherry samples are pinpointed, leading to a marked uptick in model training efficiency. Empirical validations on renowned datasets like Alpaca and WizardLM underpin our findings; with a mere 10% of conventional data input, our strategy showcases improved results. This synthesis of self-guided cherry-picking and the IFD metric signifies a transformative leap in the optimization of LLMs, promising both efficiency and resource-conscious advancements. Codes, data, and models are available: https://github.com/MingLiiii/Cherry_LLM
arXiv.org Artificial Intelligence
Sep-15-2023
- Country:
- Asia > Middle East
- UAE (0.14)
- Europe > United Kingdom
- England (0.14)
- North America > United States (0.67)
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Technology: