A Scaling Law for Token Efficiency in LLM Fine-Tuning Under Fixed Compute Budgets
Lagasse, Ryan, Kierans, Aidan, Ghosh, Avijit, Dori-Hacohen, Shiri
–arXiv.org Artificial Intelligence
We introduce a scaling law for fine-tuning large language models (LLMs) under fixed compute budgets that explicitly accounts for data composition. Conventional approaches measure training data solely by total tokens, yet the number of examples and their average token length--what we term dataset volume --play a decisive role in model performance. Experiments on the BRICC dataset Salavati et al. (2024) and subsets of the MMLU dataset Hendrycks et al. (2021), evaluated under multiple subsampling strategies, reveal that data composition significantly affects token efficiency. These results motivate refined scaling laws for practical LLM fine-tuning in resource-constrained settings. Code will be made available upon publication.
arXiv.org Artificial Intelligence
Jun-4-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > Montenegro (0.04)
- North America > United States
- Connecticut > Tolland County > Storrs (0.14)
- Asia > Middle East
- Genre:
- Research Report (1.00)
- Technology: