Improving Diversity in Language Models: When Temperature Fails, Change the Loss
Verine, Alexandre, Bronnec, Florian Le, Zheng, Kunhao, Allauzen, Alexandre, Chevaleyre, Yann, Negrevergne, Benjamin
–arXiv.org Artificial Intelligence
Increasing diversity in language models is a challenging yet essential objective. A common approach is to raise the decoding temperature. In this work, we investigate this approach through a simplistic yet common case to provide insights into why decreasing temperature can improve quality (Precision), while increasing it often fails to boost coverage (Recall). Our analysis reveals that for a model to be effectively tunable through temperature adjustments, it must be trained toward coverage. To address this, we propose rethinking loss functions in language models by leveraging the Precision-Recall framework. Our results demonstrate that this approach achieves a substantially better trade-off between Precision and Recall than merely combining negative log-likelihood training with temperature scaling. These findings offer a pathway toward more versatile and robust language modeling techniques.
arXiv.org Artificial Intelligence
Aug-14-2025
- Country:
- Asia
- Europe
- Denmark > Capital Region
- Copenhagen (0.04)
- France > Île-de-France
- Denmark > Capital Region
- North America > Canada (0.04)
- Oceania > Australia
- Genre:
- Research Report > New Finding (1.00)
- Technology: