Controlling Summarization Length Through EOS Token Weighting
Belligoli, Zeno, Stergiadis, Emmanouil, Fainman, Eran, Gusev, Ilya
–arXiv.org Artificial Intelligence
Controlling the length of generated text can be crucial in various text-generation tasks, including summarization. Existing methods often require complex model alterations, limiting compatibility with pre-trained models. We address these limitations by developing a simple approach for controlling the length of automatic text summaries by increasing the importance of correctly predicting the EOS token in the cross-entropy loss computation. The proposed methodology is agnostic to architecture and decoding algorithms and orthogonal to other inference-time techniques to control the generation length. We tested it with encoder-decoder and modern GPT-style LLMs, and show that this method can control generation length, often without affecting the quality of the summary.
arXiv.org Artificial Intelligence
Jun-6-2025
- Country:
- Asia > China
- Hong Kong (0.04)
- North America
- Canada (0.04)
- United States > Massachusetts
- Middlesex County > Cambridge (0.04)
- Oceania > Australia
- Asia > China
- Genre:
- Research Report (0.40)
- Technology: