Length Controlled Generation for Black-box LLMs
Gu, Yuxuan, Wang, Wenjie, Feng, Xiaocheng, Zhong, Weihong, Zhu, Kun, Huang, Lei, Chua, Tat-Seng, Qin, Bing
–arXiv.org Artificial Intelligence
Large language models (LLMs) have demonstrated impressive instruction following capabilities, while still struggling to accurately manage the length of the generated text, which is a fundamental requirement in many real-world applications. Existing length control methods involve fine-tuning the parameters of LLMs, which is inefficient and suboptimal for practical use. In this paper, we propose a novel iterative sampling framework for text length control, integrating the Metropolis-Hastings algorithm with an importance sampling acceleration strategy. This framework efficiently and reliably regulates LLMs to generate length-constrained text without modifying the underlying parameters, thereby preserving the original capabilities of LLMs. Experimental results demonstrate that our framework achieves almost 100\% success rates of length control on Llama3.1 for tasks such as length-controlled abstractive summarization and length-constrained instruction following, with minimal additional computational overhead. This also highlights the significant potential of our method for precise length control across a broader range of applications, without compromising the versatility of LLMs.
arXiv.org Artificial Intelligence
Dec-19-2024
- Country:
- Asia (0.46)
- Europe > Germany (0.28)
- North America > United States (0.46)
- Genre:
- Research Report > New Finding (0.34)
- Industry:
- Health & Medicine > Therapeutic Area (0.46)
- Transportation > Air (0.51)
- Technology: