KL-Divergence Guided Temperature Sampling
Chang, Chung-Ching, Reitter, David, Aksitov, Renat, Sung, Yun-Hsuan
–arXiv.org Artificial Intelligence
Temperature sampling is a conventional approach to diversify large language model predictions. As temperature increases, the prediction becomes diverse but also vulnerable to hallucinations -- generating tokens that are sensible but not factual. One common approach to mitigate hallucinations is to provide source/grounding documents and the model is trained to produce predictions that bind to and are attributable to the provided source. It appears that there is a trade-off between diversity and attribution. To mitigate any such trade-off, we propose to relax the constraint of having a fixed temperature over decoding steps, and a mechanism to guide the dynamic temperature according to its relevance to the source through KL-divergence. Our experiments justifies the trade-off, and shows that our sampling algorithm outperforms the conventional top-k and top-p algorithms in conversational question-answering and summarization tasks.
arXiv.org Artificial Intelligence
Nov-29-2023
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe
- Portugal (0.04)
- Ireland (0.04)
- United Kingdom (0.47)
- Belgium (0.04)
- Isle of Man (0.04)
- France (0.04)
- Switzerland (0.04)
- Austria (0.04)
- Spain > Galicia
- Madrid (0.04)
- Russia > Central Federal District
- Moscow Oblast > Moscow (0.04)
- North America
- Canada > Manitoba
- Winnipeg Metropolitan Region > Winnipeg (0.04)
- Mexico (0.04)
- United States
- California > Los Angeles County
- Los Angeles (0.04)
- New York > Bronx County
- New York City (0.04)
- California > Los Angeles County
- Canada > Manitoba
- Asia > Middle East
- Genre:
- Contests & Prizes (0.94)
- Research Report (1.00)
- Industry:
- Health & Medicine (1.00)
- Law (0.68)
- Leisure & Entertainment > Sports
- Hockey (1.00)
- Technology: