Rationales Are Not Silver Bullets: Measuring the Impact of Rationales on Model Performance and Reliability
Zhu, Chiwei, Xu, Benfeng, Yang, An, Lin, Junyang, Wang, Quan, Zhou, Chang, Mao, Zhendong
–arXiv.org Artificial Intelligence
Training language models with rationales augmentation has been shown to be beneficial in many existing works. In this paper, we identify that such a prevailing view does not hold consistently. We conduct comprehensive investigations to thoroughly inspect the impact of rationales on model performance as well as a novel perspective of model reliability. The results lead to several key findings that add new insights upon existing understandings: 1) Rationales can, at times, deteriorate model performance; 2) Rationales can, at times, improve model reliability, even outperforming their untrained counterparts; 3) A linear correspondence exists in between the performance and reliability improvements, while both are driven by the intrinsic difficulty of the task. These findings provide informative regulations on the broad utilization of rationales and raise critical implications on the procedure of explicitly aligning language models with implicit human thoughts. Codes can be found at https://github.com/Ignoramus0817/rationales.
arXiv.org Artificial Intelligence
Jun-2-2025
- Country:
- Asia
- Europe
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Monaco (0.04)
- Portugal > Lisbon
- Lisbon (0.04)
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- United Kingdom (0.04)
- Ireland > Leinster
- North America > United States
- California > San Francisco County
- San Francisco (0.04)
- Georgia > Fulton County
- Atlanta (0.04)
- Texas > Travis County
- Austin (0.04)
- Washington > King County
- Seattle (0.04)
- California > San Francisco County
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Education (0.46)
- Technology: