Wait, We Don't Need to "Wait"! Removing Thinking Tokens Improves Reasoning Efficiency
Wang, Chenlong, Feng, Yuanning, Chen, Dongping, Chu, Zhaoyang, Krishna, Ranjay, Zhou, Tianyi
–arXiv.org Artificial Intelligence
Recent advances in large reasoning models have enabled complex, step-by-step reasoning but often introduce significant overthinking, resulting in verbose and redundant outputs that hinder efficiency. In this study, we examine whether explicit self-reflection, signaled by tokens such as "Wait" and "Hmm", is necessary for advanced reasoning. We propose NoWait, a simple yet effective approach that disables explicit self-reflection by suppressing these tokens during inference. Extensive experiments on ten benchmarks across textual, visual, and video reasoning tasks show that NoWait reduces chain-of-thought trajectory length by up to 27%-51% in five R1-style model series, without compromising model utility. NoWait thus offers a plug-and-play solution for efficient and utility-preserving multimodal reasoning.
arXiv.org Artificial Intelligence
Jun-19-2025
- Country:
- Asia
- Middle East > Jordan (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- North America > United States
- Ohio (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.34)
- Industry:
- Education (0.68)
- Technology:
- Information Technology > Artificial Intelligence
- Cognitive Science > Problem Solving (1.00)
- Machine Learning > Neural Networks
- Deep Learning (1.00)
- Natural Language
- Chatbot (0.94)
- Large Language Model (1.00)
- Representation & Reasoning (1.00)
- Vision (0.68)
- Information Technology > Artificial Intelligence