DiPT: Enhancing LLM reasoning through diversified perspective-taking

Just, Hoang Anh, Dabas, Mahavir, Huang, Lifu, Jin, Ming, Jia, Ruoxi

arXiv.org Artificial Intelligence 

Correct reasoning steps are important for language models to achieve high performance on many tasks, such as commonsense reasoning, question answering, and mathematical problem-solving [Wei et al., 2022, Kojima et al., 2022, Suzgun et al., 2022]. One way to elicit reasoning is through the chain-of-thought (CoT) method Wei et al. [2022], Kojima et al. [2022], which asks the model to provide step-by-step reasoning. Another approach encourages the model to provide similar problems Yasunaga et al. [2024] as the query, indirectly compelling the model to first understand the original query. Similarly, repeating and rephrasing the query Deng et al. [2023], Mekala et al. [2023] requires the model to first understand the problem and then modify the query into its own words. This rephrasing might help simplify the problem for the model. Additionally, reasoning can be generated by indirectly providing reasoning examples in demonstrations, referred to as in-context learning (ICL) Brown et al. [2020], Min et al. [2022], Xie et al. [2021]. While these methods have demonstrated significant performance improvements, language models are still prone to errors due to incorrect context understanding or analytical steps. Furthermore, they are subject to instability when requests are paraphrased. This instability is particularly concerning in the context of adversarial prompts, where recent research [Zou et al., 2023, Zeng et al., 2024] has shown that adversaries can intentionally rewrite prompts to coax safety-aligned language models into generating objectionable content that they would not generate otherwise.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found