LogiCoT: Logical Chain-of-Thought Instruction-Tuning
Liu, Hanmeng, Teng, Zhiyang, Cui, Leyang, Zhang, Chaoli, Zhou, Qiji, Zhang, Yue
–arXiv.org Artificial Intelligence
Generative Pre-trained Transformer 4 (GPT-4) demonstrates impressive chain-of-thought reasoning ability. Recent work on self-instruction tuning, such as Alpaca, has focused on enhancing the general proficiency of models. These instructions enable the model to achieve performance comparable to GPT-3.5 on general tasks like open-domain text generation and paraphrasing. However, they fall short of helping the model handle complex reasoning tasks. To bridge the gap, this paper presents LogiCoT, a new instruction-tuning dataset for Logical Chain-of-Thought reasoning with GPT-4. We elaborate on the process of harvesting instructions for prompting GPT-4 to generate chain-of-thought rationales. LogiCoT serves as an instruction set for teaching models of logical reasoning and elicits general reasoning skills.
arXiv.org Artificial Intelligence
Oct-28-2023
- Country:
- Europe > Italy (0.14)
- North America > Canada (0.14)
- Genre:
- Research Report (0.64)
- Industry:
- Education > Curriculum > Subject-Specific Education (0.67)
- Technology: