Enhancing Zero-Shot Chain-of-Thought Reasoning in Large Language Models through Logic
Abstract: Recent advancements in LLMs have showcased their remarkable generalizability across various domains. However, their reasoning abilities still have significant room for improvement, especially when confronted with scenarios requiring multi-step reasoning. Although LLMs possess extensive knowledge, their reasoning often fails to effectively utilize this knowledge to establish a coherent thinking paradigm. These models sometimes show hallucinations as their reasoning procedures are unconstrained by logical principles. Aiming at improving the zero-shot chain-of-thought reasoning ability of LLMs, we propose LoT (Logical Thoughts), a self-improvement prompting framework that leverages principles rooted in symbolic logic, particularly Reductio ad Absurdum, to systematically verify and rectify the reasoning processes step by step. Experimental evaluations conducted on language tasks in diverse domains, including arithmetic, commonsense, symbolic, causal inference, and social problems, demonstrate the efficacy of enhanced reasoning by logic. The implementation code for LoT can be accessed at: https://github.com/xf-zhao/LoT.
- David Agler. 2012. Symbolic Logic: Syntax, Semantics, and Proof. Rowman & Littlefield Publishers, Lanham, Md.
- Michael Ahn et al. 2022. Do As I Can, Not As I Say: Grounding Language in Robotic Affordances.
- A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity. arXiv preprint arXiv:2302.04023.
- Sparks of Artificial General Intelligence: Early experiments with GPT-4.
- Training Verifiers to Solve Math Word Problems.
- Antonia Creswell and Murray Shanahan. 2022. Faithful Reasoning Using Large Language Models.
- Selection-Inference: Exploiting Large Language Models for Interpretable Logical Reasoning.
- Artur d’Avila Garcez and Luis C. Lamb. 2020. Neurosymbolic AI: The 3rd Wave. arXiv:2012.05876 [cs].
- Maieutic Prompting: Logically Consistent Reasoning with Recursive Explanations. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 1266–1279, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Large Language Models are Zero-Shot Reasoners. In Advances in Neural Information Processing Systems.
- Making Large Language Models Better Reasoners with Step-Aware Verifier.
- Let’s Verify Step by Step.
- Program Induction by Rationale Generation: Learning to Solve and Explain Algebraic Word Problems. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 158–167, Vancouver, Canada. Association for Computational Linguistics.
- Deductive Verification of Chain-of-Thought Reasoning.
- Self-Refine: Iterative Refinement with Self-Feedback.
- Between words and characters: A Brief History of Open-Vocabulary Modeling and Tokenization in NLP.
- Improving Coherence and Consistency in Neural Sequence Models with Dual-System, Neuro-Symbolic Reasoning.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
- Neuro-symbolic artificial intelligence. AI Communications, 34(3):197–209.
- Toolformer: Language Models Can Teach Themselves to Use Tools.
- Aarohi Srivastava et al. 2023. Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models.
- Self-Consistency Improves Chain of Thought Reasoning in Language Models.
- Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. In Advances in Neural Information Processing Systems.
- Generating Sequences by Learning to Self-Correct.
- Tree of Thoughts: Deliberate Problem Solving with Large Language Models.
- ReAct: Synergizing Reasoning and Acting in Language Models. In The Eleventh International Conference on Learning Representations.
- Chat with the Environment: Interactive Multimodal Perception using Large Language Models.
- Progressive-Hint Prompting Improves Reasoning in Large Language Models.
- Least-to-Most Prompting Enables Complex Reasoning in Large Language Models.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.