Inductive or Deductive? Rethinking the Fundamental Reasoning Abilities of LLMs (2408.00114v2)
Abstract: Reasoning encompasses two typical types: deductive reasoning and inductive reasoning. Despite extensive research into the reasoning capabilities of LLMs, most studies have failed to rigorously differentiate between inductive and deductive reasoning, leading to a blending of the two. This raises an essential question: In LLM reasoning, which poses a greater challenge - deductive or inductive reasoning? While the deductive reasoning capabilities of LLMs, (i.e. their capacity to follow instructions in reasoning tasks), have received considerable attention, their abilities in true inductive reasoning remain largely unexplored. To investigate into the true inductive reasoning capabilities of LLMs, we propose a novel framework, SolverLearner. This framework enables LLMs to learn the underlying function (i.e., $y = f_w(x)$), that maps input data points $(x)$ to their corresponding output values $(y)$, using only in-context examples. By focusing on inductive reasoning and separating it from LLM-based deductive reasoning, we can isolate and investigate inductive reasoning of LLMs in its pure form via SolverLearner. Our observations reveal that LLMs demonstrate remarkable inductive reasoning capabilities through SolverLearner, achieving near-perfect performance with ACC of 1 in most cases. Surprisingly, despite their strong inductive reasoning abilities, LLMs tend to relatively lack deductive reasoning capabilities, particularly in tasks involving ``counterfactual'' reasoning.
- Graph of thoughts: Solving elaborate problems with large language models. arXiv preprint arXiv:2308.09687.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
- Large language models as tool makers. arXiv preprint arXiv:2305.17126.
- Explaining answers with entailment trees. arXiv preprint arXiv:2104.08661.
- A survey for in-context learning. arXiv preprint arXiv:2301.00234.
- Large language models are not abstract reasoners. arXiv preprint arXiv:2305.19555.
- Demystifying prompts in language models via perplexity estimation. arXiv preprint arXiv:2212.04037.
- Folio: Natural language reasoning with first-order logic. arXiv preprint arXiv:2209.00840.
- Jie Huang and Kevin Chen-Chuan Chang. 2022. Towards reasoning in large language models: A survey. arXiv preprint arXiv:2212.10403.
- Large language models are zero-shot reasoners. Advances in neural information processing systems, 35:22199–22213.
- What makes good in-context examples for gpt-3333? arXiv preprint arXiv:2101.06804.
- Large language models as general pattern machines. arXiv preprint arXiv:2307.04721.
- OpenAI. 2023. Gpt-4 technical report. ArXiv, abs/2303.08774.
- Logic-lm: Empowering large language models with symbolic solvers for faithful logical reasoning. arXiv preprint arXiv:2305.12295.
- Creator: Tool creation for disentangling abstract and concrete reasoning of large language models. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 6922–6939.
- Learning to retrieve prompts for in-context learning. arXiv preprint arXiv:2112.08633.
- Clutrr: A diagnostic benchmark for inductive reasoning from text. arXiv preprint arXiv:1908.06177.
- Large language models are in-context semantic reasoners rather than symbolic reasoners. arXiv preprint arXiv:2305.14825.
- Large language models still can’t plan (a benchmark for llms on planning and reasoning about change). arXiv preprint arXiv:2206.10498.
- Emergent abilities of large language models. arXiv preprint arXiv:2206.07682.
- Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
- Reasoning or reciting? exploring the capabilities and limitations of language models through counterfactual tasks. arXiv preprint arXiv:2307.02477.
- Are large language models really good logical reasoners? a comprehensive evaluation from deductive, inductive and abductive views. arXiv preprint arXiv:2306.09841.
- Llms and the abstraction and reasoning corpus: Successes, failures, and the importance of object-based representations. arXiv preprint arXiv:2305.18354.
- Language models as inductive reasoners. arXiv preprint arXiv:2212.10923.
- Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
- Reclor: A reading comprehension dataset requiring logical reasoning. arXiv preprint arXiv:2002.04326.
- Least-to-most prompting enables complex reasoning in large language models. arXiv preprint arXiv:2205.10625.
- Kewei Cheng (8 papers)
- Jingfeng Yang (31 papers)
- Haoming Jiang (52 papers)
- Zhengyang Wang (48 papers)
- Binxuan Huang (21 papers)
- Ruirui Li (33 papers)
- Shiyang Li (24 papers)
- Zheng Li (326 papers)
- Yifan Gao (69 papers)
- Xian Li (116 papers)
- Bing Yin (56 papers)
- Yizhou Sun (149 papers)