Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 162 tok/s
Gemini 2.5 Pro 56 tok/s Pro
GPT-5 Medium 38 tok/s Pro
GPT-5 High 35 tok/s Pro
GPT-4o 104 tok/s Pro
Kimi K2 164 tok/s Pro
GPT OSS 120B 426 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Brain-Inspired Two-Stage Approach: Enhancing Mathematical Reasoning by Imitating Human Thought Processes (2403.00800v1)

Published 23 Feb 2024 in cs.CL, cs.AI, and cs.LG

Abstract: Although LLMs demonstrate emergent abilities in solving math word problems, there is a challenging task in complex multi-step mathematical reasoning tasks. To improve model performance on mathematical reasoning tasks, previous work has conducted supervised fine-tuning on open-source models by improving the quality and quantity of data. In this paper, we propose a novel approach, named Brain, to imitate human thought processes to enhance mathematical reasoning abilities, using the Frontal Lobe Model to generate plans, and then employing the Parietal Lobe Model to generate code and execute to obtain answers. First, we achieve SOTA performance in comparison with Code LLaMA 7B based models through this method. Secondly, we find that plans can be explicitly extracted from natural language, code, or formal language. Our code and data are publicly available at https://github.com/cyzhh/Brain.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Palm 2 technical report.
  2. Training verifiers to solve math word problems.
  3. Towards a unified view of answer calibration for multi-step reasoning.
  4. G-llava: Solving geometric problem with multi-modal large language model.
  5. Tora: A tool-integrated reasoning agent for mathematical problem solving.
  6. Reasoning with language model is planning with world model.
  7. A closer look at the self-verification abilities of large language models in logical reasoning.
  8. Metagpt: Meta programming for a multi-agent collaborative framework.
  9. Large language models cannot self-correct reasoning yet.
  10. Decomposed prompting: A modular approach for solving complex tasks.
  11. Making large language models better reasoners with step-aware verifier.
  12. Let’s verify step by step.
  13. Deductive verification of chain-of-thought reasoning.
  14. Wizardmath.
  15. OpenAI. 2023. Gpt-4 technical report.
  16. Refiner: Reasoning feedback on intermediate representations.
  17. Direct preference optimization: Your language model is secretly a reward model.
  18. Mathematical discoveries from program search with large language models. Nature.
  19. Code llama: Open foundation models for code.
  20. Deepseekmath: Pushing the limits of mathematical reasoning in open language models.
  21. Adaplanner: Adaptive planning from feedback with language models.
  22. Mirac Suzgun and Adam Tauman Kalai. 2024. Meta-prompting: Enhancing language models with task-agnostic scaffolding.
  23. Llama: Open and efficient foundation language models.
  24. Solving olympiad geometry without human demonstrations. Nature, 625(7995):476–482.
  25. Mathcoder: Seamless code integration in llms for enhanced mathematical reasoning.
  26. Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models.
  27. Math-shepherd: A label-free step-by-step verifier for llms in mathematical reasoning.
  28. Fine-grained human feedback gives better rewards for language model training.
  29. Get an a in math: Progressive rectification prompting.
  30. Tree of thoughts: Deliberate problem solving with large language models.
  31. Outcome-supervised verifiers for planning in mathematical reasoning.
  32. Metamath: Bootstrap your own mathematical questions for large language models.
  33. Scaling relationship on learning mathematical reasoning with large language models.
  34. Mammoth: Building math generalist models through hybrid instruction tuning.
  35. Cumulative reasoning with large language models.
  36. Verify-and-edit: A knowledge-enhanced chain-of-thought framework. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5823–5840, Toronto, Canada. Association for Computational Linguistics.
  37. Progressive-hint prompting improves reasoning in large language models.
  38. Solving math word problems via cooperative reasoning induced language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4471–4485, Toronto, Canada. Association for Computational Linguistics.
  39. Large language models can learn rules.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.