How well do Large Language Models perform in Arithmetic tasks?
(2304.02015)Abstract
Large language models have emerged abilities including chain-of-thought to answer math word problems step by step. Solving math word problems not only requires abilities to disassemble problems via chain-of-thought but also needs to calculate arithmetic expressions correctly for each step. To the best of our knowledge, there is no work to focus on evaluating the arithmetic ability of LLMs. In this work, we propose an arithmetic dataset MATH 401 to test the latest LLMs including GPT-4, ChatGPT, InstrctGPT, Galactica, and LLaMA with various arithmetic expressions and provide a detailed analysis of the ability of LLMs. MATH 401 and evaluation codes are released at \url{https://github.com/GanjinZero/math401-llm}.
We're not able to analyze this paper right now due to high demand.
Please check back later (sorry!).
Generate a summary of this paper on our Pro plan:
We ran into a problem analyzing this paper.