Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.168
|View full text |Cite
|
Sign up to set email alerts
|

Are NLP Models really able to Solve Simple Math Word Problems?

Abstract: The problem of designing NLP solvers for math word problems (MWP) has seen sustained research activity and steady gains in the test accuracy. Since existing solvers achieve high performance on the benchmark datasets for elementary level MWPs containing one-unknown arithmetic word problems, such problems are often considered "solved" with the bulk of research attention moving to more complex MWPs. In this paper, we restrict our attention to English MWPs taught in grades four and lower. We provide strong evidenc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
129
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 108 publications
(129 citation statements)
references
References 22 publications
(28 reference statements)
0
129
0
Order By: Relevance
“…The tokenization scheme could be the cause for limited extrapolation, since language models get better at arithmetic when numbers are tokenized at the digit/character level (Nogueira et al, 2021;Wallace et al, 2019). For arithmetic word problems, state of the art solvers rely on predicting an equation, which is then filled in with specific numeric values from the question (Patel et al, 2021), altogether bypassing the need for encoding numbers into embeddings.…”
Section: Resultsmentioning
confidence: 99%
“…The tokenization scheme could be the cause for limited extrapolation, since language models get better at arithmetic when numbers are tokenized at the digit/character level (Nogueira et al, 2021;Wallace et al, 2019). For arithmetic word problems, state of the art solvers rely on predicting an equation, which is then filled in with specific numeric values from the question (Patel et al, 2021), altogether bypassing the need for encoding numbers into embeddings.…”
Section: Resultsmentioning
confidence: 99%
“…Datasets We conduct experiments on four datasets across two different languages: MAWPS (Koncel-Kedziorski et al, 2016), Math23k (Wang et al, 2017), MathQA (Amini et al, 2019), and SVAMP (Patel et al, 2021). The dataset statistics can be found in Table 2…”
Section: Methodsmentioning
confidence: 99%
“…S2T/G2T GTS (Xie and Sun, 2019) 82.6 Graph2Tree 85.6 Roberta-GTS (Patel et al, 2021) 88.5 Roberta-Graph2Tree (Patel et al, 2021) 88 adapt the dataset to filter out some questions that are unsolvable. We consider the operations "addition", "subtraction", "multiplication", and "division" for MAWPS and SVAMP, and an extra "exponentiation" for MathQA and Math23k.…”
Section: S2smentioning
confidence: 99%
See 1 more Smart Citation
“…However, while using deep learning to solve MWPs, existing methods (Xie and Sun, 2019;Zhang et al, 2020) get stuck in memorizing procedures. Patel et al (2021) provide evidence that these methods rely on shallow heuristics to generate equations. We look at this issue and think it is because they focus on text understanding or equation generation for one problem.…”
Section: Introductionmentioning
confidence: 99%