2021
DOI: 10.48550/arxiv.2107.13435
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MWP-BERT: Numeracy-Augmented Pre-training for Math Word Problem Solving

Abstract: Math word problem (MWP) solving is the task of transforming a sequence of natural language problem descriptions to executable math equations. An MWP solver not only needs to understand complex scenarios described in the problem texts, but also identify the key mathematical variables and associate text descriptions with math equation logic. Although recent sequence modeling MWP solvers have gained credits on the math-text contextual understanding, pre-trained language models (PLM) have not been explored for sol… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…GenBERT [47] and NF-NSM [39] enhance the numerical reasoning capabilities of models by incorporating numerical data into the training process of PLMs. MWP-BERT [95] further enhances the model's Mathematical Language Models: A Survey 111:9 capacity to represent and calculate numerical values by incorporating numeric attributes into symbol placeholders. MathBERT [136] employs additional joint training of text and formulas to effectively capture the semantic-level structural information of formulas.…”
Section: Non-autoregression Lmsmentioning
confidence: 99%
“…GenBERT [47] and NF-NSM [39] enhance the numerical reasoning capabilities of models by incorporating numerical data into the training process of PLMs. MWP-BERT [95] further enhances the model's Mathematical Language Models: A Survey 111:9 capacity to represent and calculate numerical values by incorporating numeric attributes into symbol placeholders. MathBERT [136] employs additional joint training of text and formulas to effectively capture the semantic-level structural information of formulas.…”
Section: Non-autoregression Lmsmentioning
confidence: 99%
“…Instead, these models can undergo fine-tuning directly on the pre-trained counterparts, streamlining the training process and yielding outstanding results [17,18]. In the work of MWp-bert [19], pre-trained models were employed for solving mathematical word problems. GeoQA+ [20] utilized pre-trained models to transform the textual information of geometric problems, thereby augmenting the dataset.…”
Section: Pre-training Model In Nlpmentioning
confidence: 99%
“…Previous work has seen modest success on simpler or specialized mathematics problem benchmarks. Techniques based on co-training output to verify (7,8) or predict expression trees to evaluate (9)(10)(11)(12)(13)(14) are able to solve elementary schoollevel math problems, such as MAWPS and Math23k, with over 80% accuracy. However, these approaches does not extend to high-school, math Olympiad, nor university-level courses.…”
Section: Related Workmentioning
confidence: 99%
“…The parabola y = a + bx + cx 2 goes through the points (x, y) = (1, 4) and (2, 8) and (3,14). Find and solve a matrix equation for the unknowns (a, b, c).…”
Section: Topic Elimination Using Matrices Original Questionmentioning
confidence: 99%