2022
DOI: 10.48550/arxiv.2206.14858
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Solving Quantitative Reasoning Problems with Language Models

Abstract: Language models have achieved remarkable performance on a wide range of tasks that require natural language understanding. Nevertheless, state-of-the-art models have generally struggled with tasks that require quantitative reasoning, such as solving mathematics, science, and engineering problems at the college level. To help close this gap, we introduce Minerva , a large language model pretrained on general natural language data and further trained on technical content. The model achieves state-of-the-art perf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
64
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(64 citation statements)
references
References 16 publications
0
64
0
Order By: Relevance
“…In Fig. 6, we take datasets sampled from our generative data model, (19) and (23), and map them through our random feature model, (28), for a fixed set of features weights, u, in order to verify that their spectra has the properties discussed in §2.2. We see that the power-law portion of each spectrum is controlled by the minimum of the number of features, N , and the size of the dataset, T .…”
Section: Random Feature Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…In Fig. 6, we take datasets sampled from our generative data model, (19) and (23), and map them through our random feature model, (28), for a fixed set of features weights, u, in order to verify that their spectra has the properties discussed in §2.2. We see that the power-law portion of each spectrum is controlled by the minimum of the number of features, N , and the size of the dataset, T .…”
Section: Random Feature Modelmentioning
confidence: 99%
“…After such a study, a number of follow ups appeared showing even more general applicability and more detailed understanding [20][21][22][23] -and even improved performance scaling with data size from a power law to an exponential falloff with clever pruning [24]. 1 At the same time, autoregressive generative modeling with transformers has continued to be applied to broader AI tasks such as coding [27], quantitative reasoning [28], and even on the suite of computer vision tasks, with the advent of the Vision Transformer (ViT) [29] family of models.…”
Section: Introductionmentioning
confidence: 99%
“…Pre-trained models, such as BERT [19], can be fine-tuned to achieve high performance on downstream tasks such as sentiment analysis. Minerva [20], a recent language model, has shown promising performance on even more complex tasks such as undergraduate-level STEM problems. Pre-training on large amounts of unlabeled sequences also holds a promise for biology.…”
Section: Introductionmentioning
confidence: 99%
“…Machine learning tools designed for text prediction [56,57] can be fine-tuned to "auto-complete" mathematical proofs, given a formal problem statement [34,58], even to the point of generating correct solutions to International Math Olympiad problems [59]. Recently, large language models have demonstrated capabilities in solving chemistry problems [60,61], as well as answering scientific question-and-answer problems invoking quantitative reasoning [62] Formal proofs in science and engineering mathematics may in the future provide useful, high-quality data for artificial intelligences aiming to learn, reason, and discover in science [63,64,65].…”
Section: Discussionmentioning
confidence: 99%