2021
DOI: 10.48550/arxiv.2102.06203
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Proof Artifact Co-training for Theorem Proving with Language Models

Abstract: Labeled data for imitation learning of theorem proving in large libraries of formalized mathematics is scarce as such libraries require years of concentrated effort by human specialists to be built. This is particularly challenging when applying large Transformer language models to tactic prediction, because the scaling of performance with respect to model size is quickly disrupted in the data-scarce, easily-overfitted regime. We propose PACT (Proof Artifact Co-Training), a general methodology for extracting a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 7 publications
(20 citation statements)
references
References 27 publications
0
18
0
Order By: Relevance
“…In this section, in order to study baseline performances associated with existing systems, we report pass rates achieved by GPT-f [12] applied to Metamath, GPT-f/PACT [12,7] applied to Lean as well as a baseline prover implemented in Lean denoted as the tidy baseline. Pass rates are reported as Pass@N where N is the number of proof search attempts per statement.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…In this section, in order to study baseline performances associated with existing systems, we report pass rates achieved by GPT-f [12] applied to Metamath, GPT-f/PACT [12,7] applied to Lean as well as a baseline prover implemented in Lean denoted as the tidy baseline. Pass rates are reported as Pass@N where N is the number of proof search attempts per statement.…”
Section: Methodsmentioning
confidence: 99%
“…They also proposed a methodology to train a value function to further guide proof search, achieving a 56.22% pass rate on the held-out test set. Large language models were applied to Lean by Han et al [7]. They created an environment around the Lean prover targeted to machine learning and propose a dataset extracted from low level proof artifacts that is shown to boost performance when used as a self-supervised co-training objective.…”
Section: Neural Theorem Provingmentioning
confidence: 99%
See 2 more Smart Citations
“…For automated theorem proving and mathematical reasoning, various datasets and accompanying approaches have been proposed (Kaliszyk et al, 2017;Bansal et al, 2019) including more recent work with equational logic (Piepenbrock et al, 2021) and language models (Rabe et al, 2020;Han et al, 2021).…”
Section: Related Workmentioning
confidence: 99%