2022
DOI: 10.48550/arxiv.2205.12615
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Autoformalization with Large Language Models

Abstract: Autoformalization is the process of automatically translating from natural language mathematics to formal specifications and proofs. A successful autoformalization system could advance the fields of formal verification, program synthesis, and artificial intelligence. While the long-term goal of autoformalization seemed elusive for a long time, we show large language models provide new prospects towards this goal. We make the surprising observation that LLMs can correctly translate a significant portion (25.3%)… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(10 citation statements)
references
References 20 publications
0
10
0
Order By: Relevance
“…); further sub-types include declarations combined with assumptions, such as "Let k be an integer such that n = 2k" (which are existentially loaded and are in need of verification) and justified claims ("Since x = 3(a + b), x is a multiple of 3"). The sentences 12 written in this CNL are converted into an internal list format whose crucial ingredients are a list of the of variables occuring in the sentence, its type (assumption, declaration, claim, annotation,...) and its actual content (which can be empty, as in the case of annotations). Thus, the sentence "Therefore, x is even" would be translated as…”
Section: The Diproche Cnl and The Internal List Formatmentioning
confidence: 99%
See 1 more Smart Citation
“…); further sub-types include declarations combined with assumptions, such as "Let k be an integer such that n = 2k" (which are existentially loaded and are in need of verification) and justified claims ("Since x = 3(a + b), x is a multiple of 3"). The sentences 12 written in this CNL are converted into an internal list format whose crucial ingredients are a list of the of variables occuring in the sentence, its type (assumption, declaration, claim, annotation,...) and its actual content (which can be empty, as in the case of annotations). Thus, the sentence "Therefore, x is even" would be translated as…”
Section: The Diproche Cnl and The Internal List Formatmentioning
confidence: 99%
“…At first, the experiences reported in Avigad et al [1] with using large language models for autoformalization appear to be discouraging for this plan: Only about 11 percent of the natural language inputs were formalized correctly ( [1], p. 3). Much better results were reported in [12], where more than 25 percent of the natural language inputs (which were problems for math competitions) were translated correctly into Isabelle (ibid., p. 1). Still, for reliably checking even a simple natural language argument consisting of typically more than 10 sentences with sufficient reliability to be of didactical use to beginner's students, anything considerably below a hundred percent is not good enough.…”
Section: Introductionmentioning
confidence: 96%
“…LLMs are Transformers [43], which is the state of the art neural architecture for natural language proccessing. Additionally, Transformers have shown remarkable performance when being applied to classical problems in verification (e.g., [19,41,26,9]), reasoning (e.g., [28,51]), as well as the auto-formalization [36] of mathematics and formal specifications (e.g., [50,20,22]).…”
Section: Large Language Modelsmentioning
confidence: 99%
“…The term autoformalization (Wang et al, 2018;Szegedy, 2020) has been coined for tasks of translating between natural language and formal specifications or proofs. Closest to our work is a very recent, independently developed, effort in translating between natural language and formal proofs using very large language models (Wu et al, 2022).…”
Section: Related Workmentioning
confidence: 99%