2020
DOI: 10.1007/978-3-030-51054-1_6
|View full text |Cite
|
Sign up to set email alerts
|

Deep Generation of Coq Lemma Names Using Elaborated Terms

Abstract: Coding conventions for naming, spacing, and other essentially stylistic properties are necessary for developers to effectively understand, review, and modify source code in large software projects. Consistent conventions in verification projects based on proof assistants, such as Coq, increase in importance as projects grow in size and scope. While conventions can be documented and enforced manually at high cost, emerging approaches automatically learn and suggest idiomatic names in Java-like languages by appl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
11
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1

Relationship

2
3

Authors

Journals

citations
Cited by 9 publications
(11 citation statements)
references
References 51 publications
0
11
0
Order By: Relevance
“…We evaluated an earlier version of ROOSTERIZE using a corpus derived from the MathComp family of Coq projects, finding that the toolchain significantly outperforms strong baselines on automatic metrics [7]. Moreover, we found encouraging results in a qualitative case study where the maintainer of a medium-sized Coq project manually evaluated over 150 name suggestions generated by ROOSTERIZE.…”
Section: Introductionmentioning
confidence: 88%
See 2 more Smart Citations
“…We evaluated an earlier version of ROOSTERIZE using a corpus derived from the MathComp family of Coq projects, finding that the toolchain significantly outperforms strong baselines on automatic metrics [7]. Moreover, we found encouraging results in a qualitative case study where the maintainer of a medium-sized Coq project manually evaluated over 150 name suggestions generated by ROOSTERIZE.…”
Section: Introductionmentioning
confidence: 88%
“…Users should first obtain a model, e.g., by downloading a pre-trained model. The following command downloads the model we pre-trained on our MathComp corpus [7]:…”
Section: A Command Linementioning
confidence: 99%
See 1 more Smart Citation
“…Despite the high accuracy achieved by our preliminary implementation even when using the baseline n-gram model, we believe our spacing prediction (based only on raw token streams) needs significant tuning for practical use. For example, newlines before Qed sentences often get mispredicted, and unlike for name suggestions [3], it is usually inconvenient to inspect more than the top-1 suggestion for spacing. Moreover, for MathComp, we were able to construct, with help from maintainers, a sufficiently large corpus with strict adherence to conventions; for other projects, it may be more challenging, e.g., due to project size or lack of consensus on conventions.…”
Section: Challenges and Future Directionsmentioning
confidence: 99%
“…As a first step, we here outline initial models to learn and suggest space formatting in Coq files, with a preliminary implementation for Coq 8.10, and evaluated using on a corpus based on MathComp 1.9.0 which comprises 164k lines of Coq code from four core projects [3].…”
mentioning
confidence: 99%