2021
DOI: 10.48550/arxiv.2108.07732
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Program Synthesis with Large Language Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
238
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 96 publications
(240 citation statements)
references
References 0 publications
2
238
0
Order By: Relevance
“…Left-to-Right Language Models (Figure 2, left) Auto-regressive, Left-to-right LMs, predict the probability of a token given the previous tokens. In code modeling, CodeGPT (124M) (Lu et al, 2021), CodeParrot (1.5B) (Tunstall et al, 2022), GPT-Neo (2.7B) (Black et al, 2021), GPT-J (6B) (Wang & Komatsuzaki, 2021), Codex (12B) (Chen et al, 2021), GPT-NeoX (20B) (Black et al, 2022), and Google's (137B) (Austin et al, 2021) belong to this category. The left-to-right nature of these models makes them highly useful for program generation tasks, such as code completion.…”
Section: Pretraining Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Left-to-Right Language Models (Figure 2, left) Auto-regressive, Left-to-right LMs, predict the probability of a token given the previous tokens. In code modeling, CodeGPT (124M) (Lu et al, 2021), CodeParrot (1.5B) (Tunstall et al, 2022), GPT-Neo (2.7B) (Black et al, 2021), GPT-J (6B) (Wang & Komatsuzaki, 2021), Codex (12B) (Chen et al, 2021), GPT-NeoX (20B) (Black et al, 2022), and Google's (137B) (Austin et al, 2021) belong to this category. The left-to-right nature of these models makes them highly useful for program generation tasks, such as code completion.…”
Section: Pretraining Methodsmentioning
confidence: 99%
“…These models excel at useful downstream tasks like code completion (Raychev et al, 2014) and synthesizing code from natural language descriptions (Desai et al, 2016). The current state-of-the-art large language models for code, such as Austin et al (2021), have shown significant progress for AI-based programming assistance. Most notably, one of the largest of these models, Codex (Chen et al, 2021) has been deployed in the real-world production tool GitHub Copilot 1 , as an in-IDE developer assistant that automatically generates code based on the user's context.…”
Section: Introductionmentioning
confidence: 99%
“…BLEU score is computed as the overlapping fraction of n-grams between the machine-generated text and the reference text. The metric has however been shown to not be a reliable measure for source code Allamanis et al, 2018;Austin et al, 2021). Computational Accuracy @ k (CA@k): Recent work in code synthesis has adopted the CA@k metric (Austin et al, 2021;Roziere et al, 2020) to evaluate code generation models.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…The metric has however been shown to not be a reliable measure for source code Allamanis et al, 2018;Austin et al, 2021). Computational Accuracy @ k (CA@k): Recent work in code synthesis has adopted the CA@k metric (Austin et al, 2021;Roziere et al, 2020) to evaluate code generation models. To compute CA@k, k samples are generated from the model, and the problem is considered solved if any of the generated k samples pass the unit tests associated with the problem.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…Neural models originally developed for natural language processing show promising performance for modeling computer programming languages (Chen et al, 2021;Austin et al, 2021). This has led to a number of interesting applications, and deep-learning models are now successfully and routinely applied in tools that assist developers in writing and understanding programs and code.…”
Section: Introductionmentioning
confidence: 99%