2022
DOI: 10.48550/arxiv.2201.11227
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Synchromesh: Reliable code generation from pre-trained language models

Abstract: Large pre-trained language models have been used to generate code, providing a flexible interface for synthesizing programs from natural language specifications. However, they often violate syntactic and semantic rules of their output language, limiting their practical usability. In this paper, we propose SYNCHROMESH: a framework for substantially improving the reliability of pre-trained models for code generation. SYNCHROMESH comprises two components. First, it retrieves few-shot examples from a training bank… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 15 publications
(24 citation statements)
references
References 13 publications
0
24
0
Order By: Relevance
“…In contrast to most existing neuro-symbolic reasoning frameworks, e.g., [24], instead of using a pretrained or jointly-trained semantic parser, we introduce the use of large language-to-code models for parsing. Specifically, we use Codex [6,11] with the Synchromesh framework [27]. By specifying only a small number of examples of language input and expected programs, we gain perfect parsing capabilities across unseen categories and re-…”
Section: Semantic Parsermentioning
confidence: 99%
“…In contrast to most existing neuro-symbolic reasoning frameworks, e.g., [24], instead of using a pretrained or jointly-trained semantic parser, we introduce the use of large language-to-code models for parsing. Specifically, we use Codex [6,11] with the Synchromesh framework [27]. By specifying only a small number of examples of language input and expected programs, we gain perfect parsing capabilities across unseen categories and re-…”
Section: Semantic Parsermentioning
confidence: 99%
“…Usually this is accomplished by semantic retrieval using the testing input as the query. Previous works have studied how to build sentence-level retrievers (Poesia et al, 2022;Rubin et al, 2021). Our work goes beyond sentences and studies how to retrieve dialogues.…”
Section: Dialogue Retrievermentioning
confidence: 99%
“…The first is similarity-based retrieval. Poesia et al (2022) and Das et al (2021) define a similarity metric between semantic parsing results and use this similarity as the training objective for the retriever.…”
Section: Related Workmentioning
confidence: 99%
“…The study also suggested the future direction of this research domain to improve automatic code generation using natural language by analyzing the current trend of approaches. [30] proposes Synchromesh, a framework to improve the coding reliability of pre-trained models. Using Target Similarity Tuning, this framework retrieves a few-shot example from a training bank.…”
Section: Ptm For Code Generation and Understandingmentioning
confidence: 99%