The World Wide Web Conference 2019
DOI: 10.1145/3308558.3313632
|View full text |Cite
|
Sign up to set email alerts
|

CoaCor: Code Annotation for Code Retrieval with Reinforcement Learning

Abstract: To accelerate software development, much research has been performed to help people understand and reuse the huge amount of available code resources. Two important tasks have been widely studied: code retrieval, which aims to retrieve code snippets relevant to a given natural language query from a code base, and code annotation, where the goal is to annotate a code snippet with a natural language description. Despite their advancement in recent years, the two tasks are mostly explored separately. In this work,… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
111
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 92 publications
(113 citation statements)
references
References 57 publications
(145 reference statements)
1
111
1
Order By: Relevance
“…For example, the BLEU score [24] of the code summarization model in CoaCor is not satisfactory though it improves the performance of existing code retrieval models significantly. Different from what claimed in [34], we respectively argue that generating summaries close to human-provided queries is naturally fit to code retrieval. The compromise of BLEU score, which represents the similarity between the generated summaries and human-written ones, can be avoided if we can model the inner connection between the two tasks better.…”
Section: Introductioncontrasting
confidence: 79%
See 2 more Smart Citations
“…For example, the BLEU score [24] of the code summarization model in CoaCor is not satisfactory though it improves the performance of existing code retrieval models significantly. Different from what claimed in [34], we respectively argue that generating summaries close to human-provided queries is naturally fit to code retrieval. The compromise of BLEU score, which represents the similarity between the generated summaries and human-written ones, can be avoided if we can model the inner connection between the two tasks better.…”
Section: Introductioncontrasting
confidence: 79%
“…Two VAEs are trained jointly to reconstruct their inputs as much as possible with regularization that captures the closeness between the latent variables of code and description, which will be used for measuring similarity. Similarly, Yao et al [34] constructed a neural networkbased code annotation model to describe the functionality of an entire code snippet. It produces meaningful words that can be used for code retrieval where these words and a natural language query are projected into a vector space to measure the cosine similarity between them.…”
Section: Related Work 61 Code Retrievalmentioning
confidence: 99%
See 1 more Smart Citation
“…Several of the latest code search techniques that find code given a natural language query rely on machine learning techniques (e.g.,NCS [10], DeepCS [8], UNIF [38], MMAN [39], TBCAA [40], and CoaCor [41]). NCS proposes an enhanced word embedding for a natural language query [10].…”
Section: Code Search Systemsmentioning
confidence: 99%
“…This technique aims to capture semantics by incorporating API call information into ASTs which is otherwise abstracted as the same AST node type. CoaCor [41] uses reinforcement learning to build a code annotation framework for effective code retrieval. By generating detailed code annotations using multiple keywords, CoaCor improves the performance of existing code retrieval models.…”
Section: Code Search Systemsmentioning
confidence: 99%