Proceedings of the 1st Workshop on Natural Language Processing for Programming (NLP4Prog 2021) 2021
DOI: 10.18653/v1/2021.nlp4prog-1.5
|View full text |Cite
|
Sign up to set email alerts
|

CoTexT: Multi-task Learning with Code-Text Transformer

Abstract: We present CoTexT, a pre-trained, transformerbased encoder-decoder model that learns the representative context between natural language (NL) and programming language (PL). Using self-supervision, CoTexT is pretrained on large programming language corpora to learn a general understanding of language and code. CoTexT supports downstream NL-PL tasks such as code summarizing/documentation, code generation, defect detection, and code debugging. We train CoTexT on different combinations of available PL corpus inclu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
43
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 70 publications
(43 citation statements)
references
References 13 publications
0
43
0
Order By: Relevance
“…PLBART [3] employed a denoising objective over code and natural language via token masking, token deletion, and token infilling as noising strategies. CoText [54] was build on top of T5 with a special focus on multi-task learning over multiple programming languages. Another model variant was introduced by Fried et al [27,InCoder].…”
Section: Pre-trained Language Models For Programming Languagementioning
confidence: 99%
See 1 more Smart Citation
“…PLBART [3] employed a denoising objective over code and natural language via token masking, token deletion, and token infilling as noising strategies. CoText [54] was build on top of T5 with a special focus on multi-task learning over multiple programming languages. Another model variant was introduced by Fried et al [27,InCoder].…”
Section: Pre-trained Language Models For Programming Languagementioning
confidence: 99%
“…to the Biomedical [37,46,9], Legal [14], Cyber Security [2], and Finance [6] domains, while simultaneously expanding to include signals from modalities other than natural language, e.g. Vision [65,18,15,13,64,25], Proteins [11], Time Series [71,72,55] and Code [42,70,26,31,54].…”
Section: Introductionmentioning
confidence: 99%
“…A short time ago, BARTpho , a large pretrained sequence-to-sequence model for Vietnamese inheriting BART style (Lewis et al, 2019), demonstrated the effectiveness of pretrained language models on Vietnamese abstractive summarization. Nevertheless, there are some past works that have shown that T5 architecture (Raffel et al, 2019) might outperform BART in some aspects (i.e., (Phan et al, 2021a)). Inspired by that, we propose ViT5, trained on the Vietnamese monolingual subset of CC100, following the architecture and training methodology in Raffel et al (2019).…”
Section: Introductionmentioning
confidence: 99%
“…Encouraged by the excellent recent results of data-driven APR approaches [15,4,30,21,5], in this work we wanted to investigate whether is it worth ne-tuning the GPT-2 model. At the time of writing this article the top three approaches are CoTexT [22], PLBART [1] and DeepDebug [5]. Although none of these approaches use the GPT-2 model, their operating principle is similar.…”
Section: Introductionmentioning
confidence: 99%