Proceedings of the 30th IEEE/ACM International Conference on Program Comprehension 2022
DOI: 10.1145/3524610.3527905
|View full text |Cite
|
Sign up to set email alerts
|

Learning to represent programs with heterogeneous graphs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 42 publications
(23 citation statements)
references
References 31 publications
0
23
0
Order By: Relevance
“…A crucial aspect of deep learning for code assistance is the choice of representation and model architecture. Recent approaches have explored grammar-based representations [1], graph-based representations [2], and hybrid models that integrate symbolic reasoning with deep learning [3]. These methods have shown promise in tasks such as code completion, repair, refactoring, and optimization.…”
Section: Deep Learning Methods For Code 21 Representations and Model ...mentioning
confidence: 99%
See 1 more Smart Citation
“…A crucial aspect of deep learning for code assistance is the choice of representation and model architecture. Recent approaches have explored grammar-based representations [1], graph-based representations [2], and hybrid models that integrate symbolic reasoning with deep learning [3]. These methods have shown promise in tasks such as code completion, repair, refactoring, and optimization.…”
Section: Deep Learning Methods For Code 21 Representations and Model ...mentioning
confidence: 99%
“…Grammar-based representations utilize the syntactic structure of programming languages to build deep learning models capable of understanding and generating code. One approach involves the use of Abstract Syntax Trees (ASTs) [1], which provide a hierarchical representation of code, capturing its structure and semantics. AST-based models have demonstrated success in tasks such as code completion and bug detection.…”
Section: Grammar-based Representationsmentioning
confidence: 99%
“…We also think that the source code before and after source code modeling is also symmetric. There exists a lot of excellent literature on source code analysis [46][47][48][49][50][51][52][53][54][55][56][57][58], and we have shown the representative source code models from the past five years in Table 1. From Table 1, we know that scholars have considered structural information (e.g., AST, API) in recent years, instead of just treating the source code as pure text.…”
Section: Source Code Modelingmentioning
confidence: 99%
“…According to the different representations of source code, we divide the source code modeling into four types: Token-based, Tree-based, Graph-based, and other source code modeling. [40] IJCAI AST CCD LSTM [6] ICPC AST SCS Seq2seq [17] IJCAJ × (API, comments) (API, code, comments) SCS LSTM [19] ASE AST, sourcecode SCS Bi-LSTM [43] ICLR AST, (token, path, token) PF, SCS HOPE [50] MSR identifier, AST, CFG, Bytecode CCD RNN [51] ICLR variable/statetrace PR Word2vec [53] ESEC abstractedsymbolictraces ECM GGNN [55] ICLR AST, PDG PF MLP [62] ASE × tokens SCS RNN [63] AAAI AST SCS, SCC GRU [22] ICSE text, ASTnodetokens SCS Bi-LSTM [46] ICSE AST, ST-trees SCC, CCD Bi-LSTM [64] POPL AST, (token, path, token) PF Bi-LSTM [65] ASE text SCS Bi-LSTM [23] ICSE × AST, codesequence SCS BERT [24] Access functionalkeywords SCS Transformer [25] arXiv × comments, code SCS, CR Transformer [26] ACL × AST SCS GRU [44] ESE tokens, AST SCS Regularizer [56] IST AST SCG GRU [66] JCRD (code, API, comments), (function, comments) SCS GRU [67] ACL text, ASTnodetokens SCS GNN [68] arXiv AST, context SCS Seq2Seq [69] ICPC × (seq, comment), (context, comment) SCS API2Com [70] ICPC AST, API, seq SCS…”
Section: Source Code Modelingmentioning
confidence: 99%
“…Nevertheless, employing the resulting retrained models (henceforth PTM-Cs) for SE tasks is not ideal, as there are code-specific characteristics that may not be properly taken into account by these models, such as the syntactic [17], [18] and semantic structures [19] inherent in source code [20]. Consequently, SE researchers have developed a number of pre-trained models of source code (henceforth CodePTMs) that take into account code-specific characteristics in the past few years [21]- [26].…”
Section: Introductionmentioning
confidence: 99%