2022
DOI: 10.48550/arxiv.2207.08466
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

What does Transformer learn about source code?

Abstract: In the field of source code processing, the transformer-based representation models have shown great powerfulness and have achieved state-of-the-art (SOTA) performance in many tasks. Although the transformer models process the sequential source code, pieces of evidence show that they may capture the structural information (e.g., in the syntax tree, data flow, control flow, etc.) as well. We propose the aggregated attention score, a method to investigate the structural information learned by the transformer. We… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 12 publications
0
1
0
Order By: Relevance
“…They find that CAT-scores for source code models are correlated with their performance on code summarization, and that the CAT-scores vary per layer and per language, with a tendency for higher scores in the earlier layers. Zhang et al [43] define a similar metric over the attention matrix, an aggregated attention score, with which they can derive a graph of relationships between tokens.…”
Section: Related Workmentioning
confidence: 99%
“…They find that CAT-scores for source code models are correlated with their performance on code summarization, and that the CAT-scores vary per layer and per language, with a tendency for higher scores in the earlier layers. Zhang et al [43] define a similar metric over the attention matrix, an aggregated attention score, with which they can derive a graph of relationships between tokens.…”
Section: Related Workmentioning
confidence: 99%