2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE) 2021
DOI: 10.1109/icse43902.2021.00040
|View full text |Cite
|
Sign up to set email alerts
|

Traceability Transformed: Generating More Accurate Links with Pre-Trained BERT Models

Abstract: Software traceability establishes and leverages associations between diverse development artifacts. Researchers have proposed the use of deep learning trace models to link natural language artifacts, such as requirements and issue descriptions, to source code; however, their effectiveness has been restricted by availability of labeled data and efficiency at runtime. In this study, we propose a novel framework called TVace BERT (T-BERT) to generate trace links between source code and natural language artifacts.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
72
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 85 publications
(72 citation statements)
references
References 27 publications
0
72
0
Order By: Relevance
“…Another alternative is to define suitable intermediate training tasks. We have found initial evidence of this, and a recent paper adds further evidence, in the context of traceability [83]. However, it used a very similar task and dataset.…”
Section: Intermediate Task Trainingmentioning
confidence: 57%
“…Another alternative is to define suitable intermediate training tasks. We have found initial evidence of this, and a recent paper adds further evidence, in the context of traceability [83]. However, it used a very similar task and dataset.…”
Section: Intermediate Task Trainingmentioning
confidence: 57%
“…Outside of code authoring, transfer learning has been enabled by Transformer-based models in other software artifacts. Lin et al [24] apply pretrained BERT models for learning relationships between issues and commits in a software repos-itory. Sharma et al [25] detect code smells in programming languages where sufficient training data is not available by transferring knowledge from other data-rich languages.…”
Section: Resultsmentioning
confidence: 99%
“…In the future, we plan to consider more diverse information of a post into account, such as the attached pictures, author information, etc. Also, we are interested in applying PTM4Tag on more SQA sites such as Freecode 14 , AskUbuntu 15 , etc., to further evaluate its effectiveness and generalizability. We release our replication package 16 to facilitate future research.…”
Section: Discussionmentioning
confidence: 99%
“…Recent trends in the NLP domain have led to the rapid development of transfer learning. Especially, substantial work has shown that pre-trained language models learn practical and generic language representations which could achieve outstanding performance in various downstream tasks simply by fine-tuning, i.e., without training a new model from scratch [10,15,20]. With proper training manner, the model can effectively capture the semantics of individual words based on their surrounding context and reflect the meaning of the whole sentence.…”
Section: Pre-trained Language Modelsmentioning
confidence: 99%