Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering 2018
DOI: 10.1145/3238147.3238190
|View full text |Cite
|
Sign up to set email alerts
|

Neural-machine-translation-based commit message generation: how far are we?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
177
3

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 206 publications
(181 citation statements)
references
References 49 publications
1
177
3
Order By: Relevance
“…This is opposite to the conclusion achieved by Liu et al [23]. One possible reason is that the tasks between ours and Liu et al's [23] are different, i.e., Liu et al aim at producing texts based on code, while we focus on generating texts for dialogues and modeling code is different from modeling dialogue texts [53], [54]. The higher BLEU-4 score of the proposed RRGen model than that of the NMT model explains that the response generated by the RRGen model is more similar to developers' response than the response generated by the NMT model.…”
Section: Evaluation Using An Automatic Metriccontrasting
confidence: 99%
See 1 more Smart Citation
“…This is opposite to the conclusion achieved by Liu et al [23]. One possible reason is that the tasks between ours and Liu et al's [23] are different, i.e., Liu et al aim at producing texts based on code, while we focus on generating texts for dialogues and modeling code is different from modeling dialogue texts [53], [54]. The higher BLEU-4 score of the proposed RRGen model than that of the NMT model explains that the response generated by the RRGen model is more similar to developers' response than the response generated by the NMT model.…”
Section: Evaluation Using An Automatic Metriccontrasting
confidence: 99%
“…All the three metrics are rated on a 1-5 scale (5 for fully satisfying the rating scheme, 1 for completely not satisfying the rating scheme, and 3 for the borderline cases), since a 5-point scale is widely used in prior software engineering studies [3], [23], [58]. Besides the three metrics, each participant is asked to rank responses generated by the three tools and those from developers based on their preference.…”
Section: B Survey Designmentioning
confidence: 99%
“…After applying several filters, Jiang et al obtain a set of 32 commits that could be used by NMT1 algorithm [7]. Liu et al made a cleaned version of this dataset by removing the noisy messages [11] (cleaned dataset). The noisy messages are categorized into two categories: (1) The messages generated by development tools (called bot messages).…”
Section: Datasetmentioning
confidence: 99%
“…A higher BLEU_4 score for a generated message shows that the message is more similar to the one written by the human developer. Liu et al [11] also consider this metric as a textual similarity distance metric used in the second step of NNGen described above.…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, the text content (e.g., class/method/variable names, comments) of programs, which is used in tasks like code recommendation and program comprehension, is often not expressed in a consistent and normative way. For example, a recent study by Liu et al [8] revealed that a large part of the commit messages that are used as references are noisy, for example they may be bot messages generated by tools or trivial messages containing little or redundant information.…”
mentioning
confidence: 99%