2021
DOI: 10.1007/s10664-021-09946-8
|View full text |Cite
|
Sign up to set email alerts
|

Are datasets for information retrieval-based bug localization techniques trustworthy?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 74 publications
0
9
0
Order By: Relevance
“…Kochhar et al [30] investigate potential biases in datasets used for IRFL experiments, showing that the inclusion of bug reports where the reporter already localizes the fault causes a bias and distorts retrieval performance results. Kim and Lee [23] point out issues regarding representability and generalizability of benchmarks used in IRFL. Hirsch and Hofer [27] survey public bug localization benchmarks and identify issues in data availability, data quality, and reproducibility.…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Kochhar et al [30] investigate potential biases in datasets used for IRFL experiments, showing that the inclusion of bug reports where the reporter already localizes the fault causes a bias and distorts retrieval performance results. Kim and Lee [23] point out issues regarding representability and generalizability of benchmarks used in IRFL. Hirsch and Hofer [27] survey public bug localization benchmarks and identify issues in data availability, data quality, and reproducibility.…”
Section: Related Workmentioning
confidence: 99%
“…Along with MRR and TopN, MAP is one of the most popular performance metrics used in IRFL research [23]. The Retrieved documents in the context of IRFL experiments is the set of bug fix locations in the codebase, most often presented at file level, and commonly referred to as ground truth in IRFL.…”
Section: Definitionsmentioning
confidence: 99%
See 3 more Smart Citations