2020 IEEE 20th International Conference on Software Quality, Reliability and Security (QRS) 2020
DOI: 10.1109/qrs51102.2020.00071
|View full text |Cite
|
Sign up to set email alerts
|

A Practical, Principled Measure of Fuzzer Appeal: A Preliminary Study

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 27 publications
0
3
0
Order By: Relevance
“…al [17], e.g., using 10 or more runs of 24 hours each in experimental trials. We will make every effort to identify and protect against the usual threats to validity in fuzzing experiments, by using a range of benchmark subjects and avoiding pitfalls such as measuring only crash counts bucketed crashes, rather than making an effort to identify actual distinct faults [9] (or using only crashes, not crashes and code coverage results).…”
Section: Proposed Evaluationmentioning
confidence: 99%
“…al [17], e.g., using 10 or more runs of 24 hours each in experimental trials. We will make every effort to identify and protect against the usual threats to validity in fuzzing experiments, by using a range of benchmark subjects and avoiding pitfalls such as measuring only crash counts bucketed crashes, rather than making an effort to identify actual distinct faults [9] (or using only crashes, not crashes and code coverage results).…”
Section: Proposed Evaluationmentioning
confidence: 99%
“…Various proposals for handling these problems have been posed; for instance Gavrilov et al [16] propose using multiple versions of a program and detecting differences exposed by fuzzers as a richer evaluation measure (they also provide a more in-depth examination of the weakness of the coverage and seeded-bug measures discussed above). However, such an approach requires the availability of multiple versions of a program, and is not fundamentally tied to measuring bug detection (if outputs differ but are not flagged as faulty, this is seen as a difference in appeal, regardless of oracle strength).…”
Section: Introductionmentioning
confidence: 99%
“…Unfortunately, out of the numerous fuzzing evaluation research papers available [15], [18], [19], [20], [21], [22], [16] none recommends the use of mutation analysis for fuzzing. Indeed, none of the papers we examined [1], [23], [24], [25] actually used mutation score as a means of evaluation.…”
Section: Introductionmentioning
confidence: 99%