2020
DOI: 10.1016/j.cose.2020.101859
|View full text |Cite
|
Sign up to set email alerts
|

We need to talk about antiviruses: challenges & pitfalls of AV evaluations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 32 publications
(19 citation statements)
references
References 7 publications
0
18
0
Order By: Relevance
“…A common, automated method for labeling malware is scanning with a single antivirus. However, antivirus signatures are frequently incorrect and can lack family information entirely [22,23]. Furthermore, antivirus vendors often use different names for the same malware family, which we refer to as aliases.…”
Section: Malware Dataset Labeling Strategiesmentioning
confidence: 99%
See 1 more Smart Citation
“…A common, automated method for labeling malware is scanning with a single antivirus. However, antivirus signatures are frequently incorrect and can lack family information entirely [22,23]. Furthermore, antivirus vendors often use different names for the same malware family, which we refer to as aliases.…”
Section: Malware Dataset Labeling Strategiesmentioning
confidence: 99%
“…However, all approaches to malware grouping run into different versions of the same set of problems. Malware is written by an active adversary who attempts to evade and mislead analysts, including complex code obfuscations and misdirection via code theft to slow analysts and thwart automation [27,28,29,30]. Even standard countermeasures such as packing, which hides the original source code of a program from static analysis, are poorly understood and difficult to circumvent [31].…”
Section: Malware Dataset Labeling Strategiesmentioning
confidence: 99%
“…Only recently have some of these AV biases been categorized and described. Botacin et al (2020) has shown that the detection rate of AV products may vary by country (i.e., is this malware global, or country specific, in its proliferation), executable type (e.g., COM files vs. DLL), and family type (e.g., ransomware vs trojans). These biases will naturally be embedded into any model and evaluation built from labels that are AV produced.…”
Section: Data Collection Challengesmentioning
confidence: 99%
“…These biases will naturally be embedded into any model and evaluation built from labels that are AV produced. Further, using older files to try and maximize confidence is not a guaranteed workaround, since AV engines will have label regressions over time, where they stop detecting sufficiently old files as malicious (Botacin et al, 2020).…”
Section: Data Collection Challengesmentioning
confidence: 99%
“…The global vs. local dichotomy is not a new observation. Previous studies have shown that other AV products can have different FP and false negative (FN) rates for regional malware (e.g., United States vs. Brazil) and malware types (e.g., Trojan vs. ransomware) [2]. This can cause excessive false alerts, which users find unacceptable and may lead them to abandon the malware detectors.…”
Section: Introductionmentioning
confidence: 99%