2020
DOI: 10.1109/tvcg.2019.2934631
|View full text |Cite
|
Sign up to set email alerts
|

Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Abstract: 1 2 Recall: 0.81 in the poisoned model 0.90 in the victim model 7 5 4 B A C E D F G 3 G.1 6 G.2 Fig. 1. Reliability attack on spam filters.(1) Poisoning instance #40 has the largest impact on the recall value, which is (2) also depicted in the model overview.(3) There is heavy overlap among instances in the two classes as well the poisoning instances. (4) Instance #40 has been successfully attacked causing a number of innocent instances to have their labels flipped. (5) The flipped instances are very close to … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
46
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 58 publications
(47 citation statements)
references
References 59 publications
0
46
0
1
Order By: Relevance
“…Black box methods are often designed for very specific goals. Ma et al [MXLM20], for example, support the identification of feature combinations that elicit a specific response from a model. Ye et al [YXX*19] enable users to assess and increase the quality of training data labels.…”
Section: Related Workmentioning
confidence: 99%
“…Black box methods are often designed for very specific goals. Ma et al [MXLM20], for example, support the identification of feature combinations that elicit a specific response from a model. Ye et al [YXX*19] enable users to assess and increase the quality of training data labels.…”
Section: Related Workmentioning
confidence: 99%
“…Based on these measures, developers can iteratively identify those features that cause discrimination and remove them from the model. Researchers are also interested in exploring potential vulnerabilities in models that prevent them from being reliably applied to real-world applications [84,91]. Cao et al [84] proposed AEVis to analyze how adversarial examples fool neural networks.…”
Section: Analyzing Training Resultsmentioning
confidence: 99%
“…It then employs a river-based metaphor to show the diverging and merging patterns of the extracted datapaths, which reveal where the adversarial samples mislead the model. Ma et al [91] designed a series of visual representations from overview to detail to reveal how data poisoning will make a model misclassify a specific sample. By comparing the distributions of the poisoned and normal training data, experts can deduce the reason for the misclassification of the attacked sample.…”
Section: Analyzing Training Resultsmentioning
confidence: 99%
“…Security vulnerabilities. When research is conducted in ML, there is always a factor that is not often taken into account at first: “how do we secure ML models from unethical attacks?” An instance of this idea is published by Ma et al [MXLM20], explaining how visualization can assist in avoiding vulnerabilities of adversarial attacks in ML. Specifically, their focus is on how to avoid data poisoning attacks from the models, data instances, features, and local structures perspectives with the use of their VA approach.…”
Section: Discussion and Research Opportunitiesmentioning
confidence: 99%
“…Biologists and doctors, for instance, are interested in being able to compare data structures and receive guidance on where to focus on. Ma et al [MXLM20] employ a multi‐faceted visualization schema intended to aid the analysis of ML experts for the domain of adversarial attacks.…”
Section: In‐depth Categorization Of Trust Against Facets Of Interamentioning
confidence: 99%