2023
DOI: 10.1145/3608482
|View full text |Cite
|
Sign up to set email alerts
|

The Privacy Issue of Counterfactual Explanations: Explanation Linkage Attacks

Abstract: Black-box machine learning models are used in an increasing number of high-stakes domains, and this creates a growing need for Explainable AI (XAI). However, the use of XAI in machine learning introduces privacy risks, which currently remain largely unnoticed. Therefore, we explore the possibility of an explanation linkage attack , which can occur when deploying instance-based strategies to find counterfactual explanations. To counter such an attack, we propose k … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 51 publications
(89 reference statements)
0
1
0
Order By: Relevance
“…Several attempts have been made at combining explanations and privacy preservation (Franco et al, 2021 ; Rahman et al, 2021 ; Bárcena et al, 2022 ; Ariffin et al, 2023 ). Some works investigated the impact of XAI on privacy and found that the privacy of models can indeed be compromised, depending on the XAI method used (Zhao et al, 2021 ; Goethals et al, 2023 ; Lucieri et al, 2023 ; Spartalis et al, 2023 ; Yan et al, 2023 ). As a result, methods defending the privacy of explainable models have been proposed (Montenegro et al, 2021 ; Nguyen et al, 2023 ; Pentyala et al, 2023 ).…”
Section: Related Workmentioning
confidence: 99%
“…Several attempts have been made at combining explanations and privacy preservation (Franco et al, 2021 ; Rahman et al, 2021 ; Bárcena et al, 2022 ; Ariffin et al, 2023 ). Some works investigated the impact of XAI on privacy and found that the privacy of models can indeed be compromised, depending on the XAI method used (Zhao et al, 2021 ; Goethals et al, 2023 ; Lucieri et al, 2023 ; Spartalis et al, 2023 ; Yan et al, 2023 ). As a result, methods defending the privacy of explainable models have been proposed (Montenegro et al, 2021 ; Nguyen et al, 2023 ; Pentyala et al, 2023 ).…”
Section: Related Workmentioning
confidence: 99%