2018
DOI: 10.1007/978-3-030-00063-9_15
|View full text |Cite
|
Sign up to set email alerts
|

Black-Box Model Explained Through an Assessment of Its Interpretable Features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1
1

Relationship

2
4

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 6 publications
0
11
0
Order By: Relevance
“…An exhaustive overview of the existing XAI techniques for NLP models, applied in different contexts, such as social networks, medical, and cybersecurity, is presented in [34]. Many works exploit feature-perturbation strategies in the explanation process, analyzing the model reactions to produce prediction-local explanations, like in [2,28,32,35,40,48]. This straightforward idea is very powerful but requires a careful selection of the input features to be perturbed.…”
Section: Domain-specific Approachesmentioning
confidence: 99%
See 3 more Smart Citations
“…An exhaustive overview of the existing XAI techniques for NLP models, applied in different contexts, such as social networks, medical, and cybersecurity, is presented in [34]. Many works exploit feature-perturbation strategies in the explanation process, analyzing the model reactions to produce prediction-local explanations, like in [2,28,32,35,40,48]. This straightforward idea is very powerful but requires a careful selection of the input features to be perturbed.…”
Section: Domain-specific Approachesmentioning
confidence: 99%
“…T-EBAnO proposes a new local and global explanation process for state-of-the-art deep NLP models. By exploiting a perturbation-based strategy similar to that described in [48], which was successfully tailored to image data, T-EBAnO fills in the gap of missing customized solutions for explaining deep NLP models by introducing a totally redesigned architecture and experimental section. Specifically, we introduce (i) a novel feature extraction process specifically tailored to textual data and deep natural language models, (ii) new perturbation strategies, (iii) an improved version of the index proposed in [48] able to quantify the influence of the input feature over local predictions tested in a new domain (NLP), and (iv) novel class-based global explanations, besides extending the experiments to new models and use cases, and presenting a human evaluation of the exploited index and proposed explanations.…”
Section: Task-specific Approachesmentioning
confidence: 99%
See 2 more Smart Citations
“…For instance, the work in [62] uses a model-agnostic perturbation-based technique that produces local explanations by training a simpler interpretable model that approximates the prediction locally. Instead, the work in [63] produces local explanations of deep convolutional neural networks by a process of perturbation over interpretable features extracted from the hypercolumns. The authors of [64] propose a model-agnostic explainability technique based on a game-theoretic approach that iteratively removes combinations of inputs features to measure the features' importance.…”
Section: Explainable Artificial Intelligencementioning
confidence: 99%