2023
DOI: 10.48550/arxiv.2301.09656
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Selective Explanations: Leveraging Human Input to Align Explainable AI

Abstract: While a vast collection of explainable AI (XAI) algorithms have been developed in recent years, they are often criticized for significant gaps with how humans produce and consume explanations. As a result, current XAI techniques are often found to be hard to use and lack effectiveness. In this work, we attempt to close these gaps by making AI explanations selective-a fundamental property of human explanations-by selectively presenting a subset from a large set of model reasons based on what aligns with the rec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 55 publications
0
2
0
Order By: Relevance
“…We argue for these high-stakes tasks, the increase in labeling time and cost is justifiable to a degree (echoing our intent of encouraging moderators to "think slow"). However, we do hope future work could look more into potential ways to improve performance while reducing time through, e.g., selectively introducing explanations on hard examples (Lai et al, 2023). This approach could aid in scaling our framework for everyday use, where the delicate balance between swift annotation and careful moderation is more prominent.…”
Section: Discussionmentioning
confidence: 99%
“…We argue for these high-stakes tasks, the increase in labeling time and cost is justifiable to a degree (echoing our intent of encouraging moderators to "think slow"). However, we do hope future work could look more into potential ways to improve performance while reducing time through, e.g., selectively introducing explanations on hard examples (Lai et al, 2023). This approach could aid in scaling our framework for everyday use, where the delicate balance between swift annotation and careful moderation is more prominent.…”
Section: Discussionmentioning
confidence: 99%
“…Thus, our system must adapt explanations to fulfill users' specific needs [77]. Meanwhile, recent research suggests a few promising directions, such as explanation selection (to ensure preference alignment) [45] and verifiability (to verify the correctness of AI outputs) [17]. Future work can be explored along with these directions to enhance the effectiveness of explanations further.…”
Section: The Mixed Effect Of Explanationsmentioning
confidence: 99%