27th International Conference on Intelligent User Interfaces 2022
DOI: 10.1145/3490099.3511160
|View full text |Cite
|
Sign up to set email alerts
|

Intuitively Assessing ML Model Reliability through Example-Based Explanations and Editing Model Inputs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(13 citation statements)
references
References 45 publications
1
12
0
Order By: Relevance
“…Example metrics that participants proposed included error rates for determinations that a piece of evidence was inconclusive and ranges for the possible values that false positive software outputs could take on (Section 5.2.3). Together, these findings echo gaps between evaluation design and real-world use contexts highlighted in prior work (e.g., [42,72,81,92]), and, importantly, demonstrate the valuable insights that public defenders develop through their everyday encounters with CFS in the U.S. criminal legal system, further motivating growing efforts in HCI to engage downstream stakeholders in designing performance evaluations of AI systems (e.g., [28,55,76,81]).…”
Section: Contextualize Design Of Performance Evaluations In Real Worl...supporting
confidence: 67%
See 3 more Smart Citations
“…Example metrics that participants proposed included error rates for determinations that a piece of evidence was inconclusive and ranges for the possible values that false positive software outputs could take on (Section 5.2.3). Together, these findings echo gaps between evaluation design and real-world use contexts highlighted in prior work (e.g., [42,72,81,92]), and, importantly, demonstrate the valuable insights that public defenders develop through their everyday encounters with CFS in the U.S. criminal legal system, further motivating growing efforts in HCI to engage downstream stakeholders in designing performance evaluations of AI systems (e.g., [28,55,76,81]).…”
Section: Contextualize Design Of Performance Evaluations In Real Worl...supporting
confidence: 67%
“…For instance, test inputs may not be sufficiently representative of real-world settings [53,72], and performance metrics may not align with users' preferences and perceptions of ideal model performance [47,53,66]. To address this gap, a growing body of work in HCI aims to design performance evaluations grounded in downstream deployment contexts and the needs and goals of downstream stakeholders (e.g., [18,57,80,81]). This typically involves exploring users' domain-specific information needs [19,46], directly working with downstream stakeholders to collaboratively design evaluation datasets and metrics [80], and designing tools that allow users to specify their own test datasets and performance metrics [18,27,28,55,81].…”
Section: Designing Performance Evaluationsmentioning
confidence: 99%
See 2 more Smart Citations
“…Of course, there is an indisputable trade-off cost to building interactive tools and interfaces for model performance. There is continued research being done to support effective and accessible interaction for web-based tools [55], computational notebooks [62], and visualization tools [58] that are practical for data scientists to use. Adding interaction also introduces costs for the audience (e.g., false discoveries) during a presentation [38].…”
Section: Presentation Modalitiesmentioning
confidence: 99%