2020 IEEE Pacific Visualization Symposium (PacificVis) 2020
DOI: 10.1109/pacificvis48177.2020.1031
|View full text |Cite
|
Sign up to set email alerts
|

Interactive Attention Model Explorer for Natural Language Processing Tasks with Unbalanced Data Sizes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…We observe heterogeneous visual encodings adopted to support feature attributions. The most used ones are heatmaps for local post hoc feature attributions on images (Figure 7b) [vdBCR*20, HSL*21, HJZ*21, ZZM16, HCC*20, WGZ*19, SW17, CBN*20, JVW20] and text [CHS20, CGR*17, ŠSE*21, JTH*21]; matrices [WONM18, DWB21, JKV*22, PCN*19, LLL*19], node‐link diagrams [JTH*21, Vig19, JCM20, LLL*19] and custom Sankey diagrams [DWSZ20, PCN*19, HSG20, MSHB22] for self‐explainable attentive models; bar charts[WWM20, PCN*19] or averaged inputs [WGYS18, WGSY19] for global feature attribution; and enhanced line [SMM*19, CWGvW19, LYY*20, SWJ*20, ŠSE*21], area chart [KCK*19] or bar chart [MXC*20, KCK*19, SWJ*20, WWM20] for post hoc approaches to sequential data. Among them, systems that support the analysis of attentive models, and in particular of Transformers, employ the most complex and novel visualization techniques (Figure 7a) such as radial layouts [WTC21, DWB21] or grid ones [WTC21, DWB21, ŠSE*21]: these systems must show the flow of attention weights across multiple layers simultaneously to help the user understand the most important features.…”
Section: Papers Categorizationmentioning
confidence: 99%
See 1 more Smart Citation
“…We observe heterogeneous visual encodings adopted to support feature attributions. The most used ones are heatmaps for local post hoc feature attributions on images (Figure 7b) [vdBCR*20, HSL*21, HJZ*21, ZZM16, HCC*20, WGZ*19, SW17, CBN*20, JVW20] and text [CHS20, CGR*17, ŠSE*21, JTH*21]; matrices [WONM18, DWB21, JKV*22, PCN*19, LLL*19], node‐link diagrams [JTH*21, Vig19, JCM20, LLL*19] and custom Sankey diagrams [DWSZ20, PCN*19, HSG20, MSHB22] for self‐explainable attentive models; bar charts[WWM20, PCN*19] or averaged inputs [WGYS18, WGSY19] for global feature attribution; and enhanced line [SMM*19, CWGvW19, LYY*20, SWJ*20, ŠSE*21], area chart [KCK*19] or bar chart [MXC*20, KCK*19, SWJ*20, WWM20] for post hoc approaches to sequential data. Among them, systems that support the analysis of attentive models, and in particular of Transformers, employ the most complex and novel visualization techniques (Figure 7a) such as radial layouts [WTC21, DWB21] or grid ones [WTC21, DWB21, ŠSE*21]: these systems must show the flow of attention weights across multiple layers simultaneously to help the user understand the most important features.…”
Section: Papers Categorizationmentioning
confidence: 99%
“…Examples of added information are the attribution scores’ magnitude, which is encoded using colours, size [JCM20], opacity [WSP*21] or just its value, or bounding boxes [HSL*21, JKV*22, CBN*20], which highlight the most important region over images. While sorting [CHS20, MXC*20, KCK*19, Vig19, PCN*19] and filtering [WONM18, DWSZ20, JKV*22, JTH*21] by attribution scores capabilities are quite common to ease the data exploration and reduce the visual clutters, some works provide additional tools for a deeper understanding. For example, feature removal interactions guided by local attributions [HSL*21, HJZ*21, KCK*19] (e.g.…”
Section: Papers Categorizationmentioning
confidence: 99%
“…In this respect, tedious labeling of data set will be eliminated; thus, large unlabeled bug repositories can be utilized to train models in an unsupervised fashion for improved generalization. Another model that could be used to improve imbalance learning is the attention mechanism (Dong, Wu, Song and Zhang, 2020).…”
Section: Bug Localizationmentioning
confidence: 99%
“…Existing visualization methods for visually understanding machine learning models can be categorized into two classes: domain irrelevant [17], [18] and domain specific [19], [20], [21], [22], [23], [24]. Our work is in the second category with a focus in the NLP domain, which enables an in-depth understanding of the working mechanism of the model training process.…”
Section: Visualization For Understanding Nlp Modelsmentioning
confidence: 99%