Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.433
|View full text |Cite
|
Sign up to set email alerts
|

Exploring Distantly-Labeled Rationales in Neural Network Models

Abstract: Recent studies strive to incorporate various human rationales into neural networks to improve model performance, but few pay attention to the quality of the rationales. Most existing methods distribute their models' focus to distantly-labeled rationale words entirely and equally, while ignoring the potential important non-rationale words and not distinguishing the importance of different rationale words. In this paper, we propose two novel auxiliary loss functions to make better use of distantly-labeled ration… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…We also discuss how the optimization of the task-specific loss and the new loss should be coordinated. For previous methods utilizing gradient-based interpretation for model enhancement (Huang et al, 2021;Ghaeini et al, 2019), we show that they can be seen as instances of our framework.…”
Section: Introductionmentioning
confidence: 54%
See 2 more Smart Citations
“…We also discuss how the optimization of the task-specific loss and the new loss should be coordinated. For previous methods utilizing gradient-based interpretation for model enhancement (Huang et al, 2021;Ghaeini et al, 2019), we show that they can be seen as instances of our framework.…”
Section: Introductionmentioning
confidence: 54%
“…In Ghaeini et al (2019)'s work, f is a function that takes the sum of the gradients of each input embedding dimension and D is to take the sum of the gradients of rationale words. In Huang et al (2021)'s work, f is the L 1 norm that sums up the absolute value of gradients over the input embedding dimensions and D is designed in various ways to give rationale words higher attribution scores.…”
Section: Utilizing Gradient-based Methodsmentioning
confidence: 99%
See 1 more Smart Citation