Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.272
|View full text |Cite
|
Sign up to set email alerts
|

Zero-shot Text Classification via Reinforced Self-training

Abstract: Zero-shot learning has been a tough problem since no labeled data is available for unseen classes during training, especially for classes with low similarity. In this situation, transferring from seen classes to unseen classes is extremely hard. To tackle this problem, in this paper we propose a self-training based method to efficiently leverage unlabeled data. Traditional self-training methods use fixed heuristics to select instances from unlabeled data, whose performance varies among different datasets. We p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
33
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 57 publications
(33 citation statements)
references
References 31 publications
0
33
0
Order By: Relevance
“…Other approaches for few-shot learning in NLP include exploiting examples from related tasks (Yu et al, 2018;Gu et al, 2018;Dou et al, 2019;Qian and Yu, 2019;Yin et al, 2019) and using data augmentation (Xie et al, 2020;; the latter commonly relies on back-translation (Sennrich et al, 2016), requiring large amounts of parallel data. Approaches using textual class descriptors typically assume that abundant examples are available for a subset of classes (e.g., Romera-Paredes and Torr, 2015;Veeranna et al, 2016;Ye et al, 2020). In contrast, our approach requires no additional labeled data and provides an intuitive interface to leverage task-specific human knowledge.…”
Section: Introductionmentioning
confidence: 99%
“…Other approaches for few-shot learning in NLP include exploiting examples from related tasks (Yu et al, 2018;Gu et al, 2018;Dou et al, 2019;Qian and Yu, 2019;Yin et al, 2019) and using data augmentation (Xie et al, 2020;; the latter commonly relies on back-translation (Sennrich et al, 2016), requiring large amounts of parallel data. Approaches using textual class descriptors typically assume that abundant examples are available for a subset of classes (e.g., Romera-Paredes and Torr, 2015;Veeranna et al, 2016;Ye et al, 2020). In contrast, our approach requires no additional labeled data and provides an intuitive interface to leverage task-specific human knowledge.…”
Section: Introductionmentioning
confidence: 99%
“…A significant challenge for real-life development of MTC applications is severe deficiencies of annotated data for each label in the hierarchy, which demands better solutions for zero-shot learning. The existing zero-shot learning for multi-label text classification (ZS-MTC) mostly learns a matching model between the feature space of text and the label space (Ye et al, 2020). In order to learn effective representations for labels, a majority of existing work incorporates label hierarchies via a label encoder designed as Graph Neural Networks (GNNs) that can aggregate the neighboring information for labels (Chalkidis et al, 2020;Lu et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…Other approaches to few-shot learning in NLP commonly require large sets of examples from related tasks (Gu et al, 2018;Dou et al, 2019;Qian and Yu, 2019;Ye et al, 2020), parallel data for consistency training (Xie et al, 2020;, or highly specialized methods tailored towards a specific task (Laban et al, 2020). In contrast, GENPET requires no additional labeled data and provides an intuitive interface to leveraging task-specific human knowledge.…”
Section: Related Workmentioning
confidence: 99%