Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1404
|View full text |Cite
|
Sign up to set email alerts
|

Benchmarking Zero-shot Text Classification: Datasets, Evaluation and Entailment Approach

Abstract: Zero-shot text classification (0SHOT-TC) is a challenging NLU problem to which little attention has been paid by the research community. 0SHOT-TC aims to associate an appropriate label with a piece of text, irrespective of the text domain and the aspect (e.g., topic, emotion, event, etc.) described by the label. And there are only a few articles studying 0SHOT-TC, all focusing only on topical categorization which, we argue, is just the tip of the iceberg in 0SHOT-TC. In addition, the chaotic experiments in lit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
303
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 316 publications
(306 citation statements)
references
References 15 publications
3
303
0
Order By: Relevance
“…Early attempts rely on distant supervision such as Wikipedia to interpret the label name semantics and derive documentconcept relevance via explicit semantic analysis (Gabrilovich and Markovitch, 2007). Since the classifier is learned purely from general knowledge without even requiring any unlabeled domainspecific data, these methods are called dataless classification (Chang et al, 2008;Song and Roth, 2014;Yin et al, 2019). Later, topic models (Chen et al, 2015;Li et al, 2016) are exploited for seedguided classification to learn seed word-aware topics by biasing the Dirichlet priors and to infer posterior document-topic assignment.…”
Section: Weakly-supervised Text Classificationmentioning
confidence: 99%
“…Early attempts rely on distant supervision such as Wikipedia to interpret the label name semantics and derive documentconcept relevance via explicit semantic analysis (Gabrilovich and Markovitch, 2007). Since the classifier is learned purely from general knowledge without even requiring any unlabeled domainspecific data, these methods are called dataless classification (Chang et al, 2008;Song and Roth, 2014;Yin et al, 2019). Later, topic models (Chen et al, 2015;Li et al, 2016) are exploited for seedguided classification to learn seed word-aware topics by biasing the Dirichlet priors and to infer posterior document-topic assignment.…”
Section: Weakly-supervised Text Classificationmentioning
confidence: 99%
“…Recent work in NLP has shown significant progress in learning tasks from examples. Large pretrained language models have dramatically improved performance on standard benchmarks (Peters et al, 2018;Devlin et al, 2019;Raffel et al, 2019) and have shown promising results in zero shot prediction by leveraging their language understanding capabilities (Levy et al, 2017;Zhou et al, 2018;Yin et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…This coreference example shows that transforming an NLP task as textual entailment may obtain surprising advantages. There are more NLP tasks that can fit the entailment framework easily, such as text classification (Yin et al, 2019), relation extraction, summarization, etc. However, we also need to admit that reformulating into entailment may also need to fight against new challenges.…”
Section: Results and Analysesmentioning
confidence: 99%