Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval 2020
DOI: 10.1145/3397271.3401209
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging Adversarial Training in Self-Learning for Cross-Lingual Text Classification

Abstract: In cross-lingual text classification, one seeks to exploit labeled data from one language to train a text classification model that can then be applied to a completely different language. Recent multilingual representation models have made it much easier to achieve this. Still, there may still be subtle differences between languages that are neglected when doing so. To address this, we present a semisupervised adversarial training process that minimizes the maximal loss for label-preserving input perturbations… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(13 citation statements)
references
References 27 publications
0
13
0
Order By: Relevance
“…Later, realizing that the common one-hot representation of text does not admit perturbations, Miyato et al [20] instead define the adversarial perturbations on continuous word embeddings to successfully extend AT to text classification. Dong et al [7] further propose a self-learning based method to employ AT to cross-lingual text classification. In addition, some methods have also been proposed to explore the use of AT to model other tasks.…”
Section: Adversarial Training (At)mentioning
confidence: 99%
“…Later, realizing that the common one-hot representation of text does not admit perturbations, Miyato et al [20] instead define the adversarial perturbations on continuous word embeddings to successfully extend AT to text classification. Dong et al [7] further propose a self-learning based method to employ AT to cross-lingual text classification. In addition, some methods have also been proposed to explore the use of AT to model other tasks.…”
Section: Adversarial Training (At)mentioning
confidence: 99%
“…Dai et al [19] put forward a novel metric-based GAN, which used the distance-criteria to distinguish between real and fake samples. Dong et al [20] presented a semi-supervised adversarial training process for cross-lingual text classification, where the labeled data from one language could be applied to a completely different language classification. We also refer to various solutions for imbalance datasets.…”
Section: Related Work 21 Adversarial Neural Networkmentioning
confidence: 99%
“…Lai et al (2019) proposed using an unlabeled corpus in the target language to bridge the gap between the language and the domain. Dong et al (2020) 2019) concatenated contextualized representations with knowledge graph embeddings to represent author entities and used them as features for the book classification task. E-BERT (Poerner et al, 2020) inserts KB entities next to the entity names in the input sequence to improve BERT's performance for entity-centric tasks.…”
Section: Improving Cross-lingual Transfer Learningmentioning
confidence: 99%