Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Confere 2015
DOI: 10.3115/v1/p15-1046
|View full text |Cite
|
Sign up to set email alerts
|

New Transfer Learning Techniques for Disparate Label Sets

Abstract: In natural language understanding (NLU), a user utterance can be labeled differently depending on the domain or application (e.g., weather vs. calendar). Standard domain adaptation techniques are not directly applicable to take advantage of the existing annotations because they assume that the label set is invariant. We propose a solution based on label embeddings induced from canonical correlation analysis (CCA) that reduces the problem to a standard domain adaptation task and allows use of a number of transf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
58
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
6
2
2

Relationship

5
5

Authors

Journals

citations
Cited by 55 publications
(58 citation statements)
references
References 26 publications
0
58
0
Order By: Relevance
“…An intent is defined as the type of content the user is seeking. This task is part of the spoken language understanding problem (Li et al, 2009;Tur and De Mori, 2011;Kim et al, 2015c;Mesnil et al, 2015;Kim et al, 2015a;Xu and Sarikaya, 2014;Kim et al, 2015b;Kim et al, 2015d). The amount of training data we used ranges from 12k to 120k (in number of queries) across different domains, the test data was from 2k to 20k.…”
Section: Methodsmentioning
confidence: 99%
“…An intent is defined as the type of content the user is seeking. This task is part of the spoken language understanding problem (Li et al, 2009;Tur and De Mori, 2011;Kim et al, 2015c;Mesnil et al, 2015;Kim et al, 2015a;Xu and Sarikaya, 2014;Kim et al, 2015b;Kim et al, 2015d). The amount of training data we used ranges from 12k to 120k (in number of queries) across different domains, the test data was from 2k to 20k.…”
Section: Methodsmentioning
confidence: 99%
“…The closest prior work by Kim et al (2015) address a sequential labeling problem in NLU where the fine grained label sets across domains differ. However, they assume that there exists a bijective mapping between the coarse and fine-grained label sets across domains.…”
Section: Multi-task Learningmentioning
confidence: 99%
“…In fact, CCA has recently been adapted to learning latent word representations in an interesting way: by dividing each word position into a token view (which only sees surrounding context) and a type view (which only sees the word itself) and performing a CCA between these two views (Dhillon et al, 2012;Stratos et al, 2014;Stratos et al, 2015;Kim et al, 2015c). CCA is also used to induce label representations (Kim et al, 2015d) and lexicon representations (Kim et al, 2015b).…”
Section: Canonical Correlation Analysis (Cca)mentioning
confidence: 99%