2019
DOI: 10.1609/aaai.v33i01.33016260
|View full text |Cite
|
Sign up to set email alerts
|

Transfer Learning for Sequence Labeling Using Source Model and Target Data

Abstract: In this paper, we propose an approach for transferring the knowledge of a neural model for sequence labeling, learned from the source domain, to a new model trained on a target domain, where new label categories appear. Our transfer learning (TL) techniques enable to adapt the source model using the target data and new categories, without accessing to the source data. Our solution consists in adding new neurons in the output layer of the target model and transferring parameters from the source model, which are… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 34 publications
(16 citation statements)
references
References 5 publications
0
16
0
Order By: Relevance
“…Neural Network-based Models for NER Some researchers design different architectures which vary in word encoder (Chiu and Nichols 2016;Ma and Hovy 2016), sentence encoder (Huang, Xu, and Yu 2015;Ma and Hovy 2016;Chiu and Nichols 2016) and decoder (CRF) (Huang, Xu, and Yu 2015). Some works explore how to transfer learned parameters from the source domain to a new domain (Chen and Moschitti 2019;Lin and Lu 2018;Cao et al 2018). Recently, (Yang, Liang, and Zhang 2018;Reimers and Gurevych 2017) systematically analyze neural NER models to provide useful guidelines for NLP practitioners.…”
Section: Related Workmentioning
confidence: 99%
“…Neural Network-based Models for NER Some researchers design different architectures which vary in word encoder (Chiu and Nichols 2016;Ma and Hovy 2016), sentence encoder (Huang, Xu, and Yu 2015;Ma and Hovy 2016;Chiu and Nichols 2016) and decoder (CRF) (Huang, Xu, and Yu 2015). Some works explore how to transfer learned parameters from the source domain to a new domain (Chen and Moschitti 2019;Lin and Lu 2018;Cao et al 2018). Recently, (Yang, Liang, and Zhang 2018;Reimers and Gurevych 2017) systematically analyze neural NER models to provide useful guidelines for NLP practitioners.…”
Section: Related Workmentioning
confidence: 99%
“…Reference [10] focuses on transfer learning from English to Japanese, proposing the method called romanization to help dissimilar languages share a common character embedding space. Other approaches include [22], which encodes slot description to vectors and employs an attention layer to obtain slot-aware representations of user input, and [23], which uses features derived from the source model.…”
Section: B Transfer Learningmentioning
confidence: 99%
“…Then, a softmax function is used to yield the conditional probability, wherẽ Y denotes one of all possible label sequences (paths). ψ(X, Y ) is defined as the sum of emission scores (or state scores) and transition scores over all time steps (Morris and Fosler-Lussier, 2006;Chen and Moschitti, 2019):…”
Section: Modelmentioning
confidence: 99%