Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1429
|View full text |Cite
|
Sign up to set email alerts
|

Fine-grained Knowledge Fusion for Sequence Labeling Domain Adaptation

Abstract: In sequence labeling, previous domain adaptation methods focus on the adaptation from the source domain to the entire target domain without considering the diversity of individual target domain samples, which may lead to negative transfer results for certain samples. Besides, an important characteristic of sequence labeling tasks is that different elements within a given sample may also have diverse domain relevance, which requires further consideration. To take the multi-level domain relevance discrepancy int… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 23 publications
0
8
0
Order By: Relevance
“…Zhang et al (2022b) explored the semantic transfer of labels simultaneously in the entity span and the type space, thus achieving smaller discrepancies in the cross-domain transfer. The methods of model parameter transfer aim to share model parameters between different domains through knowledge distillation (Yang et al 2019;Nguyen, Gelli, and Poria 2021;Zhang et al 2021), domain prediction tasks (Lin and Lu 2018;Zhou et al 2019;Jia and Zhang 2020; or generative methods (Jia, Liang, and Zhang 2019;Chen et al 2021;). However, these methods neglect to request more source domain data to explicitly and implicitly transfer knowledge to the target domain.…”
Section: Unrecognized Entitymentioning
confidence: 99%
“…Zhang et al (2022b) explored the semantic transfer of labels simultaneously in the entity span and the type space, thus achieving smaller discrepancies in the cross-domain transfer. The methods of model parameter transfer aim to share model parameters between different domains through knowledge distillation (Yang et al 2019;Nguyen, Gelli, and Poria 2021;Zhang et al 2021), domain prediction tasks (Lin and Lu 2018;Zhou et al 2019;Jia and Zhang 2020; or generative methods (Jia, Liang, and Zhang 2019;Chen et al 2021;). However, these methods neglect to request more source domain data to explicitly and implicitly transfer knowledge to the target domain.…”
Section: Unrecognized Entitymentioning
confidence: 99%
“…Prior work on domain adaptation for coref has focused on a single dataset and often with non-neural models. Yang et al (2012) use an adaptive ensemble which adjusts members per document. Meanwhile, Zhao and Ng (2014) use an active learning approach to adapt a feature-based coref model to be on par with one trained from scratch while using far less data.…”
Section: Arrau (News)mentioning
confidence: 99%
“…These two-stage methods allow using large-scale unlabeled data in pre-training and small labeled data in fine-tuning. In order to adapt to specific tasks or domain, variants of BERT are proposed including small and practical BERT (Tsai et al, 2019;Lan et al, 2020b;Jiao et al, 2020), domain adaptive BERT (Yang et al, 2019a;Gururangan et al, 2020), and task adaptive BERT Xue et al, 2020;Jia et al, 2020). Our work performs further pre-training on BERT and proposes task-aware training objectives to improve NER.…”
Section: Two-stage Training Paradigm For Nlpmentioning
confidence: 99%