Proceedings of the 5th Workshop on Representation Learning for NLP 2020
DOI: 10.18653/v1/2020.repl4nlp-1.1
|View full text |Cite
|
Sign up to set email alerts
|

Zero-Resource Cross-Domain Named Entity Recognition

Abstract: Cross-domain named entity recognition (NER) models are able to cope with the scarcity issue of NER samples in target domains. However, most of the existing NER benchmarks lack domain-specialized entity types or do not focus on a certain domain, leading to a less effective cross-domain evaluation. To address these obstacles, we introduce a cross-domain NER dataset (CrossNER), a fully-labeled collection of NER data spanning over five diverse domains with specialized entity categories for different domains. Addit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 38 publications
(24 citation statements)
references
References 32 publications
0
24
0
Order By: Relevance
“…It has demonstrated to be useful in various applications like recommendation (Zhu et al, 2020), domain adaptation for sentiment analysis, and POS tagging (Guo et al, 2018). For NER, Liu et al (2020) proposed a Mixture of Entity Experts (MoEE) approach where they train an expert layer for each entity type, and then combine them using an MoE approach. Their approach does not include external gazetteers, and the experts provide an independent representation that is not combined with the word representation.…”
Section: Mixture-of-experts (Moe) Modelsmentioning
confidence: 99%
“…It has demonstrated to be useful in various applications like recommendation (Zhu et al, 2020), domain adaptation for sentiment analysis, and POS tagging (Guo et al, 2018). For NER, Liu et al (2020) proposed a Mixture of Entity Experts (MoEE) approach where they train an expert layer for each entity type, and then combine them using an MoE approach. Their approach does not include external gazetteers, and the experts provide an independent representation that is not combined with the word representation.…”
Section: Mixture-of-experts (Moe) Modelsmentioning
confidence: 99%
“…In general, recent works on model generalizability can be divided into two different directions: 1) adaptive training and 2) robust loss function. In adaptive training, different meta-learning [5] and fast adaptation [18,20,35] approaches have been developed and show promising result for improving the generalization of the model over different domains. Another meta-learning approach, called meta transfer learning [34], improves the generalization ability for a low-resource domain by leveraging a high-resource domain dataset.…”
Section: Related Workmentioning
confidence: 99%
“…For a fair comparison, we adopt the same network architecture BiLSTM (Hochreiter and Schmidhuber, 1997) as previous work (Bapna et al, 2017;Lee and Jha, 2019;Shah et al, 2019;Liu et al, 2020a;Liu et al, 2020b). Given an utterance with n tokens as w = [w 1 , w 2 , .…”
Section: Czsl Modelmentioning
confidence: 99%
“…The main drawback is the multiple prediction problem where a word may be predicted as multiple slot types. In contrast, (Liu et al, 2020a;Liu et al, 2020b) propose a two-stage slot filling framework. They first predict whether the tokens are slot entities or not by a BIO 3-way classifier, then identify their specific slot types based on slot type descriptions.…”
Section: Introductionmentioning
confidence: 99%