Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1127
|View full text |Cite
|
Sign up to set email alerts
|

Modeling Noisiness to Recognize Named Entities using Multitask Neural Networks on Social Media

Abstract: Recognizing named entities in a document is a key task in many NLP applications. Although current state-of-the-art approaches to this task reach a high performance on clean text (e.g. newswire genres), those algorithms dramatically degrade when they are moved to noisy environments such as social media domains. We present two systems that address the challenges of processing social media data using character-level phonetics and phonology, word embeddings, and Part-of-Speech tags as features. The first model is … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
39
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 43 publications
(39 citation statements)
references
References 31 publications
0
39
0
Order By: Relevance
“…has not yet experimented with contextual embeddings on WNUT, as side result we report a new state-of-the-art of 49.59 F1 vs. the previous best reported number of 45.55 (Aguilar et al, 2018).…”
Section: Resultsmentioning
confidence: 59%
See 2 more Smart Citations
“…has not yet experimented with contextual embeddings on WNUT, as side result we report a new state-of-the-art of 49.59 F1 vs. the previous best reported number of 45.55 (Aguilar et al, 2018).…”
Section: Resultsmentioning
confidence: 59%
“…This result is expected since most entities appear only few times in this dataset, giving our approach little evidence to aggregate and pool. Nevertheless, since recent work has not yet experimented with contextual embeddings on WNUT, as side result we report a new state-of-the-art of 49.59 F1 vs. the previous best reported number of 45.55 (Aguilar et al, 2018). Pooling operations.…”
Section: Resultsmentioning
confidence: 63%
See 1 more Smart Citation
“…After architecture search, we test the transferability of the learned architecture. In order to apply the model to other tasks, we directly use the architecture searched on WikiText-103 and train the param-Models F1 Cross-BiLSTM-CNN (Aguilar et al, 2018) (Yang and Zhang, 2018) 95.06 BiLSTM-CRF + IntNet (Xin et al, 2018) 95.29 Flair (Akbik et al, 2019) 96.72 GCDT + BERTLARGE (Liu et al, 2019b) 97 eters with the in-domain data. In our experiments, we adapt the model to CoNLL-2003, WNUT-2017NER tasks and CoNLL-2000 For the two NER tasks, it achieves new stateof-the-art F1 scores (Table 4 and Table 5).…”
Section: Transferring To Other Tasksmentioning
confidence: 99%
“…and Rei (2017) studied multi-task learning of sequence labeling with language models. Aguilar et al (2018) and Cao et al (2018) proposed multi-task learning of NER with word segmentation. Peng and Dredze (2017)'s method of multi-task learning leverages the performance of domain adaptation.…”
Section: Multi-task Learningmentioning
confidence: 99%