2016
DOI: 10.48550/arxiv.1602.07749
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Toward Mention Detection Robustness with Recurrent Neural Networks

Abstract: One of the key challenges in natural language processing (NLP) is to yield good performance across application domains and languages. In this work, we investigate the robustness of the mention detection systems, one of the fundamental tasks in information extraction, via recurrent neural networks (RNNs). The advantage of RNNs over the traditional approaches is their capacity to capture long ranges of context and implicitly adapt the word embeddings, trained on a large corpus, into a task-specific word represen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 26 publications
0
5
0
Order By: Relevance
“…Some studies [87]- [89] employ word-level representation, which is typically pre-trained over large collections of text through unsupervised algorithms such as continuous bagof-words (CBOW) and continuous skip-gram models [90] (see Figure 4 for the architectures of CBOW and skip-gram).…”
Section: Word-level Representationmentioning
confidence: 99%
See 2 more Smart Citations
“…Some studies [87]- [89] employ word-level representation, which is typically pre-trained over large collections of text through unsupervised algorithms such as continuous bagof-words (CBOW) and continuous skip-gram models [90] (see Figure 4 for the architectures of CBOW and skip-gram).…”
Section: Word-level Representationmentioning
confidence: 99%
“…The dictionary contains 205,924 words in 600 dimensional vectors. Nguyen et al [87] used word2vec toolkit to learn word embeddings for English from the Gigaword corpus augmented with newsgroups…”
Section: Word-level Representationmentioning
confidence: 99%
See 1 more Smart Citation
“…Compared to Conv-CRF, the BiLSTM-CRF model is more robust and relies less on distributed representations of inputs. Subsequently, a series of works also adopted BiLSTM as the context encoder, such as Lample et al [3], Chiu et al [4], Nguyen et al [5], Zheng et al [6], Ma et al [7], and Zheng et al [8]. Yang et al [9] adopted the GRU-CRF method and simultaneously encoded the word vector and word vector of the context.…”
Section: Related Workmentioning
confidence: 99%
“…They have been adapted to learn vector representations of words for NLP-based phenotyping [112,136], laying a foundation for computational phenotyping. Deep learning has been applied on various NLP applications, including semantic representation [146], semantic analysis [147,148], information retrieval [149,150], entity recognition [151,152], relation extraction [153][154][155][156], and event detection [157,158]. Beaulieu-Jones et al [136] developed a neural network approach to construct phenotypes to classify patient disease status.…”
Section: Deep Learningmentioning
confidence: 99%