2018
DOI: 10.1007/978-3-319-77113-7_4
|View full text |Cite
|
Sign up to set email alerts
|

Label-Dependencies Aware Recurrent Neural Networks

Abstract: In the last few years, Recurrent Neural Networks (RNNs) have proved effective on several NLP tasks. Despite such great success, their ability to model sequence labeling is still limited. This lead research toward solutions where RNNs are combined with models which already proved effective in this domain, such as CRFs. In this work we propose a solution far simpler but very effective: an evolution of the simple Jordan RNN, where labels are re-injected as input into the network, and converted into embeddings, in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0
1

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2

Relationship

4
3

Authors

Journals

citations
Cited by 9 publications
(12 citation statements)
references
References 40 publications
0
11
0
1
Order By: Relevance
“…The improved RNN proposed in this paper is based on a similar model described in [19,20], and later improved in [21,22].…”
Section: Ejordanmentioning
confidence: 99%
“…The improved RNN proposed in this paper is based on a similar model described in [19,20], and later improved in [21,22].…”
Section: Ejordanmentioning
confidence: 99%
“…Inspired by the work of Dupont et.al on document analysis [14], we propose to consider information of the previous temporal segment. We perform this by increasing the dimension of the human pose features of the current segment s by concatenating it with a one-hot context vector C s−1 corresponding to the classification of the previous temporal segment s − 1, as illustrated in Fig.…”
Section: ) Context Featuresmentioning
confidence: 99%
“…is performed. They are inspired from the Sequence-to-Sequence architecture for the overall architecture, and from the models proposed in our previous work (Dinarelli & Tellier, 2016a,b;Dinarelli et al, 2017;Dupont et al, 2017;Dinarelli & Grobol, 2019) for making predictions based on a bidirectional context on the output side (labels). Afterwards we added to this architecture some of the characteristics of the Transformer model.…”
Section: Neural Architecturesmentioning
confidence: 99%
“…The neural architecture described so far uses the same ideas introduced in (Dinarelli & Tellier, 2016a;Dinarelli et al, 2017;Dupont et al, 2017;Dinarelli & Grobol, 2019) for predicting labels using both representations of left (forward) and right (backward) contexts, and for both input-level information (words, characters, etc.) and labels.…”
Section: 5mentioning
confidence: 99%