2020
DOI: 10.1109/tnnls.2019.2910302
|View full text |Cite
|
Sign up to set email alerts
|

Recurrent Neural Networks With External Addressable Long-Term and Working Memory for Learning Long-Term Dependences

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(11 citation statements)
references
References 23 publications
0
11
0
Order By: Relevance
“…On task 1, 12, 15, 20, it yields the same performance viewing the state-of-theart of nowadays which can be seen as the best performance among all these four models. There are the tasks for which CMemN2N achieves the second-best results, such as tasks 3,5,6,8,10,11,13,14,17,18.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…On task 1, 12, 15, 20, it yields the same performance viewing the state-of-theart of nowadays which can be seen as the best performance among all these four models. There are the tasks for which CMemN2N achieves the second-best results, such as tasks 3,5,6,8,10,11,13,14,17,18.…”
Section: Resultsmentioning
confidence: 99%
“…The reason for end-to-end design is that memory networks proposed in [10] is difficult to train via backpropagation, and demands supervision at each layer of the neural network, so we have [3] to realize model which can be trained endto-end from input to output to guarantee the continuity of memory networks. We test the proposed architecture on the Facebook bAbI task 1 as in [3], [14], [24], which is a nulti-hop reasoning problem.…”
Section: Introductionmentioning
confidence: 99%
“…Specifically, by optimizing the connections of the inner LSTM cells for the performance enhancement of LSTMdominated networks, adding learnable nonlinear state-togate memory connections performs noticeably better than the vanilla LSTM for various tasks with longer sequences [79], while conducting convolutional operation on the two input-to-state/state-to-state transitions and on the previous outputs/current input of the LSTM can integrate long-term temporal dependence with time-frequency characteristics [80] and capture the contextual relationships of the data [81], respectively. Furthermore, by integrating LSTM networks with other components (such as graph neural networks [82] and external memory [83]), learning 3D contexts and the temporal dynamics of multiple studies can accurately estimate 4D changes [84], [85], while exploiting the frame-level dependencies with LSTM (or the shot-level dependencies with graph convolutional networks) [82] and remembering previous metaknowledge [86] in the optimization of performance across similarly structured tasks can perform key-shot [82] and one-shot learning [86], respectively. However, most memory mechanisms rely on weight-like storage (e.g., RNNs) or information-flow gating (e.g., LSTMs) rather than activity-based task-relevant information maintenance of working memory, which yields the best compressed transformative representation of dynamic environments for flexibility/generalizability across tasks [87].…”
Section: Working Memory Inspired Deep Learningmentioning
confidence: 99%
“…The IndRNN has the capability to handle long-term dependence that it can even work on the sequence longer than 5000 time steps. Recently, there have also been efforts such as [26] to use additional mechanisms to assist RNN in learning long-term dependence.…”
Section: B Gate-free Recurrent Neural Networkmentioning
confidence: 99%