2016
DOI: 10.48550/arxiv.1606.06630
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Multiplicative Integration with Recurrent Neural Networks

Abstract: We introduce a general and simple structural design called "Multiplicative Integration" (MI) to improve recurrent neural networks (RNNs). MI changes the way in which information from difference sources flows and is integrated in the computational building block of an RNN, while introducing almost no extra parameters. The new structure can be easily embedded into many popular RNN models, including LSTMs and GRUs. We empirically analyze its learning behaviour and conduct evaluations on several tasks using differ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 11 publications
(4 citation statements)
references
References 16 publications
0
4
0
Order By: Relevance
“…Surprisingly, there is no empirical difference between two options, before-Hadamard product and after-Hadamard product. This result may build a bridge to relate with studies on multiplicative integration with recurrent neural networks (Wu et al, 2016c).…”
Section: Non-linearitymentioning
confidence: 74%
See 3 more Smart Citations
“…Surprisingly, there is no empirical difference between two options, before-Hadamard product and after-Hadamard product. This result may build a bridge to relate with studies on multiplicative integration with recurrent neural networks (Wu et al, 2016c).…”
Section: Non-linearitymentioning
confidence: 74%
“…(6) Note that using the activation function in low-rank bilinear pooling can be found in an implementation of simple baseline for the VQA dataset (Antol et al, 2015) without an interpretation of low-rank bilinear pooling. However, notably, Wu et al (2016c) studied learning behavior of multiplicative integration in RNNs with discussions and empirical evidences.…”
Section: Nonlinear Activationmentioning
confidence: 99%
See 2 more Smart Citations