2015
DOI: 10.1007/978-3-319-09903-3_8
|View full text |Cite
|
Sign up to set email alerts
|

Analysing the Multiple Timescale Recurrent Neural Network for Embodied Language Understanding

Abstract: How the human brain understands natural language and how we can exploit this understanding for building intelligent grounded language systems is open research. Recently, researchers claimed that language is embodied in most -if not all -sensory and sensorimotor modalities and that the brain's architecture favours the emergence of language. In this chapter we investigate the characteristics of such an architecture and propose a model based on the Multiple Timescale Recurrent Neural Network, extended by embodied… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2017
2017
2018
2018

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 11 publications
(18 citation statements)
references
References 44 publications
0
18
0
Order By: Relevance
“…Although the model was reproducing learned symbolic sentences quite well in their experiments, generalisation was not possible to test, because the generation of sentences was initiated by the internal state of the Csc units, which had to be trained individually for every sentence in the model. Heinrich, Magg, and Wermter (2015) extended this model to process the language embodied in a way that visual input will trigger the model to produce a meaningful verbal utterance that appropriately represents the input. The architecture, called EMBMTRNN model, consists of similar MTRNN layers for the language network, where a verbal utterance is processed as a sequence on phoneme level based on initial activity on an overall concept level.…”
Section: Previous Mtrnn Models For Language Processingmentioning
confidence: 99%
“…Although the model was reproducing learned symbolic sentences quite well in their experiments, generalisation was not possible to test, because the generation of sentences was initiated by the internal state of the Csc units, which had to be trained individually for every sentence in the model. Heinrich, Magg, and Wermter (2015) extended this model to process the language embodied in a way that visual input will trigger the model to produce a meaningful verbal utterance that appropriately represents the input. The architecture, called EMBMTRNN model, consists of similar MTRNN layers for the language network, where a verbal utterance is processed as a sequence on phoneme level based on initial activity on an overall concept level.…”
Section: Previous Mtrnn Models For Language Processingmentioning
confidence: 99%
“…The neural dynamics in our MTRNN exhibited a dynamics which are different from those reported in Hinoshita et al (2009) and Heinrich et al (2015). Whereas the noun (or object perceptual inputs) play a significant factor in the dynamics of context layers in these two examples, our network has minimised the effects of nouns or the object perception.…”
Section: Generalisation Ability Of Mtrnnmentioning
confidence: 54%
“…They also have similar information bi-directional flows which allow the networks to recognise and to generate the time sequences. Despite their similarities, compared with LSTM, the MTRNN have other distinct features: First, from the above experiments and from other MTRNN experiments (Heinrich et al 2015;Hinoshita et al 2009), it has been shown that the fast context layers and slow context layers exhibit various dynamics to explicitly represent the relationship between the verbs and nouns. The deep LSTM, on the contrary, has not been reported to have similar dynamics.…”
Section: Hierarchical Recurrent Network and Further Developmentmentioning
confidence: 99%
See 2 more Smart Citations