Proceedings of the 5th Workshop on Cognitive Aspects of Computational Language Learning (CogACLL) 2014
DOI: 10.3115/v1/w14-0507
|View full text |Cite
|
Sign up to set email alerts
|

A multimodal corpus for the evaluation of computational models for (grounded) language acquisition

Abstract: This paper describes the design and acquisition of a German multimodal corpus for the development and evaluation of computational models for (grounded) language acquisition and algorithms enabling corresponding capabilities in robots. The corpus contains parallel data from multiple speakers/actors, including speech, visual data from different perspectives and body posture data. The corpus is designed to support the development and evaluation of models learning rather complex grounded linguistic structures, e.g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 15 publications
0
1
0
Order By: Relevance
“…Recent work on detecting fine-grained events in videos is particularly relevant (Hendricks et al, 2018;Zhukov et al, 2019;Fried et al, 2020, among others). Especially relevant is the data collected by Gaspers et al (2014), in which human subjects were asked to play simple games with a physical robot and narrate while doing so. Our data and work differs primarily in that we focus on the ability to ground to symbolic objects and physics rather than only to pixel data.…”
Section: Related Workmentioning
confidence: 99%
“…Recent work on detecting fine-grained events in videos is particularly relevant (Hendricks et al, 2018;Zhukov et al, 2019;Fried et al, 2020, among others). Especially relevant is the data collected by Gaspers et al (2014), in which human subjects were asked to play simple games with a physical robot and narrate while doing so. Our data and work differs primarily in that we focus on the ability to ground to symbolic objects and physics rather than only to pixel data.…”
Section: Related Workmentioning
confidence: 99%