2011 IEEE/RSJ International Conference on Intelligent Robots and Systems 2011
DOI: 10.1109/iros.2011.6094814
|View full text |Cite
|
Sign up to set email alerts
|

Autonomous acquisition of multimodal information for online object concept formation by robots

Abstract: This paper proposes a robot that acquires multimodal information, i.e. auditory, visual, and haptic information, fully autonomous way using its embodiment. We also propose an online algorithm of multimodal categorization based on the acquired multimodal information and words, which are partially given by human users. The proposed framework makes it possible for the robot to learn object concepts naturally in everyday operation in conjunction with a small amount of linguistic information from human users. In or… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2013
2013
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(7 citation statements)
references
References 13 publications
0
7
0
Order By: Relevance
“…Araki [44] proposes a multimodal approach (including vision, sound and haptic properties of objects) that is implemented on a real robot, and focuses on learning object concepts by using Latent Dirichlet Allocation (LDA). The multimodal data are acquired autonomously by a robot equipped with a 3D visual sensor, two arms and a small handheld observation table that serves as the platform for capturing multi-view visual images of objects.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Araki [44] proposes a multimodal approach (including vision, sound and haptic properties of objects) that is implemented on a real robot, and focuses on learning object concepts by using Latent Dirichlet Allocation (LDA). The multimodal data are acquired autonomously by a robot equipped with a 3D visual sensor, two arms and a small handheld observation table that serves as the platform for capturing multi-view visual images of objects.…”
Section: Related Workmentioning
confidence: 99%
“…While several models use specific learning algorithms [46], [41], [23], other use very generic topic models such as NMF [42] and LDA [44]. The latter ones provide a sound definition of the problem (i.e., finding the hidden cause that generates a visual feature and an associated word).…”
Section: Related Workmentioning
confidence: 99%
“…Unlike existing datasets, the datasets reported in this paper offer the following advantages to the interested researchers. First, some of these datasets were collected in a single (or similar) experimental setting to address particular problems such as object recognition and categorization [1], [2]. Second, compared with the datasets in [6], [5], our study offers the following advantages to the community: a high number of cameras that can record color images and the usage of sensors that are mounted on the iCub robot.…”
Section: Introductionmentioning
confidence: 99%
“…Other studies, e.g. [61][62][63][64], have presented solutions for merging different sensors' data to address classificationtype problems, such as object/gesture recognition or speaker identification/spatial localisation, rather than imitating demonstrated behaviours. Different approaches have been used to solve imitation learning using different sensor informations, such as hierarchical architectures based on multiple internal models [46,[57][58][59], and Gaussian Mixture Regression together with Hidden Markov Model [60].…”
Section: Introductionmentioning
confidence: 99%