2011 IEEE/RSJ International Conference on Intelligent Robots and Systems 2011
DOI: 10.1109/iros.2011.6048422
|View full text |Cite
|
Sign up to set email alerts
|

Autonomous acquisition of multimodal information for online object concept formation by robots

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
32
0

Year Published

2012
2012
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(32 citation statements)
references
References 0 publications
0
32
0
Order By: Relevance
“…In order to bring the object, the robot must find and recognize the object using multimodal information of speech of the object name and images of the object. Several researchers have studied learning objects through interaction by multimodal information so that robots can recognize the object [2], [3], [6], [7]. In [6], they proposed a language acquisition model using speech and image information.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In order to bring the object, the robot must find and recognize the object using multimodal information of speech of the object name and images of the object. Several researchers have studied learning objects through interaction by multimodal information so that robots can recognize the object [2], [3], [6], [7]. In [6], they proposed a language acquisition model using speech and image information.…”
Section: Related Workmentioning
confidence: 99%
“…In [6], they proposed a language acquisition model using speech and image information. Araki et al presented a robot that can learn the concept of an object via spoken dialog [7]. Ozasa et al proposed a method for object recognition using multimodal information in order to learn unknown objects in the object manipulation task where a human asks a robot via voice to bring an object among several objects on a table [2].…”
Section: Related Workmentioning
confidence: 99%
“…The tactile array sensor consists of 162 tactile sensors, and 162 time series sensor values are obtained by the hand grasping the object. The sensor values are approximated by using a sigmoid function, the parameters of which are used as feature vectors for the haptic information [14]. The robot grasps the object five times, and therefore, 810(= 162 × 5) sensor values are obtained from one object.…”
Section: Visual Informationmentioning
confidence: 99%
“…To obtain visual information, the robot uses a small handheld observation table with an XBee wireless controller to control the viewpoints for observing the object [9]. Considering the computational cost, Speeded-Up Robust Features (SURF) [10] is used and the descriptors with 128 dimensions are collected from each captured image and stored in the object database.…”
Section: A Superpixel-based Object Detectionmentioning
confidence: 99%