2018
DOI: 10.1016/j.neucom.2017.04.077
|View full text |Cite
|
Sign up to set email alerts
|

Multi-modal local receptive field extreme learning machine for object recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 32 publications
(9 citation statements)
references
References 1 publication
0
9
0
Order By: Relevance
“…That in turn allows for combining even heterogeneous data from very different sources and lets the machine learning model to process joint representation in order to produce an output. That type of fusion is widely spread in robotics community in areas such as object recognition [19] and scene recognition [20] tasks. Multi-modal fusion applied to robot motion planning was presented in [21] and contact-rich manipulation tasks in [7].…”
Section: Data Fusion Approachesmentioning
confidence: 99%
“…That in turn allows for combining even heterogeneous data from very different sources and lets the machine learning model to process joint representation in order to produce an output. That type of fusion is widely spread in robotics community in areas such as object recognition [19] and scene recognition [20] tasks. Multi-modal fusion applied to robot motion planning was presented in [21] and contact-rich manipulation tasks in [7].…”
Section: Data Fusion Approachesmentioning
confidence: 99%
“…They fused separately processed RGB and depth images through a CCA layer and a combining layer was introduced to the multi-view CNN. H. Liu et al [21] developed an extreme learning machine (ELM) structure using a multi-modal local receptive field (MM-LRF). There, LRF is used as a feature extractor for each modality, and a shared layer is proposed for combining the features.…”
Section: Sustainable Multi-objects Recognition Via Depth Imagesmentioning
confidence: 99%
“…Finally, they used Hough voting for the recognition of the object. H. Liu et al [21] presented a multi-modal architecture named MM-ELM-LRF. They extracted features for both of the modalities (RGB and depth) by applying ELM-LRF.…”
Section: Rgb-d Object Rgb-d Scenes Nyudv1mentioning
confidence: 99%
“…ELM can speed up learning by randomly generating input weights and hidden biases and by adopting Moore-Penrose (MP) generalized inversion to determine the output weights. Compared with the traditional gradient-based learning algorithms, ELM not only has faster learning speed and higher generalization performance, but also avoids the difficulties of stopping conditions, learning rate, learning cycle and local minimization encountered by gradient-based learning methods [8]. ELM and its variants [9][10][11] have been applied in many fields [12][13][14], including boiler combustion system modeling [15,16].…”
Section: Introductionmentioning
confidence: 99%