2018
DOI: 10.14311/nnw.2018.28.018
|View full text |Cite
|
Sign up to set email alerts
|

Csrncva: A Model of Cross-Media Semantic Retrieval Based on Neural Computing of Visual and Auditory Sensations

Abstract: Cross-media semantic retrieval (CSR) and cross-modal semantic mapping are key problems of the multimedia search engine. The cognitive function and neural structure for visual and auditory information process are an important reference for the study of brain-inspired CSR. In this paper, we analyze the hierarchy, the functionality and the structure of visual and auditory in the brain. Considering an idea from deep belief network and hierarchical temporal memory, we presented a brain-inspired intelligent model, c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2
1
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(8 citation statements)
references
References 29 publications
0
8
0
Order By: Relevance
“…How to efficiently set the parameter range still needs further exploration. e cross-modal retrieval method based on deep learning makes full use of the powerful feature extraction capabilities of deep learning models, learns the feature representation of different modal data, and then establishes semantic associations between modalities at a high level [17]. Reference [18] proposed a two-stage deep learning method for supervised cross-modal retrieval, extending the traditional norm-related analysis from 2 views to 3 views and conducting supervised learning in two stages.…”
Section: Real-valued Cross-modal Retrieval Methodmentioning
confidence: 99%
“…How to efficiently set the parameter range still needs further exploration. e cross-modal retrieval method based on deep learning makes full use of the powerful feature extraction capabilities of deep learning models, learns the feature representation of different modal data, and then establishes semantic associations between modalities at a high level [17]. Reference [18] proposed a two-stage deep learning method for supervised cross-modal retrieval, extending the traditional norm-related analysis from 2 views to 3 views and conducting supervised learning in two stages.…”
Section: Real-valued Cross-modal Retrieval Methodmentioning
confidence: 99%
“…The spiking probability is propagating among micro-columns. Micro-columns collect information from lower neighbor micro-columns and disseminate information from upper neighbor micro-columns (Liu et al 2018a). At the same time, it also receives feedback information from LDP, and prediction information from an upper neighbor.…”
Section: Hypothesis 1 (Bmm Hypothesis)mentioning
confidence: 99%
“…For instance, the node is double structure in HTM, RBM, SVM and so on. As shown in Figure 2 C, S mimicking functions from L1 to L4, and simulates memory and spatial patterns process (Liu et al 2018a). T mimicking functions of L5 and L6, and simulates memory and temporal patterns process.…”
Section: Hypothesis 1 (Bmm Hypothesis)mentioning
confidence: 99%
See 2 more Smart Citations