2020
DOI: 10.1109/tnnls.2019.2920267
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Deep Network Embedding With Integrated Structure and Attribute Information

Abstract: Network embedding is the process of learning lowdimensional representations for nodes in a network, while preserving node features. Existing studies only leverage network structure information and focus on preserving structural features. However, nodes in real-world networks often have a rich set of attributes providing extra semantic information. It has been demonstrated that both structural and attribute features are important for network analysis tasks. To preserve both features, we investigate the problem … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(13 citation statements)
references
References 39 publications
0
13
0
Order By: Relevance
“…shapes and edges) through the application of convolutions on successive layers of multiple spatial scales [17]. Multimodal NNs can utilize multiple types of inputs simultaneously [22].…”
Section: Model For Clusterability Assessmentmentioning
confidence: 99%
See 2 more Smart Citations
“…shapes and edges) through the application of convolutions on successive layers of multiple spatial scales [17]. Multimodal NNs can utilize multiple types of inputs simultaneously [22].…”
Section: Model For Clusterability Assessmentmentioning
confidence: 99%
“…We require our model to utilize two types of inputs: density grids (Section III-A) and CC vectors (Section III-B). One such architecture that is able to combine different types of representations is called a multimodal neural network (MNN) [22]. For example, whereas pixels and sound waves are difficult to relate to each other, MNN is able to combine audio and video data in a single neural network [18].…”
Section: B Multimodal Convolutional Neural Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…With hierarchical architectures and end-to-end training fashion, GNNs mainly fall into two perspectives: a spectral perspective and a vertex perspective. From a spectral perspective, graphs are converted into its spectrum signal processing [43]. For example, Spectral CNN [44] used the eigen-decomposition of graph Laplacian.…”
Section: B Graph Representation Learningmentioning
confidence: 99%
“…Visual Question Answering (VQA) is a multimodal research task that aims to answer questions related to the given image. Compared with other multimodal learning tasks (e.g., visual description [ 1 ], visual grounding [ 2 , 3 , 4 ], multimodal embedding learning [ 5 , 6 , 7 , 8 ]), VQA requires a fine-grained semantic understanding of both visual and textual content to predict the correct natural language answer. Therefore, VQA has recently emerged as an extremely challenging task and drawn considerable attention from researchers.…”
Section: Introductionmentioning
confidence: 99%