2022
DOI: 10.1007/978-3-031-19836-6_12
|View full text |Cite
|
Sign up to set email alerts
|

Uncertainty-Aware Multi-modal Learning via Cross-Modal Random Network Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…It is labelled with the suffix "partial" because it can make predictions in the case of incomplete observations as well. We applied a simple method used in multimodal modelling [28] to replace missing modality features. This method substitutes the features of a missing modality with those of other modalities.…”
Section: Model Architecturementioning
confidence: 99%
See 1 more Smart Citation
“…It is labelled with the suffix "partial" because it can make predictions in the case of incomplete observations as well. We applied a simple method used in multimodal modelling [28] to replace missing modality features. This method substitutes the features of a missing modality with those of other modalities.…”
Section: Model Architecturementioning
confidence: 99%
“…Additionally, we experimented with multimodal methods to address the problem of the entire model being unfeasible in the event of the absence of any one modality. These methods have been studied in various fields [25], [26], [27], [28]. However, in this study, considering the case when MTSO data were found to be unavailable, we introduced a simple method to replace the missing MTSO data with feature data from CCTV images; subsequently, we tested the effectiveness of this method.…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, Tian et al [49] explored the use of uncertainty estimation in fusing the softmax scores predicted using CNNs for semantic segmentation. Other notable approaches to uncertainty-aware multimodal fusion are based on optimal transport for cross-modal correspondence [50], random prior functions [51], boosted ensembles [52], and factorised deep markov models [53].…”
Section: Related Workmentioning
confidence: 99%
“…However, the combination of multimodal data is usually challenging. There are studies on fusing multimodal data according to their uncertainties, but this may face numerical instability and is difficult to transfer from one application to another [19]. Instead of directly fusing the multisensory data in a numerical space, we propose to use multimodal modules to translate them into natural language expressions that an LLM can easily digest.…”
Section: Related Workmentioning
confidence: 99%