2023
DOI: 10.1007/978-981-99-0835-6_11
|View full text |Cite
|
Sign up to set email alerts
|

Fuzzy Metadata Augmentation for Multimodal Data Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 36 publications
0
3
0
Order By: Relevance
“…Each dataset has been partitioned into segments of 7:1:2 (train:val:test)/9:1 (train: val) 23 . In the studies and performance evaluations using this dataset in the literature, no resizing was done and the images were used with the same size 24–29 …”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Each dataset has been partitioned into segments of 7:1:2 (train:val:test)/9:1 (train: val) 23 . In the studies and performance evaluations using this dataset in the literature, no resizing was done and the images were used with the same size 24–29 …”
Section: Methodsmentioning
confidence: 99%
“…23 In the studies and performance evaluations using this dataset in the literature, no resizing was done and the images were used with the same size. [24][25][26][27][28][29]…”
Section: Datasetsmentioning
confidence: 99%
“…Additionally, the authors mapped the main image quality variables to terms used in the General Image-Quality Equation (GIQE)namely, ground sample distance (GSD), relative edge response (RER), and signal-to-noise ratio (SNR)-and assessed the applicability of the GIQE function form to modeling target detector performance in the presence of significant image distortion. Gordienko et al [22] studied the performance of target detection on multimodal satellite images using the Vehicle Detection in Aerial Imagery (VEDAI) dataset, adopting the YOLO (You Only Look Once) framework, covering RGB, IR, and RGB + IR modalities, as well as different image sizes ranging from 128 × 128 to 1024 × 1024. The evaluation method included 10-fold cross-validation to ensure the model's generalization ability, mainly relying on the average precision (mAP) metric, especially mAP@0.5 at an IoU (intersection over union) threshold of 0.5, as well as the mAP range from 0.5 to 0.95; this shows that through this hybrid approach, mAP can be significantly improved at specific image sizes, providing valuable data-driven insights into how to optimize target detection systems.…”
Section: Introductionmentioning
confidence: 99%