2023
DOI: 10.1109/tcsvt.2022.3208833
|View full text |Cite
|
Sign up to set email alerts
|

RGB-T Semantic Segmentation With Location, Activation, and Sharpening

Abstract: Semantic segmentation is important for scene understanding. To address the scenes of adverse illumination conditions of natural images, thermal infrared (TIR) images are introduced. Most existing RGB-T semantic segmentation methods follow three cross-modal fusion paradigms, i.e., encoder fusion, decoder fusion, and feature fusion. Some methods, unfortunately, ignore the properties of RGB and TIR features or the properties of features at different levels. In this paper, we propose a novel feature fusion-based n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 54 publications
(8 citation statements)
references
References 62 publications
0
8
0
Order By: Relevance
“…Data augmentation is applied to the dataset used during training, including random flipping and cropping operations. To verify the semantic segmentation performance of the proposed method, this section compares it with six other semantic segmentation methods: BiSeNet [31], RTFNet [23], FuseSeg [24], GMNet [32], ABMDRNet [33], and LASNet [34]. Among them, BiSeNet is a semantic segmentation algorithm based on a single natural light image, while the others are based on the fusion of infrared and natural light images.…”
Section: Experiments Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Data augmentation is applied to the dataset used during training, including random flipping and cropping operations. To verify the semantic segmentation performance of the proposed method, this section compares it with six other semantic segmentation methods: BiSeNet [31], RTFNet [23], FuseSeg [24], GMNet [32], ABMDRNet [33], and LASNet [34]. Among them, BiSeNet is a semantic segmentation algorithm based on a single natural light image, while the others are based on the fusion of infrared and natural light images.…”
Section: Experiments Resultsmentioning
confidence: 99%
“…GMNet [32] only uses a deep feature fusion module to segment semantic regions, so its IOU metric is not good enough. ABMDRNet [33] and LASNet [34] employ different strategies to handle the fusion information of low-level and high-level data, so they obtained good results. In our method, distinct information is extracted for different layers, and the extracted high-level and global information is densely integrated into the restoration process of the original resolution of low-level information for each layer, so our method obtains the best result.…”
Section: Experiments Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Methods Modalities % mIoU GMNet [51] RGB-Infrared 49.2 LASNet [52] RGB-Infrared 42.5 EGFNet [41] RGB-Infrared 47.3 FEANet [42] RGB-Infrared 46.8 DIDFuse [43] RGB-Infrared 50.6 ReCoNet [44] RGB-Infrared 50.9 U2Fusion [53] RGB-Infrared 47.9 TarDAL [54] RGB-Infrared 48.1 SegMiF [9] RGB-Infrared 54.8 U3M (Ours) RGB-Infrared 60.8 models utilizing RGB-Infrared modalities, a combination critical for enhancing material differentiation under varying illumination conditions. Notably, our model, U3M, achieves an impressive mIoU score of 60.8, which surpasses all other models listed.…”
Section: Methodsmentioning
confidence: 99%
“…I MAGES shot under low-light or backlit conditions are visually-terrible for viewers and also degenerate the performance of down-stream vision tasks, such as action recognition [1], [2], object detection [3], [4], and semantic segmentation [5], [6]. Many efforts have been tried to increase the visibility of these images to ameliorate their low quality, including upgrading imaging devices and designing image enhancement algorithms.…”
Section: Introductionmentioning
confidence: 99%