2019 IEEE International Conference on Image Processing (ICIP) 2019
DOI: 10.1109/icip.2019.8803360
|View full text |Cite
|
Sign up to set email alerts
|

Incorporating Luminance, Depth and Color Information by a Fusion-Based Network for Semantic Segmentation

Abstract: Semantic segmentation has made encouraging progress due to the success of deep convolutional networks in recent years. Meanwhile, depth sensors become prevalent nowadays; thus, depth maps can be acquired more easily. However, there are few studies that focus on the RGB-D semantic segmentation task. Exploiting the depth information effectiveness to improve performance is a challenge. In this paper, we propose a novel solution named LDFNet, which incorporates Luminance, Depth and Color information by a fusion-ba… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
35
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(35 citation statements)
references
References 13 publications
0
35
0
Order By: Relevance
“…Related applications in multimodal fusion networks can be found in [69,70,71]. Also, the 1 × 1 convolution layer is commonly used to allow complex and learnable interaction across modalities and channels [72,70]. Besides, attention mechanism has become a powerful tool for image recognition [73,74,75].…”
Section: Semantic Image Segmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…Related applications in multimodal fusion networks can be found in [69,70,71]. Also, the 1 × 1 convolution layer is commonly used to allow complex and learnable interaction across modalities and channels [72,70]. Besides, attention mechanism has become a powerful tool for image recognition [73,74,75].…”
Section: Semantic Image Segmentationmentioning
confidence: 99%
“…Hung et al [72] presented LDFNet that contains a well-designed encoder for the non-RGB branch, aiming to fully make use of luminance, depth, and color information. Recently, RFBNet [105] was proposed with an efficient fusion mechanism that explores the interdependence between the encoders (see Figure 5).…”
Section: Early Fusionmentioning
confidence: 99%
“…In (c) as proposed in ACNet [34] and RFBNet [35], in addition to paths for each modality, an additional path for feature fusion (i.e., a path colored by magenta in the figure) is employed. For the case where one of the input modalities is equal to the output modality, in (d) [36], [37], only one modality is regarded as dominant. For intensive analysis of this dominant modality, features extracted from other modalities are auxiliarily fused to those extracted from the dominant modality.…”
Section: A Insar Analysis and Its Extensionsmentioning
confidence: 99%
“…These contributions use a large dataset such as ImageNet [35] for pre-trained models. Recent segmentation techniques have distinct characteristics denoted by their design such as: (1) network topology: pooling indices [3], skip connection [34], multi-path refinement [27], pyramid pooling [47], fusionbased architecture [15] and dense connectivity [20], (2) varying input: colour RGB or RGB-D with depth [15], [19], depth and luminance [21], and illumination invariance [1], and (3) consideration of adverse-weather conditions [8], [14], [36]. As the main objective of this work is semantic scene segmentation under foggy weather conditions, recent studies in this specific domain are specifically presented in this section.…”
Section: A Semantic Segmentationmentioning
confidence: 99%
“…Despite the general trend of performance improvement within automotive scene understanding [4], [17], [27], [47], there is still significant room for improvement across the spectrum of non-ideal operating conditions. In parallel with using recent image segmentation techniques [15], [20], [21], [40], employing the concept of image-to-image translation to map one domain onto another [22], [48] is a useful step that enables accurate semantic segmentation performance under extreme weather conditions.…”
Section: Introductionmentioning
confidence: 99%