2022
DOI: 10.1155/2022/6100292
|View full text |Cite
|
Sign up to set email alerts
|

MRFDCNet: Multireceptive Field Dense Connection Network for Real-Time Semantic Segmentation

Abstract: Semantic segmentation is widely used in automatic driving systems. To quickly and accurately classify objects in emergency situations, a large number of images need to be processed per second. To make a semantic segmentation model run on hardware with low memory and limited computing capacity, this paper proposes a real-time semantic segmentation network called MRFDCNet. This architecture is based on our proposed multireceptive field dense connection (MRFDC) module. The module uses one depthwise separable conv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 30 publications
0
2
0
Order By: Relevance
“…Notable transformer-based models include SETR [23], SegFormer [24] and RTFormer [25]. Additionally, the latest research has integrated CNNs and transformers, resulting in hybrid models like Vision Mamba [26] and VMamba [27]. Each of these models presents unique strengths and weaknesses [28].…”
Section: Semantic Segmentationmentioning
confidence: 99%
“…Notable transformer-based models include SETR [23], SegFormer [24] and RTFormer [25]. Additionally, the latest research has integrated CNNs and transformers, resulting in hybrid models like Vision Mamba [26] and VMamba [27]. Each of these models presents unique strengths and weaknesses [28].…”
Section: Semantic Segmentationmentioning
confidence: 99%
“…U-Transformer [29] and EUT [30] modify the cross-attention in the original transformer to leverage the information from the encoder, allowing a fine spatial recovery in the decoder. The cross-resolution attention employed by RTFormer [31] enables the gathering of comprehensive contextual information for high-resolution features.…”
Section: Vision Transformermentioning
confidence: 99%
“…Following BiseNet (Yu et al 2018), BiSeNetV2 (Yu et al 2021) and STDC (Fan et al 2021) make further efforts to strengthen the capability to extract rich long-range context or reduce the computational costs of the spatial branch. To balance inference speed and accuracy, DDRNet (Pan et al 2022), RTFormer (Wang et al 2022), and SeaFormer (Wan et al 2023) adopt a feature-sharing architecture that divides spatial and contextual features at the deep stages, as shown in Figure 2(b). However, these methods introduce dense fusion modules between two branches to boost the semantic information of extracted features.…”
Section: Introductionmentioning
confidence: 99%