2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW) 2021
DOI: 10.1109/iccvw54120.2021.00202
|View full text |Cite
|
Sign up to set email alerts
|

Trans4Trans: Efficient Transformer for Transparent Object Segmentation to Help Visually Impaired People Navigate in the Real World

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 68 publications
(23 citation statements)
references
References 49 publications
0
23
0
Order By: Relevance
“…In Table 8, we benchmark more than 10 semantic segmentation methods. We compare our models against RGB-only methods covering CNNbased SwiftNet [125], Fast-SCNN [126], CGNet [127], and DeepLabV3+ [128], as well as transformer-based Swin [24], SegFormer [59], and Trans4Trans [3]. We also include multimodal methods, spanning RFNet [1] designed for road-driving scene segmentation and ISSAFE [12], the only known RGB-Event method designed for traffic accident scene segmentation, as well as SA-Gate [8], a state-of-the-art RGB-D segmentation method.…”
Section: Results On Rgb-event Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…In Table 8, we benchmark more than 10 semantic segmentation methods. We compare our models against RGB-only methods covering CNNbased SwiftNet [125], Fast-SCNN [126], CGNet [127], and DeepLabV3+ [128], as well as transformer-based Swin [24], SegFormer [59], and Trans4Trans [3]. We also include multimodal methods, spanning RFNet [1] designed for road-driving scene segmentation and ISSAFE [12], the only known RGB-Event method designed for traffic accident scene segmentation, as well as SA-Gate [8], a state-of-the-art RGB-D segmentation method.…”
Section: Results On Rgb-event Datasetmentioning
confidence: 99%
“…S EMANTIC segmentation is an essential task in computer vision, which aims to transform an image input into its underlying semantically meaningful regions and enables a pixelwise dense scene understanding for many real-world applications such as automated vehicles, robotics navigation, and augmented reality [1], [2], [3]. Over the last years, pixel-wise semantic segmentation of RGB images has gained an increasing amount of attention and made significant progress on segmentation accuracy [4], [5], [6].…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, when the moving object gets closer to the user and its speed is relatively fast, the reminder of potential risk can be passed to user. The object with high velocity is often dangerous for the visually impaired and this velocity information can enhance current obstacle avoidance modules that mainly use the depth information [12,19]. We also designed a questionnaire regarding the expected feedback form from our system.…”
Section: Discussionmentioning
confidence: 99%
“…Besides, visually impaired people find it hard to maintain proper social distances from others during the Covid-19 pandemic [13]. Some assistance systems tackle this issue through Simultaneous Localization And Mapping (SLAM) and deep learning approaches [12,19], to provide accurate guidance to visually impaired people, but they are less effective in highly dynamic scenarios. To address this problem, we propose a system to help people with visual impairments perceive dynamic objects in indoor environments and understand their motion.…”
Section: Introductionmentioning
confidence: 99%
“…in adverse conditions [20], [21]. Previous standard UDA and DG approaches are of course inapplicable, as they require access to source data, which is usually unavailable in a highly automated vehicle due to storage limitations.…”
Section: Introductionmentioning
confidence: 99%