2023
DOI: 10.1088/1361-6501/acd136
|View full text |Cite
|
Sign up to set email alerts
|

PCTNet: depth estimation from single structured light image with a parallel CNN-transformer network

Abstract: Recent approaches based on convolutional neural networks significantly improve the performance of structured light image depth estimation in structured light 3D measurement. However, it remains challenging to simultaneously preserve the global structure and local details of objects for the structured light images in complex scenes. In this paper, we design a parallel CNN-Transformer network, which consists of a CNN branch, a Transformer branch, a bidirectional feature fusion module (BFFM), and a cross-feature … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 30 publications
0
5
0
Order By: Relevance
“… [ 78 ] 2023 Structured light A parallel CNN transformer network is proposed to achieve an improved depth estimation for structured light images in complex scenes. [ 79 ] 2022 Time-of-Flight (TOF) DELTAR is proposed to enable lightweight Time-of-Flight sensors to measure high-resolution and accurate depth by collaborating with color images. [ 80 ] 2020 Time-of-Flight (TOF) Based on the principle and imaging characteristics of TOF cameras, a single pixel is considered as a continuous Gaussian source, and its differential entropy is proposed as an evaluation parameter.…”
Section: Depth Perception Methods Based On Computer Visionmentioning
confidence: 99%
See 1 more Smart Citation
“… [ 78 ] 2023 Structured light A parallel CNN transformer network is proposed to achieve an improved depth estimation for structured light images in complex scenes. [ 79 ] 2022 Time-of-Flight (TOF) DELTAR is proposed to enable lightweight Time-of-Flight sensors to measure high-resolution and accurate depth by collaborating with color images. [ 80 ] 2020 Time-of-Flight (TOF) Based on the principle and imaging characteristics of TOF cameras, a single pixel is considered as a continuous Gaussian source, and its differential entropy is proposed as an evaluation parameter.…”
Section: Depth Perception Methods Based On Computer Visionmentioning
confidence: 99%
“…Additionally, as the detection distance increases, the accuracy of structured light decreases. To address these issues, current research efforts have employed strategies such as increasing power and changing coding methods [ 77 , 78 , 79 ].…”
Section: Depth Perception Methods Based On Computer Visionmentioning
confidence: 99%
“…Wang et al proposed SwinConvUNet, which employs a self-attention mechanism and the Swin Transformer for fringe-to-depth conversion, aiming to extract both local and global features [35]. Similar to these works, other researchers [36][37][38][39] presented a few end-to-end network models, such as MSUNet++, PCTNet, DF-Dnet, and DCAHINet; they focus on depth recovery through diverse multiscale feature fusion modules. The team in [40] introduced LiteF2DNet, a lightweight deep learning framework designed to reduce network weights, and they tested it on computeraided design (CAD) objects.…”
Section: Related Workmentioning
confidence: 99%
“…The radar wave formed by external interference will cause obvious interference to the identification of adverse geologic bodies. (2) Most deep learning algorithms are built based on convolutional neural network (CNN), which tend to pay more attention to the local detailed features of the image and neglect the importance of global information in target inference (Zhou et al 2023, Zhu et al 2023. Adverse geologic bodies often need to be analyzed in conjunction with the overall characteristics of the image.…”
Section: Introductionmentioning
confidence: 99%