2021
DOI: 10.1109/access.2021.3058522
|View full text |Cite
|
Sign up to set email alerts
|

PA-MVSNet: Sparse-to-Dense Multi-View Stereo With Pyramid Attention

Abstract: We thank all the participants of the study for their time and useful comments. We appreciate the support from our colleagues and classmates.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1
1

Relationship

1
9

Authors

Journals

citations
Cited by 17 publications
(7 citation statements)
references
References 41 publications
0
5
0
Order By: Relevance
“…For example, MVSTR [87], LANet [84] and TransMVSNet [14] unitize the attention mechanisms to extract dense features with global contexts. PA-MVSNet [82] and AACVP-MVSNet [78] introduce self-attention layers for hierarchical features extraction, which is able to capture multi-scale matching clues for the subsequent depth inference task. AttMVS [38] introduces an attentionenhanced matching confidence volume to improve the matching robustness.…”
Section: Related Workmentioning
confidence: 99%
“…For example, MVSTR [87], LANet [84] and TransMVSNet [14] unitize the attention mechanisms to extract dense features with global contexts. PA-MVSNet [82] and AACVP-MVSNet [78] introduce self-attention layers for hierarchical features extraction, which is able to capture multi-scale matching clues for the subsequent depth inference task. AttMVS [38] introduces an attentionenhanced matching confidence volume to improve the matching robustness.…”
Section: Related Workmentioning
confidence: 99%
“…Unlike the convolutional neural networks (CNN) [20,21] models' local image pattern, Transformer employs self-attention [22] on a tokenized image patch to model the context information of the image. In recent studies, pure Transformer architectures were introduced to many visual recognition and image classification tasks and achieved promising results.…”
Section: Transformermentioning
confidence: 99%
“…Another solution is the PAMVSNet (Zhang et al, 2021), which uses a pyramidal attention module, obtaining more in formation from the original image and generating a signifi cant improvement in the representation of features. This so lution has improved image quality and less noise, while it has an increase the execution time for the generation of multi pyramidal views.…”
Section: Related Workmentioning
confidence: 99%