2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00541
|View full text |Cite
|
Sign up to set email alerts
|

4D Visualization of Dynamic Events From Unconstrained Multi-View Videos

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
42
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 63 publications
(43 citation statements)
references
References 41 publications
1
42
0
Order By: Relevance
“…The recent neural rendering techniques bring huge potential for compelling photo-realistic free-viewpoint video generation via neural view blending [Meshry et al 2019a;Thies et al 2018] or neural scene modeling [Bemana et al 2020;Mildenhall et al 2020a;Wu et al 2020b]. Open4D [Bansal et al 2020] generates a free-viewpoint video enabling occlusion removal and time-freezing effects using around 15 mobile cameras, which is similar to our work. Such datadriven approaches get rid of the heavy reliance on reconstruction accuracy or the extremely dense capture setting.…”
Section: Introductionsupporting
confidence: 73%
See 1 more Smart Citation
“…The recent neural rendering techniques bring huge potential for compelling photo-realistic free-viewpoint video generation via neural view blending [Meshry et al 2019a;Thies et al 2018] or neural scene modeling [Bemana et al 2020;Mildenhall et al 2020a;Wu et al 2020b]. Open4D [Bansal et al 2020] generates a free-viewpoint video enabling occlusion removal and time-freezing effects using around 15 mobile cameras, which is similar to our work. Such datadriven approaches get rid of the heavy reliance on reconstruction accuracy or the extremely dense capture setting.…”
Section: Introductionsupporting
confidence: 73%
“…However, these approaches above still cannot provide a wide-range free-viewing of a large dynamic scene, let alone editing various dynamic entities. To model dynamic scenes, the prior works [Bansal et al 2020;Carranza et al 2003;Lipski et al 2010;Zitnick et al 2004] require multi-view, time-synchronized videos as input for rendering various space-time visual effects. Zitnick et al [Zitnick et al 2004] use depth maps estimated from multi-view stereo to guide viewpoint interpolation.…”
Section: Related Workmentioning
confidence: 99%
“…A key advantage of our framework is that it provides 3D representations readily integrated modern neural rendering methods, which typically requires 3D and/or camera pose ground truth annotation at training time [14,48,4,72,63]. This usually implies a 2-stage process with running Structurefrom-Motion (SfM) as precursor to training.…”
Section: Introductionmentioning
confidence: 99%
“…Consequently, these methods rely on high-quality 3D reconstruction or dense perpixel correspondence, which requires dense multi-views. Recent work [5,38] combines classical methods with datadriven approaches by learning to correct the warped views of classical methods. The sequential pipeline in these methods [5,38] do not allow end-to-end learning.…”
Section: Introductionmentioning
confidence: 99%
“…Recent work [5,38] combines classical methods with datadriven approaches by learning to correct the warped views of classical methods. The sequential pipeline in these methods [5,38] do not allow end-to-end learning.…”
Section: Introductionmentioning
confidence: 99%