2021
DOI: 10.48550/arxiv.2110.08985
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

StyleNeRF: A Style-based 3D-Aware Generator for High-resolution Image Synthesis

Jiatao Gu,
Lingjie Liu,
Peng Wang
et al.

Abstract: We propose StyleNeRF, a 3D-aware generative model for photo-realistic highresolution image synthesis with high multi-view consistency, which can be trained on unstructured 2D images. Existing approaches either cannot synthesize highresolution images with fine details or yield noticeable 3D-inconsistent artifacts. In addition, many of them lack control over style attributes and explicit 3D camera poses. StyleNeRF integrates the neural radiance field (NeRF) into a style-based generator to tackle the aforemention… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
101
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(101 citation statements)
references
References 30 publications
0
101
0
Order By: Relevance
“…GIRAFFE [26] proposes compositional neural radiance for image rendering and accelerates the rendering process with ConvNets. StyleNeRF [11] proposes the NeRF path regularization to enforce the output to be closer to the output of NeRF that has multi-view consistency. RGBD-GAN [27] samples two camera parameters to synthesize RGBD images from two views and then warps them to each other to ensure 3D consistency.…”
Section: Related Workmentioning
confidence: 99%
“…GIRAFFE [26] proposes compositional neural radiance for image rendering and accelerates the rendering process with ConvNets. StyleNeRF [11] proposes the NeRF path regularization to enforce the output to be closer to the output of NeRF that has multi-view consistency. RGBD-GAN [27] samples two camera parameters to synthesize RGBD images from two views and then warps them to each other to ensure 3D consistency.…”
Section: Related Workmentioning
confidence: 99%
“…To achieve this goal, the literature mainly follows two directions. The first line of works [21,34,44,46,72] utilize 3D-aware features to represent a scene, and apply a neural renderer, typically a CNN, on top of them for realistic image synthesis. For example, HoloGAN [44] and BlockGAN [45] learn lowresolution voxel features for objects, project them onto 2D image plane, and apply a StyleGAN-like [28] CNN to gen- erate higher-resolution images.…”
Section: Related Workmentioning
confidence: 99%
“…Giraffe [46] and GANcraft [22] instead use 3D volumetric rendering to generate 2D feature maps for the subsequent image generation. Following a similar idea, some works concurrent to ours [21,72] focus on designing better rendering networks to enable 3D-aware image generation at very high resolution. Nevertheless, an inevitable problem of these methods is the sacrifice of exact multiview consistency due to the learned black-box rendering.…”
Section: Related Workmentioning
confidence: 99%
“…Supervised Image Decomposition. Though producing impressive rendering quality, the geometry from 3D-aware generators is often incomplete or inaccurate [11,21]. As a result, the model tends to bias the image quality for highly sampled camera views (e.g.…”
Section: Introductionmentioning
confidence: 99%