2022
DOI: 10.1007/978-3-031-20050-2_11
|View full text |Cite
|
Sign up to set email alerts
|

Unleashing Transformers: Parallel Token Prediction with Discrete Absorbing Diffusion for Fast High-Resolution Image Generation from Vector-Quantized Codes

Abstract: We introduce ∞-Diff, a generative diffusion model which directly operates on infinite resolution data. By randomly sampling subsets of coordinates during training and learning to denoise the content at those coordinates, a continuous function is learned that allows sampling at arbitrary resolutions. In contrast to other recent infinite resolution generative models, our approach operates directly on the raw data, not requiring latent vector compression for context, using hypernetworks, nor relying on discrete c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(14 citation statements)
references
References 41 publications
0
14
0
Order By: Relevance
“…Some studies, such as SED ( Strudel et al, 2022 ), have pointed out the limitations of low sampling efficiency in diffusion models, which is indeed a drawback of diffusion models. To address this issue, in the field of computer vision, there have been a few studies that propose different efficient sampling strategies ( Bond-Taylor et al, 2022 ; Xiao, Kreis & Vahdat, 2022 ; Watson et al, 2022 ; Vahdat, Kreis & Kautz, 2021 ; Zhang & Chen, 2021 ). These methods have demonstrated the ability to double the sampling speed in many cases.…”
Section: Future Directionsmentioning
confidence: 99%
“…Some studies, such as SED ( Strudel et al, 2022 ), have pointed out the limitations of low sampling efficiency in diffusion models, which is indeed a drawback of diffusion models. To address this issue, in the field of computer vision, there have been a few studies that propose different efficient sampling strategies ( Bond-Taylor et al, 2022 ; Xiao, Kreis & Vahdat, 2022 ; Watson et al, 2022 ; Vahdat, Kreis & Kautz, 2021 ; Zhang & Chen, 2021 ). These methods have demonstrated the ability to double the sampling speed in many cases.…”
Section: Future Directionsmentioning
confidence: 99%
“…Diffusion models are widely used in Artificial Intelligence Generated Content (AIGC), which are of great importance in generative models. Diffusion models have illustrated their power in image generation [14,18,36,45], detection [7], segmentation [1,3,8], image-to-image translation [23,24], super resolution [38], image inpainting [5], image editing [31], text-to-image [2,15,25], video generation [17,40], point cloud [30,49,52], and human motion synthesis [22,39]. DDPM-Segmentation [3] is the first work to apply the diffusion model to semantic segmentation, which pre-trains a diffusion model and then trains classifiers for each pixel.…”
Section: Diffusion Modelmentioning
confidence: 99%
“…Diffusion models naturally have the ability to denoise noisy samples to the ideal data. Recently, the diffusion probability models (DPM) have illustrated their great power in generative tasks [5,22,25,31], but their potential in BEV perception tasks has not been fully explored. In this work, we propose DiffBEV, a novel framework that utilizes conditional DPM to improve quality of the BEV feature and push the boundary of BEV perception.…”
Section: Introductionmentioning
confidence: 99%
“…More recently, diffusion model-based approaches are gaining increasing popularity due to their exceptional results in image generation (Sohl-Dickstein et al, 2015;Ho et al, 2020;Yang et al, 2022;Bond-Taylor et al, 2021;Chung et al, 2022b;Batzolis et al, 2022;Bansal et al, 2022;Liu et al, 2022;Ku et al, 2022;Benton et al, 2022;Horwitz & Hoshen, 2022;Horita et al, 2022;Li et al, 2022). Besides, these methods enjoy the advantage of being able to perform inpainting without the need for degradation-specific training (Song & Ermon, 2019a).…”
Section: Related Workmentioning
confidence: 99%