2021
DOI: 10.1111/cgf.142619
|View full text |Cite
|
Sign up to set email alerts
|

Semantics‐Guided Latent Space Exploration for Shape Generation

Abstract: We introduce an approach to incorporate user guidance into shape generation approaches based on deep networks. Generative networks such as autoencoders and generative adversarial networks are trained to encode shapes into latent vectors, effectively learning a latent shape space that can be sampled for generating new shapes. Our main idea is to enable users to explore the shape space with the use of high-level semantic keywords. Specifically, the user inputs a set of keywords that describe the general attribut… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 33 publications
0
6
0
Order By: Relevance
“…AI-generated content (AIGC), also known as generative AI (GAI), recently gained a surge of development (Zhang et al 2023a,b;Cao et al 2023;Balaji et al 2022), covering several areas such as image (Ramesh et al 2021Saharia et al 2022;Yu et al 2022;Rombach et al 2022), text (Raffel et al 2020;Radford et al 2018Radford et al , 2019Brown et al 2020;Ope-nAI 2023;Vinyals et al 2015), 3D (Fu et al 2022;Jahan, Guan, and Van Kaick 2021;Liu et al 2022;Mildenhall et al 2021), and speech (Qian et al 2014;Ze, Senior, and Schuster 2013;Zen and Sak 2015). Since ChatGPT emerged, people have been amazed by its performance while recognizing the reasoning potential in the generated text content (Bang et al 2023).…”
Section: Introductionmentioning
confidence: 99%
“…AI-generated content (AIGC), also known as generative AI (GAI), recently gained a surge of development (Zhang et al 2023a,b;Cao et al 2023;Balaji et al 2022), covering several areas such as image (Ramesh et al 2021Saharia et al 2022;Yu et al 2022;Rombach et al 2022), text (Raffel et al 2020;Radford et al 2018Radford et al , 2019Brown et al 2020;Ope-nAI 2023;Vinyals et al 2015), 3D (Fu et al 2022;Jahan, Guan, and Van Kaick 2021;Liu et al 2022;Mildenhall et al 2021), and speech (Qian et al 2014;Ze, Senior, and Schuster 2013;Zen and Sak 2015). Since ChatGPT emerged, people have been amazed by its performance while recognizing the reasoning potential in the generated text content (Bang et al 2023).…”
Section: Introductionmentioning
confidence: 99%
“…Existing works (Chen et al (2018); Jahan et al (2021); Liu et al (2022)) typically rely on paired text-shape data for model training. Yet, collecting 3D shapes is already very challenging on its own, let alone the tedious manual annotations needed to construct the text-shape pairs.…”
Section: Introductionmentioning
confidence: 99%
“…Finally, an obvious benefit of our method in comparison to methods based on natural language [13,24] is that it needs no manual labeling of the data to enable the exploration. In comparison to sketching approaches [18], artistic skills are not required from the user, while in comparison to classic parametric models [8], the requirements on data pre-processing are not as stringent.…”
Section: Overview Of Our Methodsmentioning
confidence: 99%
“…Learning of the latent space can be accomplished with the use of an autoencoder network composed of an encoder and decoder [24]. The encoder takes a 3D shape as input, e.g., encoded as a 3D volume, and returns the latent representation of the shape.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation