2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00775
|View full text |Cite
|
Sign up to set email alerts
|

XMP-Font: Self-Supervised Cross-Modality Pre-training for Few-Shot Font Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 32 publications
(13 citation statements)
references
References 21 publications
0
5
0
Order By: Relevance
“…To address this, SA-VAE (Sun et al 2017) and EMD (Zhang, Zhang, and Cai 2018) generate unseen fonts by disentangling style and content representations. To enable the generator to capture local style characteristics, some methods (Wu, Yang, and Hsu 2020;Huang et al 2020;Cha et al 2020;Park et al 2021a,b;Liu et al 2022;Kong et al 2022)…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…To address this, SA-VAE (Sun et al 2017) and EMD (Zhang, Zhang, and Cai 2018) generate unseen fonts by disentangling style and content representations. To enable the generator to capture local style characteristics, some methods (Wu, Yang, and Hsu 2020;Huang et al 2020;Cha et al 2020;Park et al 2021a,b;Liu et al 2022;Kong et al 2022)…”
Section: Related Workmentioning
confidence: 99%
“…Although these methods have achieved remarkable success in font generation, they still suffer from complex character generation and large style variation transfer, leading to severe stroke missing, artifacts, blurriness, layout errors, and style inconsistency as shown in Figure 1(b)(c). Retrospectively, most font generation approaches (Park et al 2021a,b;Xie et al 2021;Tang et al 2022;Liu et al 2022;Kong et al 2022;Wang et al 2023) adopt a GANbased (Goodfellow et al 2014) framework which potentially suffers from unstable training due to their adversarial training nature. Moreover, most of these methods perceive content information through only single-scale highlevel features, omitting the fine-grained details that are crucial to preserving the source content, especially for complex characters.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…To improve the quality of the generated images, [13] proposed the Deep Feature Similarity (DFS) architecture to leverage the feature similarity between the input content and style images to synthesize target images. Recently, researchers [9,19,20,[44][45][46] have made significant progress by exploiting the compositionality of compositional scripts. However, our experimental results indicate poor performance for the constructed multi-language dataset.…”
Section: Font Generationmentioning
confidence: 99%
“…Disentanglement is mostly used for few-shot or one-shot font generation tasks [24,25,26,27,28,29,30]. Additional techniques for inter-or intra-radical style consistency for compounded characters with multiple radicals (such as Chinese and Korean letters) are introduced [26,31,32,33]. AGISNet [28] and TET-GAN [29,30] are proposed for dealing with more decorative and colorful font styles.…”
Section: Disentanglement For Font Imagesmentioning
confidence: 99%