2019
DOI: 10.1109/access.2019.2953454
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Layer Autoencoder for Zero-Shot Learning

Abstract: Zero-shot learning (ZSL) is the task of recognizing samples from their related classes which have never been seen during model training. ZSL is generally realized through learning a common embedding space where both high dimensional visual features and some pre-defined semantics can be mapped. However, this kind of solutions usually suffers from domain shift. In addition, the limitation and subjectivity of manual semantic information can also affect the classification results. To address these challenges, this… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(1 citation statement)
references
References 41 publications
0
1
0
Order By: Relevance
“…The space embedding methods [ 8 , 9 , 10 ] rely on an embedding space, in which the attribute migration from seen classes to unseen classes is completed. The generative model methods utilize different generative models, such as generative adversarial networks (GANs) [ 16 , 17 ], variational autoencoders (VAEs) [ 18 , 19 ], and flow-based models (Flows) [ 20 ] to directly generate visual features of unseen classes, and then transform the zero-shot learning problem into a traditional supervised learning problem. For example, [ 16 ] has proposed a triple discriminator GAN (TDGAN), which employs a GAN with three discriminators to synthesize visual features for images of unseen classes.…”
Section: Related Workmentioning
confidence: 99%
“…The space embedding methods [ 8 , 9 , 10 ] rely on an embedding space, in which the attribute migration from seen classes to unseen classes is completed. The generative model methods utilize different generative models, such as generative adversarial networks (GANs) [ 16 , 17 ], variational autoencoders (VAEs) [ 18 , 19 ], and flow-based models (Flows) [ 20 ] to directly generate visual features of unseen classes, and then transform the zero-shot learning problem into a traditional supervised learning problem. For example, [ 16 ] has proposed a triple discriminator GAN (TDGAN), which employs a GAN with three discriminators to synthesize visual features for images of unseen classes.…”
Section: Related Workmentioning
confidence: 99%