2022
DOI: 10.1016/j.patcog.2021.108246
|View full text |Cite
|
Sign up to set email alerts
|

Integrated generalized zero-shot learning for fine-grained classification

Abstract: Embedding learning (EL) and feature synthesizing (FS) are two of the popular categories of fine-grained GZSL methods. EL or FS using global features cannot discriminate fine details in the absence of local features. On the other hand, EL or FS methods exploiting local features either neglect direct attribute guidance or global information. Consequently, neither method performs well.In this paper, we propose to explore global and direct attribute-supervised local visual features for both EL and FS categories in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(4 citation statements)
references
References 24 publications
0
4
0
Order By: Relevance
“…At last, we evaluate SAT on object detection and identify images of bicycle and electric bicycle. According to our training protocol, SAT model appears perfect performance, such as mAP (Albuquerque et al,2021) and maxDets (Shermin et al,2021). The Average Precision (AP) reaches 95%, the small area metric reaches 90%, the medium area metric reaches 94.8% and the large area metric reaches 94.5%, as shown in Table 4.…”
Section: Object Detectionmentioning
confidence: 83%
“…At last, we evaluate SAT on object detection and identify images of bicycle and electric bicycle. According to our training protocol, SAT model appears perfect performance, such as mAP (Albuquerque et al,2021) and maxDets (Shermin et al,2021). The Average Precision (AP) reaches 95%, the small area metric reaches 90%, the medium area metric reaches 94.8% and the large area metric reaches 94.5%, as shown in Table 4.…”
Section: Object Detectionmentioning
confidence: 83%
“…This model is trained to master conditional probability distribution; P(y|x, T). This learning type is suitable for machine translation, where the label is absent in the target language [244][245][246].…”
Section: Zero-shot Learningmentioning
confidence: 99%
“…Various modalities have been used as auxiliary information: word embeddings (Frome et al, 2013;Xian et al, 2016), hierarchical embeddings (Kampffmeyer et al, 2019), attributes (Farhadi et al, 2009;Akata et al, 2015) or Wikipedia articles (Elhoseiny et al, 2017;Zhu et al, 2018;Elhoseiny et al, 2016;Qiao et al, 2016). Most recent work uses generative models conditioned on class descriptions to synthesize training examples for unseen categories (Long et al, 2017;Kodirov et al, 2017;Felix et al, 2018;Xian et al, 2019;Vyas et al, 2020;, attention-enabled feature extractors (Yu et al, 2018;Zhu et al, 2019;Shermin et al, 2022;Chen et al, 2022). The multi-modal and often finegrained nature of the standard and generalised (G)ZSL task renders it related to our problem.…”
Section: Zero/few Shot Learningmentioning
confidence: 99%