2017
DOI: 10.1109/tpami.2016.2643667
|View full text |Cite
|
Sign up to set email alerts
|

Write a Classifier: Predicting Visual Classifiers from Unstructured Text

Abstract: Abstract-People typically learn through exposure to visual concepts associated with linguistic descriptions. For instance, teaching visual object categories to children is often accompanied by descriptions in text or speech. In a machine learning context, these observations motivates us to ask whether this learning process could be computationally modeled to learn visual classifiers. More specifically, the main question of this work is how to utilize purely textual description of visual classes with no trainin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
39
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
3
2
1

Relationship

2
4

Authors

Journals

citations
Cited by 36 publications
(39 citation statements)
references
References 51 publications
0
39
0
Order By: Relevance
“…Baselines and Competing Methods: The performance of our approach is compared to six state-of-the-art algorithms: SJE [6], MCZSL [3], ZSLNS [36], ESZSL [39], WAC [14]. The source code of ESZSL and ZSLNS are available online, and we get the code of WAC [14,13] from its author. For MCZSL and SJE, since their source codes are not available, we directly copy the highest scores for non-attribute settings reported in [3,6].…”
Section: Performance Evaluationmentioning
confidence: 99%
See 2 more Smart Citations
“…Baselines and Competing Methods: The performance of our approach is compared to six state-of-the-art algorithms: SJE [6], MCZSL [3], ZSLNS [36], ESZSL [39], WAC [14]. The source code of ESZSL and ZSLNS are available online, and we get the code of WAC [14,13] from its author. For MCZSL and SJE, since their source codes are not available, we directly copy the highest scores for non-attribute settings reported in [3,6].…”
Section: Performance Evaluationmentioning
confidence: 99%
“…Note that some of these methods were applied on attributes prediction (e.g., ZSLNS [36], SynC [10], ESZSL [39] ) or image-sentence similarity (e.g.,Order Embedding [46]). We used the publicly available code of these methods and other text-based methods like (ZSLNS [36], WAC [14], WAC-kernel [13]) to apply them on our setting. Note that the conventional split setting for zero-shot learning is Super-Category Shared splitting (i.e.…”
Section: Performance Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…[18]). The proposed approaches mainly rely on learning a similarity function between text descriptions and images (either linearly [8,25] or non-linearly via deep neural networks [18] or kernels [7]). At test-time, classification is performed by associating the image to the class with the highest similarity to the corresponding class-level text.…”
Section: Related Workmentioning
confidence: 99%
“…Many existing zero-shot learning approaches make use of deep features (i.e. vectors of activations from some late layer in a network pretrained on some largescale task) to learn joint embeddings with class descriptions [32,1,3,5,23,8,9,7]. These higher-level features collapse many underlying concepts in the pursuit of class discrimination; consequentially, accessing lower-level concepts and recombining them in new ways to represent novel classes is difficult with these features.…”
mentioning
confidence: 99%