2018 IEEE International Conference on Multimedia and Expo (ICME) 2018
DOI: 10.1109/icme.2018.8486612
|View full text |Cite
|
Sign up to set email alerts
|

Visual Confusion Label Tree for Image Classification

Abstract: Convolution neural network models are widely used in image classification tasks. However, the running time of such models is so long that it is not the conforming to the strict real-time requirement of mobile devices. In order to optimize models and meet the requirement mentioned above, we propose a method that replaces the fully-connected layers of convolution neural network models with a tree classifier. Specifically, we construct a Visual Confusion Label Tree based on the output of the convolution neural ne… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 4 publications
0
5
0
Order By: Relevance
“…DDT [25] Despite efforts to reduce computational costs, the model may still require significant resources for training and inference. • Output Layer: The output layer merges the input image and the deconvolved image using an addition function.…”
Section: Methods Limitationmentioning
confidence: 99%
See 1 more Smart Citation
“…DDT [25] Despite efforts to reduce computational costs, the model may still require significant resources for training and inference. • Output Layer: The output layer merges the input image and the deconvolved image using an addition function.…”
Section: Methods Limitationmentioning
confidence: 99%
“…DDT [25] is an advanced neural network specifically developed for the purpose of image denoising. It utilises a distinctive architecture that effectively combines local and global data in a simultaneous manner.…”
Section: Related Workmentioning
confidence: 99%
“…To obtain a suitable number of superclasses, we perform a series of ablation experiments on our MMF model with a singe label structure H A . Referring to the number of superclasses in H S , We vary the number of H A 's superclasses in [18,20,25,30] for CIFAR100 , and [15,18,20,30,40,50] for Car196. According to the results of Table 1, we select the H A with 30 superclasses for CIFAR100, and 15 for Car196.…”
Section: Ablation Studymentioning
confidence: 99%
“…However, these relations may be inconsistent with the appearances, which weakens the performance of classification tasks. Therefore, a lot of work builds visual information tree structures [13,3,19,20,21,16,11,29]. Some build the tree structures based on the confusion matrix [13,3,19,20,21], which is constructed by the results of a classifier.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation