2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298904
|View full text |Cite
|
Sign up to set email alerts
|

Deep roto-translation scattering for object classification

Abstract: Dictionary learning algorithms or supervised deep convolution networks have considerably improved the efficiency of predefined feature representations such as SIFT. We introduce a deep scattering convolution network, with complex wavelet filters over spatial and angular variables. This representation brings an important improvement to results previously obtained with predefined features over object image databases such as Caltech and CIFAR. The resulting accuracy is comparable to results obtained with unsuperv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
187
0
1

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 157 publications
(188 citation statements)
references
References 23 publications
0
187
0
1
Order By: Relevance
“…It is worth pointing out on CIFAR-10, the other reported methods [22,7,26,21,10] are usually based on different unsupervised networks and supervised classifiers for evaluation, making it difficult to make a direct comparison between them. The results still suggest that the state-of-the-art performances can be reached by AETs, as their error rates are very close to the pre-assumptive lower bound set by the fully supervised counterpart.…”
Section: Resultsmentioning
confidence: 99%
“…It is worth pointing out on CIFAR-10, the other reported methods [22,7,26,21,10] are usually based on different unsupervised networks and supervised classifiers for evaluation, making it difficult to make a direct comparison between them. The results still suggest that the state-of-the-art performances can be reached by AETs, as their error rates are very close to the pre-assumptive lower bound set by the fully supervised counterpart.…”
Section: Resultsmentioning
confidence: 99%
“…Roto-Scat + SVM [28] 17.7 ExamplarCNN [11] 15.7 DCGAN [32] 17.2 Scattering [27] 15.3 RotNet + non-linear [14] 10.94 RotNet + conv [14] 8.84 AET-affine + non-linear [37] 9.77 AET-affine + conv [37] 8.05 AET + non-linear [37] 9.41 AET + conv [37] 7.82 AVT + non-linear 8.96 AVT + conv 7.75 Table 2: Error rates of different classifiers trained on top of the learned representations on CIFAR 10, where n-FC denotes a classifier with n fully connected layers and conv denotes the third NIN block as a convolutional classifier. Two AET variants are chosen for a fair direct comparison since they are based on the same architecture as the AVT and have outperformed the other unsupervised representations before [37].…”
Section: Methodsmentioning
confidence: 99%
“…Recently there has been an explosion of interest into CNNs with predefined transformation equivariances, beyond translation [2,3,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19] [20, 21,22,23]. However, with the exception of Cohen and Welling [23] (projections on sphere), Kondor [22] (point clouds), and Thomas et al [20] (point clouds), these have mainly focused on the 2D scenario.…”
Section: Related Workmentioning
confidence: 99%