Proceedings of 2010 IEEE International Symposium on Circuits and Systems 2010
DOI: 10.1109/iscas.2010.5537907
|View full text |Cite
|
Sign up to set email alerts
|

Convolutional networks and applications in vision

Abstract: Abstract-Intelligent tasks, such as visual perception, auditory perception, and language understanding require the construction of good internal representations of the world (or "features"), which must be invariant to irrelevant variations of the input while, preserving relevant information. A major question for Machine Learning is how to learn such good features automatically. Convolutional Networks (ConvNets) are a biologicallyinspired trainable architecture that can learn invariant features. Each stage in a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
925
0
16

Year Published

2012
2012
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 1,700 publications
(942 citation statements)
references
References 38 publications
1
925
0
16
Order By: Relevance
“…A logarithmic non-linearity is first applied to invariant scattering coefficients to linearize their power law behavior across scales. This is similar to the normalization strategies used by bag of words [10] and deep neural networks [5].…”
Section: Hierarchical Architecturementioning
confidence: 88%
See 3 more Smart Citations
“…A logarithmic non-linearity is first applied to invariant scattering coefficients to linearize their power law behavior across scales. This is similar to the normalization strategies used by bag of words [10] and deep neural networks [5].…”
Section: Hierarchical Architecturementioning
confidence: 88%
“…Most image representations build localized invariants over small image patches, for example with SIFT descriptors [15]. These invariant coefficients are then aggregated into more invariant global image descriptors, for example with bag of words [10] or multiple layers of deep neural network [4,5]. We follow a similar strategy by first computing invariants over image patches and then aggregating them at the global image scale.…”
Section: Hierarchical Architecturementioning
confidence: 99%
See 2 more Smart Citations
“…For instance, it would be interesting to have a global network able to decompose the "concept" of a face as a combination of two eyes, one nose and one mouth, which would be themselves "concepts" from a lower layer. For this reason, using the proposed architecture on top of a standard convolutional network [45,42] may not be an optimal solution, since it would force the network to categorize each image as a whole. This rather suggests to fully integrate the proposed network into a convolutional architecture, which would progressively categorize subparts of an image in a hierarchical way.…”
Section: Perspectives and Future Workmentioning
confidence: 99%