2014
DOI: 10.48550/arxiv.1409.0575
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ImageNet Large Scale Visual Recognition Challenge

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
297
1
5

Year Published

2017
2017
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 289 publications
(303 citation statements)
references
References 51 publications
0
297
1
5
Order By: Relevance
“…For all our experiments, we adopt the "Base" ViT model architecture (Dosovitskiy et al, 2020) pretrained on Ima-geNet (Russakovsky et al, 2014). Unless differently specified, we use clips of size 8×224×224, with frames sampled at a rate of 1/16.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…For all our experiments, we adopt the "Base" ViT model architecture (Dosovitskiy et al, 2020) pretrained on Ima-geNet (Russakovsky et al, 2014). Unless differently specified, we use clips of size 8×224×224, with frames sampled at a rate of 1/16.…”
Section: Methodsmentioning
confidence: 99%
“…We show these results in Figure 5, where we also compare our method with SlowFast R50 (Feichtenhofer et al, 2019b), and I3D R50 (Carreira & Zisserman, 2017) trained on the same subsets. We note that all 3 architectures are pretrained on ImageNet (Russakovsky et al, 2014).…”
Section: The Importance Of Pretraining and Dataset Scalementioning
confidence: 99%
“…In this section, we assess the practicality of XOR query and the proposed algorithm using real data collected from Amazon Mechanical Turk. We designed a binary classification task using 600 images of dogs and cats sampled from ImageNet [25]. Each human intelligent task (HIT) was designed to include 20 degree-1 queries and 20 degree-4 XOR queries.…”
Section: B Real Datasetsmentioning
confidence: 99%
“…However, the reported approach has substantial shortcomings and cannot be used for fast conversion of pre-trained neural networks on mobile devices. First of all it has a requirement to train threshold on the full Im-ageNet dataset [19]. Besides, it has no examples demonstrating the accuracy of networks used as standards for mobile platforms.…”
Section: Quantization With Training / Fine-tuningmentioning
confidence: 99%
“…Adam optimizer [13] is used for training, and cosine annealing with the reset of optimizer parameters -for learning rate. Training is carried out on approximately 10% part of Im-ageNet dataset [19]. Testing is done on the validation set.…”
Section: Scaling the Weights For Mobilenet-v2 (With Relu6)mentioning
confidence: 99%