2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.00810
|View full text |Cite
|
Sign up to set email alerts
|

GistNet: a Geometric Structure Transfer Network for Long-Tailed Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 35 publications
(11 citation statements)
references
References 14 publications
0
11
0
Order By: Relevance
“…During training, the tail class-specific features are fused with the head class-generic features to generate new features to augment the tail classes. GIST [38] proposes to transfer the geometric information of the feature distribution boundaries of the head classes to the tail classes by increasing the classifier weights of the tail classes. The motivation of the recently proposed CMO [45] is very intuitive, it argues that the images from the head classes have rich backgrounds, so the images from the tail classes can be pasted directly onto the rich background images of the head classes to increase the richness of the tail images.…”
Section: Information Augmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…During training, the tail class-specific features are fused with the head class-generic features to generate new features to augment the tail classes. GIST [38] proposes to transfer the geometric information of the feature distribution boundaries of the head classes to the tail classes by increasing the classifier weights of the tail classes. The motivation of the recently proposed CMO [45] is very intuitive, it argues that the images from the head classes have rich backgrounds, so the images from the tail classes can be pasted directly onto the rich background images of the head classes to increase the richness of the tail images.…”
Section: Information Augmentationmentioning
confidence: 99%
“…The imbalance of sample numbers in the dataset gives rise to the challenge of long-tailed visual recognition. Most previous works assume that head classes are always easier to be learned than tail classes, e.g., class re-balancing [8,14,24,34,37,46,52], information augmentation [23,31,35,38,39,44,56,64,67], decoupled training [10,16,29,30,71,76], and ensemble learning [20,36,57,58,61,72,77] have been proposed to improve the performance of tail classes. However, recent studies [3,50] have shown that classification dif- ficulty is not always correlated with the number of samples, e.g., the performance of some tail classes is even higher than that of the head classes.…”
Section: Introductionmentioning
confidence: 99%
“…To verify the effectiveness, we compare the proposed method ResCom against four groups of stateof-the-art methods on common-used long-tailed datasets: Logits modification methods, including LDAM [3], Causal Norm [44], Balanced Softmax [37], and LADE [22]. Architecture modification methods, like BBN [56], ELF [15], ResLT [11], GistNet [33], RIDE [49], and DiVE [20]. Multi-stage methods such as Decouple [27], DisAlign [53], MiSLAS [54], SSD [30].…”
Section: Comparison With State-of-the-artsmentioning
confidence: 99%
“…Class rebalancing [2,3,7,14,17,18,20,23,26,28,33,34,37,39,41,42,45], for example, aims to boost the weight of losses arising from tail classes, thereby pushing the decision boundary away from the tail class and improving the probability of correctly classifying the underlying distribution. Information augmentation [4,8,12,15,19,21,22,24,25,32,36,38,39,43], on the other hand, expands the observed distribution of the tail classes by introducing prior knowledge to facilitate the model learning of the underlying distribution. It is important to note that these methods default to the model being able to learn adequately and fairly at least for the samples in the training domain.…”
Section: Introductionmentioning
confidence: 99%