2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00471
|View full text |Cite
|
Sign up to set email alerts
|

Feature Erasing and Diffusion Network for Occluded Person Re-Identification

Abstract: Occlusion perturbation presents a significant challenge in person re-identification (re-ID), and existing methods that rely on external visual cues require additional computational resources and only consider the issue of missing information caused by occlusion. In this paper, we propose a simple yet effective framework, termed Erasing, Transforming, and Noising Defense Network (ETNDNet), which treats occlusion as a noise disturbance and solves occluded person re-ID from the perspective of adversarial defense.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
47
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 115 publications
(47 citation statements)
references
References 71 publications
0
47
0
Order By: Relevance
“…On the P-DukeMTMC dataset, our FCFormer outperforms QPM by 1.0% and 8.3% in terms of Rank-1 accuracy and mAP respectively. On the Occluded-REID dataset, FCFormer produces the comparable results with FED [28], achieving 84.9% Rank-1. We fail to achieve the highest Rank-1 performance on Occluded-REID for the following reasons: First, the Transformer has poor cross-domain generalization ability on small datasets.…”
Section: Performance Under Transfer Settingmentioning
confidence: 89%
See 1 more Smart Citation
“…On the P-DukeMTMC dataset, our FCFormer outperforms QPM by 1.0% and 8.3% in terms of Rank-1 accuracy and mAP respectively. On the Occluded-REID dataset, FCFormer produces the comparable results with FED [28], achieving 84.9% Rank-1. We fail to achieve the highest Rank-1 performance on Occluded-REID for the following reasons: First, the Transformer has poor cross-domain generalization ability on small datasets.…”
Section: Performance Under Transfer Settingmentioning
confidence: 89%
“…Table II shows the results on three occluded datasets, i.e., Occluded-Duke and P-DukeMTMC. As table shows, two classes of methods will be compared: CNN based ReID methods [3], [4], [6], [7], [10], [13], [16], [29]- [31], [33]- [39], [42]- [44] and Transformer based ReID methods [5], [20], [21], [23], [28]. The results show that the proposed FCFormer consistently achieves the competitive performance on Occluded datasets.…”
Section: Comparison With the State-of-the-art Modelsmentioning
confidence: 99%
“…However, the incomplete information caused by occlusion increases its difficulty significantly. Recent works [25] [36] attempted to solve the occlusion problem by utilizing pose information provided by the estimation algorithm as crucial supplementary guidance. Miao et al [25] proposed a pose-guided feature alignment (PGFA) method that employs pose key-points information to generate the attention map to direct discriminative feature learning.…”
Section: Occluded Person Re-identificationmentioning
confidence: 99%
“…Consequently, these methods significantly degrade when dealing with occluded images. While recent endeavors have facilitated person Re-ID under occlusion conditions [39,36,31,18,28,16], two main problems associated with occlusions still need to be addressed. Firstly, the presence of obstacles will vanish some parts of the human body, missing and misaligned extracted features.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, with the exploration of attention mechanisms for various vision tasks, it has also been adopted for occluded person Re-ID to eliminate the interference of noisy information [44,29,12]. During the process of attention learning, many data augmentation strategies [1,36,51] generate artificial occlusion, which directs the attention to person and forces it to avoid occluded regions. Currently, the most widely used artificial occlusion methods are random erasing [48] or using the background as occlusion [1].…”
Section: Introductionmentioning
confidence: 99%