2022
DOI: 10.1609/aaai.v36i1.19987
|View full text |Cite
|
Sign up to set email alerts
|

Modality-Adaptive Mixup and Invariant Decomposition for RGB-Infrared Person Re-identification

Abstract: RGB-infrared person re-identification is an emerging cross-modality re-identification task, which is very challenging due to significant modality discrepancy between RGB and infrared images. In this work, we propose a novel modality-adaptive mixup and invariant decomposition (MID) approach for RGB-infrared person re-identification towards learning modality-invariant and discriminative representations. MID designs a modality-adaptive mixup scheme to generate suitable mixed modality images between RGB and infrar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 60 publications
(3 citation statements)
references
References 29 publications
0
3
0
Order By: Relevance
“…In this section, a comparison between the proposed method and state-of-the-art VI Re-ID approaches is conducted, including Cross-modal feature learning: DDAG [ 21 ], AGW [ 4 ], cm-SSFT [ 45 ], GLMC [ 22 ] (best method), MPANet [ 12 ], LBA [ 46 ], CM-NAS [ 19 ], MMN [ 47 ], MID [ 48 ]; Modal transformation: cmGAN [ 13 ], AlignGAN [ 49 ], Xmodal [ 17 ], SFANet [ 50 ], AGMNet [ 51 ] (best method), and PMT [ 25 ]. According to Table 1 , the results on the two datasets demonstrate that the proposed GSMEN outperforms state-of-the-art methods, achieving the outstanding performance.…”
Section: Methodsmentioning
confidence: 99%
“…In this section, a comparison between the proposed method and state-of-the-art VI Re-ID approaches is conducted, including Cross-modal feature learning: DDAG [ 21 ], AGW [ 4 ], cm-SSFT [ 45 ], GLMC [ 22 ] (best method), MPANet [ 12 ], LBA [ 46 ], CM-NAS [ 19 ], MMN [ 47 ], MID [ 48 ]; Modal transformation: cmGAN [ 13 ], AlignGAN [ 49 ], Xmodal [ 17 ], SFANet [ 50 ], AGMNet [ 51 ] (best method), and PMT [ 25 ]. According to Table 1 , the results on the two datasets demonstrate that the proposed GSMEN outperforms state-of-the-art methods, achieving the outstanding performance.…”
Section: Methodsmentioning
confidence: 99%
“…In order to alleviate the problem that single-modal visible light data cannot provide useful information at night (Zhu et al 2021;Rao et al 2021;Wang et al 2022a;Li, Wu, and Zheng 2021;Zheng et al 2015;Zhou et al 2023;Li et al 2022;Lei et al 2008), and eliminate the huge domain heterogeneity between cross-modal data (Zhang et al 2022;Tian et al 2021;Wu et al 2021;Chen et al 2021;Wu et al 2017;Nguyen et al 2017;Huang et al 2022;Farooq et al 2022;Kim et al 2023;Feng, Wu, and Zheng 2023;Zhang and Wang 2023;Zheng et al 2023;He et al 2015), Zheng et al (Zheng et al 2021) construct the first multi-modal person ReID dataset, RGBNT201. And Zheng et al (Li et al 2020) propose the benchmark multi-modal vehicle ReID datasets, RGBN300 and RGBNT100.…”
Section: Related Workmentioning
confidence: 99%
“…Inter modality mixing: [22] learns a dynamical and local linear interpolation between the different regions of cross-modality images in data-dependent fashion to mix up the RGB and infrared (IR) images. We explored both the static and dynamic mixing methods and found the static has a better performance.…”
Section: Related Workmentioning
confidence: 99%