TY - JOUR
T1 - Exploring modality-shared appearance features and modality-invariant relation features for cross-modality person Re-IDentification
AU - Huang, Nianchang
AU - Liu, Jianan
AU - Luo, Yongjiang
AU - Zhang, Qiang
AU - Han, Jungong
N1 - Funding Information:
This work is supported by the National Natural Science Foundation of China under Grant no. 61773301 . It is also supported by the Shaanxi Innovation Team Project under Grant No.2018TD-012.
Publisher Copyright:
© 2022 Elsevier Ltd
PY - 2023/3/1
Y1 - 2023/3/1
N2 - Most existing cross-modality person Re-IDentification works rely on discriminative modality-shared features for reducing cross-modality variations and intra-modality variations. Despite their preliminary success, such modality-shared appearance features cannot capture enough modality-invariant discriminative information due to a massive discrepancy between RGB and IR images. To address this issue, on top of appearance features, we further capture the modality-invariant relations among different person parts (referred to as modality-invariant relation features), which help to identify persons with similar appearances but different body shapes. To this end, a Multi-level Two-streamed Modality-shared Feature Extraction (MTMFE) sub-network is designed, where the modality-shared appearance features and modality-invariant relation features are first extracted in a shared 2D feature space and a shared 3D feature space, respectively. The two features are then fused into the final modality-shared features such that both cross-modality variations and intra-modality variations can be reduced. Besides, a novel cross-modality center alignment loss is proposed to further reduce the cross-modality variations. Experimental results on several benchmark datasets demonstrate that our proposed method exceeds state-of-the-art algorithms by a wide margin.
AB - Most existing cross-modality person Re-IDentification works rely on discriminative modality-shared features for reducing cross-modality variations and intra-modality variations. Despite their preliminary success, such modality-shared appearance features cannot capture enough modality-invariant discriminative information due to a massive discrepancy between RGB and IR images. To address this issue, on top of appearance features, we further capture the modality-invariant relations among different person parts (referred to as modality-invariant relation features), which help to identify persons with similar appearances but different body shapes. To this end, a Multi-level Two-streamed Modality-shared Feature Extraction (MTMFE) sub-network is designed, where the modality-shared appearance features and modality-invariant relation features are first extracted in a shared 2D feature space and a shared 3D feature space, respectively. The two features are then fused into the final modality-shared features such that both cross-modality variations and intra-modality variations can be reduced. Besides, a novel cross-modality center alignment loss is proposed to further reduce the cross-modality variations. Experimental results on several benchmark datasets demonstrate that our proposed method exceeds state-of-the-art algorithms by a wide margin.
KW - Cross-modality person Re-IDentification
KW - Modality-invariant relation features
KW - Modality-shared appearance features
KW - Thermal infrared images
KW - Visible images
U2 - 10.1016/j.patcog.2022.109145
DO - 10.1016/j.patcog.2022.109145
M3 - Article
AN - SCOPUS:85141449349
SN - 0031-3203
VL - 135
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 109145
ER -