Visible-Infrared Cross-modality Person Re-identification Based on Whole-Individual Training
2021
Abstract Visible-infrared cross-modality person re-identification (VI-ReID) aims to search person images across cameras of different modalities, which can make up for the problem that ReID cannot be performed through visible images in a dark environment. The difficulty of VI-ReID task is the huge discrepancy between the visible modality and the infrared modality. In this paper, a novel whole-individual training (WIT) model is proposed for VI-ReID, which is based on the idea of pulling in the whole and distinguishing the individuals. Specifically, the model is divided into a whole part and an individual part. Two loss functions are developed in the whole part, namely center maximum mean discrepancy (CMMD) loss and intra-class heterogeneous center (ICHC) loss. Ignoring identity difference and treating each modality as a whole, the CMMD loss pulls in the centers of the two modalities. Ignoring modality difference and treating each identify as a whole, the ICHC loss pulls images with the same identity to its cross-modality center. In the individual part, a cross-modality triplet (CMT) loss is employed, which can distinguish the pedestrian images with different identities. The WIT model can help the network identify pedestrian images in an all-round way. Experiments show that the VI-ReID performance of the proposed method is better than existing technologies on two most popular benchmark datasets SYSU-MM01 and RegDB.
Keywords:
- Correction
- Source
- Cite
- Save
- Machine Reading By IdeaReader
30
References
1
Citations
NaN
KQI