Channel exchange and adversarial learning guided cross-modal person re-identification

作者:

Highlights:

摘要

The extensive progress of Re-ID has been obtained in the visible modality. Because the security monitoring system automatically switches from visible modality to infrared modality in the dark situation, the research of the infrared-visible modality-based cross-modal person Re-ID (IV-Re-ID) task increases much attention. However, the existing heterogeneous physical properties result in a substantial semantic gap between the visible and infrared modality data, further increasing the challenge of IV-Re-ID. This paper proposes a Cross-modal Channel Exchange Network (CmCEN) for the task of IV-Re-ID. First, a non-local attention mechanism and a semi-weight share mechanism are utilized in the backbone network to enhance the discriminative capability of both local and global data representations. Then, a channel exchange module is designed to measure each channel’s significance effectively, and the useless channels could be replaced by the critical channels from the other modality data. Finally, a discriminator with adversarial loss is designed in the generative adversarial module. It can generate a similar distribution of two different modality images with the same person identity and a different distribution of two different modality images with different identities under the supervision of the adversarial loss to further improve the robustness of the learned latent feature space. The evaluation results on two cross-modal datasets demonstrated that the CmCEN achieves competitively and even higher performance comparing the SOTA models in terms of accuracy and mAP of IV-Re-ID.

论文关键词:Person Re-ID,Channel exchange,Cross-modality,Adversarial learning

论文评审过程:Received 5 August 2022, Revised 31 August 2022, Accepted 7 September 2022, Available online 22 September 2022, Version of Record 5 October 2022.

论文官网地址:https://doi.org/10.1016/j.knosys.2022.109883