Your browser doesn't support javascript.
loading
Cross-modal group-relation optimization for visible-infrared person re-identification.
Zhu, Jianqing; Wu, Hanxiao; Chen, Yutao; Xu, Heng; Fu, Yuqing; Zeng, Huanqiang; Liu, Liu; Lei, Zhen.
Affiliation
  • Zhu J; College of Engineering, Huaqiao University, Quanzhou, China.
  • Wu H; College of Information Science and Engineering, Huaqiao University, Xiamen, China; School of Computer Science and Artificial Intelligence, Wuhan University of Technology, Wuhan, China.
  • Chen Y; College of Engineering, Huaqiao University, Quanzhou, China.
  • Xu H; College of Engineering, Huaqiao University, Quanzhou, China.
  • Fu Y; College of Engineering, Huaqiao University, Quanzhou, China.
  • Zeng H; College of Engineering, Huaqiao University, Quanzhou, China. Electronic address: zeng0043@hqu.edu.cn.
  • Liu L; School of Artificial Intelligence and State Key Lab of Software Development Environment, Beihang University, Beijing, China. Electronic address: liuliubh@buaa.edu.cn.
  • Lei Z; State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China; School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China; Centre for Artificial Intelligence and Robotics, Hong Kong Institute
Neural Netw ; 179: 106576, 2024 Jul 31.
Article in En | MEDLINE | ID: mdl-39121790
ABSTRACT
Visible-infrared person re-identification (VIPR) plays an important role in intelligent transportation systems. Modal discrepancies between visible and infrared images seriously confuse person appearance discrimination, e.g., the similarity of the same class of different modalities is lower than the similarity between different classes of the same modality. Worse still, the modal discrepancies and appearance discrepancies are coupled with each other. The prevailing practice is to disentangle modal and appearance discrepancies, but it usually requires complex decoupling networks. In this paper, rather than disentanglement, we propose to measure and optimize modal discrepancies. We explore a cross-modal group-relation (CMGR) to describe the relationship between the same group of people in two different modalities. The CMGR has great potential in modal invariance because it considers more stable groups rather than individuals, so it is a good measurement for modal discrepancies. Furthermore, we design a group-relation correlation (GRC) loss function based on Pearson correlations to optimize CMGR, which can be easily integrated with the learning of VIPR's appearance features. Consequently, our CMGR model serves as a pivotal constraint to minimize modal discrepancies, operating in a manner similar to a loss function. It is applied solely during the training phase, thereby obviating the need for any execution during the inference phase. Experimental results on two public datasets (i.e., RegDB and SYSU-MM01) demonstrate that our CMGR method is superior to state-of-the-art approaches. In particular, on the RegDB dataset, with the help of CMGR, the rank-1 identification rate has improved by more than 7% compared to the case of not using CMGR.
Key words

Full text: 1 Collection: 01-internacional Database: MEDLINE Language: En Journal: Neural Netw Year: 2024 Document type: Article

Full text: 1 Collection: 01-internacional Database: MEDLINE Language: En Journal: Neural Netw Year: 2024 Document type: Article