应用科学学报2024,Vol.42Issue(5) :884-892.DOI:10.3969/j.issn.0255-8297.2024.05.014

跨通道交互注意力机制驱动的双流网络跨模态行人重识别

Cross-Modal Person Re-identification Driven by Cross-Channel Interactive Attention Mechanism in Dual-Stream Networks

何磊 栗风永 秦川
应用科学学报2024,Vol.42Issue(5) :884-892.DOI:10.3969/j.issn.0255-8297.2024.05.014

跨通道交互注意力机制驱动的双流网络跨模态行人重识别

Cross-Modal Person Re-identification Driven by Cross-Channel Interactive Attention Mechanism in Dual-Stream Networks

何磊 1栗风永 1秦川2
扫码查看

作者信息

  • 1. 上海电力大学计算机科学与技术学院,上海 201306
  • 2. 上海理工大学光电信息与计算机工程学院,上海 200093
  • 折叠

摘要

现有的跨模态行人重识别方法不能同时兼顾模态间与模态内的目标行人差异,很难提升检索准确度.为解决这一问题,引入跨通道交互的注意力机制,增强行人特征的鲁棒提取能力,有效抑制冗余特征的提取并获得更具辨别力的特征表达.进一步,联合异质中心三元组损失、三元组损失和身份损失进行监督学习,有效结合了行人特征的跨模态类间差异和类内差异.实验证明了所提方法的有效性.与7个已有的经典方法相比,所提方法在两个标准数据集RegDB与SYSU-MM01上都取得了较好的性能效果.

Abstract

Existing cross-modal person re-identification methods often fail to take into account the difference of target person between modes and within modes,making it difficult to further improve the retrieval accuracy.To solve this problem,this paper introduces the cross-channel interaction attention mechanism to enhance the robust extraction of person features,effectively suppresses the extraction of irrelevant features and achieves more discriminative feature expression.Furthermore,hetero-center triplet loss,triplet loss and identity loss are combined for supervised learning,effectively integrating the inter-modal and intra-class differences in person features.Experimental results demonstrate the effectiveness of the proposed method,which outperforms seven existing methods on two standard datasets,RegDB and SYSU-MM01.

关键词

跨模态/行人重识别/卷积神经网络/注意力机制

Key words

cross-modal/person re-identification/convolutional neural network/attention mechanism

引用本文复制引用

基金项目

国家自然科学基金(U1936213)

上海市自然科学基金(20ZR1421600)

出版年

2024
应用科学学报
上海大学 中国科学院上海技术物理研究所

应用科学学报

CSTPCDCSCD北大核心
影响因子:0.594
ISSN:0255-8297
参考文献量16
段落导航相关论文