基于多核扩展卷积的无监督视频行人重识别
DOI:
作者:
作者单位:

1.兰州理工大学 电气工程与信息工程学院 甘肃 兰州;2.西北民族大学 数学与计算机科学学院 甘肃 兰州

作者简介:

通讯作者:

基金项目:

国家自然科学基金(基金号62061042);甘肃省工业过程先进控制重点实验室开放基金项目(2022KX10)


Unsupervised Video Person Re-identification Based On Multiple Kernel Dilated Convolution
Author:
Affiliation:

1.School of Electrical Engineering and Information Engineering,Lanzhou University of Technology;2.College of Mathematics and Computer Science,Northwest Minzu University

Fund Project:

Project(62061042) supported by the National Natural Science Foundation of China; Project(2022KX10) supported by the Key Laboratory of Gansu Advanced Control for Industrial Processes

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
    摘要:

    行人重识别旨在跨监控摄像头下检索出特定的行人目标。由于存在姿态变化、物体遮挡、背景干扰等问题,导致对行人重识别效果不佳,提出一种利用多核扩展卷积的无监督视频行人重识别方法,使提取到的行人特征得到更充分的利用。首先,将预训练的ResNet50作为编码器,在编码器中引入多核扩展卷积模块增加卷积核感受野,以捕获局部和全局特征,通过解码器来获取更多的语义信息,增强特征表示;其次,在解码器的输出引入多尺度特征融合模块融合相邻层中的特征,进一步减少不同特征通道层之间的语义差距,以产生更鲁棒的特征表示;最后,在PRID2011、iLIDS-VID和DukeMTMC-VideoReID等主流数据集进行实验验证,结果表明改进后的模型获取更高的性能指标和识别精度,能够更好的表达行人特征。

    Abstract:

    Person re-identification aims to retrieve specific person objects across surveillance cameras. Due to problems such as attitude changes, object occlusion, and background interference, the person re-identification effect is not good. An unsupervised video person re-identification method using multi-core expanded convolution is proposed to make full use of the extracted person features. First, the pre-trained ResNet50 is used as the encoder, and the multi-core expanded convolution module is introduced into the encoder to increase the convolution kernel receptive field to capture local and global features, and obtain more semantic information through the decoder to enhance feature representation; Secondly, a multi-scale feature fusion module is introduced at the output of the decoder to fuse the features in adjacent layers, further reducing the semantic gap between different feature channel layers to produce more robust feature representations; finally, in PRID2011, iLIDS-VID Experimental verification with mainstream datasets such as DukeMTMC-VideoReID shows that the improved model has higher performance indicators and recognition accuracy, and can better express person characteristics.

    参考文献
    相似文献
    引证文献
引用本文
分享
文章指标
  • 点击次数:
  • 下载次数:
历史
  • 收稿日期:2023-11-27
  • 最后修改日期:2024-02-26
  • 录用日期:2024-04-26
  • 在线发布日期: