基于生成对抗网络的跨视角步态特征提取
作者:
作者单位:

作者简介:

通讯作者:

中图分类号:

基金项目:


Cross-View Gait Feature Extraction Using Generative Adversarial Networks
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    步态是一种能够在远距离、非侵犯的条件下识别身份的生物特征,但在实际场景中,步态很容易受到拍摄视角、行走环境、物体遮挡、着装等因素的影响.在跨视角识别问题上,现有方法只注重将多种视角的步态模板转化到固定视角下,且视角跨度的增大加深了错误的累积.为了提取有效的步态特征用于跨视角步态识别,本文提出了一种基于生成对抗网络的跨视角步态特征提取方法,该方法只需训练一个模型即可将步态模板转换到任意视角下的正常行走状态,并最大化地保留原本的身份特征信息,从而提高步态识别的准确率.在CASIA-B和OUMVLP数据集上的实验结果表明,该方法在解决跨视角步态识别问题上具有一定的鲁棒性和可行性.

    Abstract:

    Gait is a biological feature that can recognize identity at a long distance and without invasion. However, the performance of gait recognition can be adversely affected by many factors such as view angle, walking environment, occlusion, and clothing, among others. For cross-view gait recognition, the existing cross-view methods focus on transforming gait templates to a specific view angle, which may accumulate the transformation error in a large variation of view angles. To extract invariant gait features, we propose a method which is based on generative adversarial networks. In the proposed method, a gait template could be transformed to any view angle and normal walking state by training only one model. At the same time, the method maintain effective identity information to the most extent and improving the accuracy of gait recognition. Experiments on CASIA-B and OUMVLP datasets indicate that compared with several published approaches, the proposed method achieves competitive performance and is more robust and interpretable to cross-view gait recognition.

    参考文献
    相似文献
    引证文献
引用本文

秦月红,王敏.基于生成对抗网络的跨视角步态特征提取.计算机系统应用,2020,29(1):164-170

复制
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2019-06-25
  • 最后修改日期:2019-07-16
  • 录用日期:
  • 在线发布日期: 2019-12-30
  • 出版日期: 2020-01-15
文章二维码
您是第位访问者
版权所有:中国科学院软件研究所 京ICP备05046678号-3
地址:北京海淀区中关村南四街4号 中科院软件园区 7号楼305房间,邮政编码:100190
电话:010-62661041 传真: Email:csa (a) iscas.ac.cn
技术支持:北京勤云科技发展有限公司

京公网安备 11040202500063号