留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于步态的摄像机网络跨视域行人跟踪

宋淑婕 万九卿

宋淑婕,万九卿. 基于步态的摄像机网络跨视域行人跟踪[J]. 北京航空航天大学学报,2023,49(8):2154-2166 doi: 10.13700/j.bh.1001-5965.2021.0610
引用本文: 宋淑婕,万九卿. 基于步态的摄像机网络跨视域行人跟踪[J]. 北京航空航天大学学报,2023,49(8):2154-2166 doi: 10.13700/j.bh.1001-5965.2021.0610
SONG S J,WAN J Q. Gait based cross-view pedestrian tracking with camera network[J]. Journal of Beijing University of Aeronautics and Astronautics,2023,49(8):2154-2166 (in Chinese) doi: 10.13700/j.bh.1001-5965.2021.0610
Citation: SONG S J,WAN J Q. Gait based cross-view pedestrian tracking with camera network[J]. Journal of Beijing University of Aeronautics and Astronautics,2023,49(8):2154-2166 (in Chinese) doi: 10.13700/j.bh.1001-5965.2021.0610

基于步态的摄像机网络跨视域行人跟踪

doi: 10.13700/j.bh.1001-5965.2021.0610
基金项目: 北京市自然科学基金(4192031); 国家自然科学基金(61873015)
详细信息
    作者简介:

    宋淑婕 女,硕士研究生。主要研究方向:目标检测跟踪与识别

    万九卿 男,博士,副教授,硕士生导师。主要研究方向:信号处理、目标检测跟踪与识别

    通讯作者:

    E-mail:wanjiuqing@buaa.edu.cn

  • 中图分类号: TP391

Gait based cross-view pedestrian tracking with camera network

Funds: Beijing Municipal Natural Science Foundation (4192031); National Natural Science Foundation of China (61873015)
More Information
  • 摘要:

    非重叠视域摄像机网络行人目标跨视域跟踪是智能视觉监控的基本问题之一。针对基于外观一致性假设的行人跨视域跟踪方法对光照或衣着变化敏感的问题,提出一种融合基于2D骨架图的步态特征与时空约束的跨视域行人跟踪方法。从单视域局部轨迹提取骨架集合计算步态特征,建立跨视域目标跟踪问题的整数线性规划模型,模型参数由步态特征相似度和时空约束定义,利用对偶分解算法实现问题的分布式求解。通过步态特征与更加精细化的时空约束融合,显著提升了仅基于步态特征的跨视域跟踪算法对于光照和衣着变化的鲁棒性,克服了单独使用步态或时空特征时判别力较弱的问题。在公开数据集上的测试结果表明,所提方法跟踪准确,且对光照和衣着变化具有鲁棒性。

     

  • 图 1  智能摄像机网络及其拓扑

    Figure 1.  A smart camera network and its topology

    图 2  目标通过时摄像机收集到的视频帧

    Figure 2.  Video frames collected by camera as target passes

    图 3  说明示例

    Figure 3.  Illustrating example

    图 4  修正的骨架模型

    Figure 4.  Modified skeleton model

    图 5  生成骨架图像

    Figure 5.  Skeleton image generated

    图 6  一个步态周期的骨架图像集合

    Figure 6.  Set of skeleton images in one walking cycle

    图 7  对偶分解算法中的二分图子问题

    Figure 7.  Bipartite graph subproblem in dual decomposition algorithm

    图 8  摄像机网络的布局和视域

    Figure 8.  Layouts of camera networks and camera’s FOV

    图 9  光照变化效果

    Figure 9.  Effects of lighting variation

    图 10  换装效果

    Figure 10.  Effect of clothed changing

    图 11  VBOLO数据集拍摄场景

    Figure 11.  Two stations of VBOLO dataset

    图 12  演员在2个场景出现的示例[33]

    Figure 12.  Examples of actors appearing in two stations

    图 13  9个演员Rank-1~Rank-10查询结果的对应观测

    Figure 13.  Corresponding observations of 9 actors from Rank-1 to Rank-10

    图 14  光照改变下跨视域跟踪示例

    Figure 14.  Examples of cross-view tracking under lighting variations

    图 15  换装情况下跨视域跟踪示例

    Figure 15.  Examples of cross-view tracking under clothing changes

    表  1  时空观测

    Table  1.   Space-time observation

    时空属性状态
    位置CamE
    进入的时间09:10:21 a.m.
    离开的时间09:10:27 a.m.
    进入的方向左边界
    离开的方向右边界
    下载: 导出CSV

    表  2  NLPR_MCT数据集的细节

    Table  2.   Details of NLPR_MCT dataset

    子数据集相机数持续时间/min帧率/(帧·s−1目标数$T{P_s}$$ T{P_c} $
    Dataset13202023571853334
    Dataset23202025588419408
    Dataset34 3.525 1418187152
    Dataset442425 4942615256
      注:TPs为数据集提供的单视域的轨迹数,TPc为数据集提供的跨视域的轨迹数。
    下载: 导出CSV

    表  3  MCT及模拟数据集上Rank-n准确度比较

    Table  3.   Comparison of Rank-n accuracy on MCT and simulated datasets %

    子数据集Rank-1Rank-5
    原始光照换装原始光照换装
    RGBSSP步态RGBSSP步态RGBSSP步态RGBSSP步态RGBSSP步态RGBSSP步态
    Dataset16.3545.517.86 2.59 2.6 26.42 0.53 3.2 17.53 15.3454.5 40.82 4.154.241.977.417.938.14
    Dataset28.89 50.411.960.731.18.99 0.37 1.1 11.15 16.67 64.1 26.452.193.017.272.965.223.38
    Dataset319.0869.135.1 16.1160.535.1 9.21 13.8 29.8 53.29 85.5 65.5647.6578.362.2535.5355.365.56
    Dataset421.1485.540.7314.1155.036.95 5.62 8.4 36.95 41.87 95.6 66.5337.974.364.6624.134.961.04
    下载: 导出CSV

    表  4  VBOLO数据集上步态特征Rank-n准确度比较

    Table  4.   Comparison of Rank-n accuracyof gait feature on VBOLO dataset %

    Rank-1Rank-5 Rank-10 Rank-15 Rank-20
    56.5791.9297.9898.99100
    下载: 导出CSV

    表  5  跨视域跟踪方法的性能比较

    Table  5.   Performance comparison of cross-view tracking methods

    方法${ {{e} } }^{ {\rm{c} } }$W
    Dataset1Dataset2Dataset3Dataset4 Dataset1Dataset2Dataset3Dataset4
    ICLM[37]13 30 32 62 0.9610.9270.7900.758
    CRF[37]54 81 51 700.8380.8010.6650.727
    EGM[29]55 121 39 1570.83530.70340.74170.3845
    PMCSHR[38]112 167 44 1100.6620.5910.7110.633
    Hfutdspmct[30]86 141 40 1550.74250.65440.73680.3945
    AdbTeam[30]227 267 131 2160.32040.34560.13820.1563
    TRACTA[39]121 176 126 1810.63270.54480.13980.2870
    本文24 81 85 950.92790.80140.43700.6286
    下载: 导出CSV

    表  6  光照改变和换装情况下基于步态、RGB和SSP特征的跨视域跟踪方法结果对比

    Table  6.   Comparison of results of cross-view tracking methods based on gait, RGB and SSP feature under lighting variations and clothes changing

    子数据集$e^{ {\rm{c} } }$W
    原始光照改变换装原始光照改变换装
    RGBSSP步态RGBSSP步态RGBSSP步态RGBSSP步态RGBSSP步态RGBSSP步态
    Dataset1751830834326 54 59 300.775 40.945 80.908 90.751 40.870 80.921 80.838 20.823 30.909 5
    Dataset214039561320.904 45711473620.656 81140.860 90.674 00.818 80.859 50.720 50.719 90.847 2
    Dataset312036558665559495600.210 50.699 90.517 50.358 20.488 20.533 90.381 50.374 90.508 2
    Dataset4164367815181100115114820.359 20.859 00.695 00.403 00.678 30.601 40.550 60.554 50.675 6
    下载: 导出CSV

    表  7  有无时空约束的结果比较

    Table  7.   Comparison of results with and without space-time constraints

    有/无时控约束$e^{ {\rm{c} } }$W
    Dataset1Dataset2Dataset3Dataset4Dataset1Dataset2Dataset3Dataset4
    无时空约束276342 120 2030.159 90.148 20.195 70.200 7
    有时空约束 30 56 55780.908 90.860 90.517 50.695 0
    下载: 导出CSV

    表  8  基于剪影和基于骨架的步态特征对比

    Table  8.   Silhouette vs skeleton based gait feature

    输入$e^{ {\rm{c} } }$W
    Dataset1Dataset2Dataset3Dataset4Dataset1Dataset2Dataset3Dataset4
    剪影3788931100.891 50.782 40.365 70.551 2
    骨架305655 780.908 90.860 90.517 50.695 0
    下载: 导出CSV
  • [1] GRAY D, TAO H. Viewpoint invariant pedestrian recognition with an ensemble of localized features[C]//European Conference on Computer Vision. Berlin: Springer, 2008: 262-275.
    [2] VARMA M, ZISSERMAN A. A statistical approach to texture classification from single images[J]. International Journal of Computer Vision, 2005, 62(1-2): 61-81. doi: 10.1007/s11263-005-4635-4
    [3] AHONEN T, HADID A, PIETIKÄINEN M. Face description with local binary patterns: Application to face recognition[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2006, 28(12): 2037-2041. doi: 10.1109/TPAMI.2006.244
    [4] LOWE D G. Distinctive image features from scale-invariant keypoints[J]. International Journal of Computer Vision, 2004, 60(2): 91-110. doi: 10.1023/B:VISI.0000029664.99615.94
    [5] BAY H, TUYTELAARS T, VAN GOOL L. SURF: Speeded up robust features[C]//European Conference on Computer Vision. Berlin: Springer, 2006: 404-417.
    [6] CALONDER M, LEPETIT V, STRECHA C, et al. BRIEF: Binary robust independent elementary features[C]//European Conference on Computer Vision. Berlin: Springer, 2010: 778-792.
    [7] ABDEL-HAKIM A E, FARAG A A. CSIFT: A SIFT descriptor with color invariant characteristics[C]//2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2006: 1978-1983.
    [8] BAK S, BRÉMOND F. Re-identification by covariance descriptors[M]. Berlin: Springer, 2014: 71-91.
    [9] WAN F B, WU Y, QIAN X L, et al. When person re-identification meets changing clothes[C]//2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. Piscataway: IEEE Press, 2020: 3620-3628.
    [10] 王科俊, 丁欣楠, 邢向磊, 等. 多视角步态识别综述[J]. 自动化学报, 2019, 45(5): 841-852. doi: 10.16383/j.aas.2018.c170559

    WANG K J, DING X N, XING X L, et al. A survey of multi-view gait recognition[J]. Acta Automatica Sinica, 2019, 45(5): 841-852(in Chinese). doi: 10.16383/j.aas.2018.c170559
    [11] HAN J, BHANU B. Individual recognition using gait energy image[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2006, 28(2): 316-322. doi: 10.1109/TPAMI.2006.38
    [12] GIANARIA E, BALOSSINO N, GRANGETTO M, et al. Gait characterization using dynamic skeleton acquisition[C]//2013 IEEE 15th International Workshop on Multimedia Signal Processing. Piscataway: IEEE Press, 2013: 440-445.
    [13] GÜLER R A, NEVEROVA N, KOKKINOS I. DensePose: Dense human pose estimation in the wild[C]//2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2018: 7297-7306.
    [14] VERLEKAR T. Gait analysis in unconstrained environments[D]. Lisbon: University of Lisbon, 2019.
    [15] JAN NORDIN M D, SAADOON A. A survey of gait recognition based on skeleton model for human identification[J]. Research Journal of Applied Sciences, Engineering and Technology, 2016, 12(7): 756-763. doi: 10.19026/rjaset.12.2751
    [16] ZHANG D, SHAH M. Human pose estimation in videos[C]//2015 IEEE International Conference on Computer Vision. Piscataway: IEEE Press, 2016: 2012-2020.
    [17] LIN B, ZHANG S, BAO F. Gait recognition with multiple-temporal-scale 3D convolutional neural network[C]//Proceedings of the 28th ACM International Conference on Multimedia. New York: ACM, 2020: 3054-3062.
    [18] LI N, ZHAO X, MA C. JointsGait: A model-based gait recognition method based on gait graph convolutional networks and joints relationship pyramid mapping[EB/L]. (2020-12-09) [2021-10-01]. https://arxiv.org/abs/2005.08625v1.
    [19] CHAO H B, HE Y W, ZHANG J P, et al. GaitSet: Regarding gait as a set for cross-view gait recognition[J]. Proceedings of the Conference on Artificial Intelligence, 2019, 33(1): 8126-8133. doi: 10.1609/aaai.v33i01.33018126
    [20] CHEN W H, CAO L J, CHEN X T, et al. An equalized global graph model-based approach for multicamera object tracking[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2017, 27(11): 2367-2381. doi: 10.1109/TCSVT.2016.2589619
    [21] CHEN X J, BHANU B. Integrating social grouping for multitarget tracking across cameras in a CRF model[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2017, 27(11): 2382-2394. doi: 10.1109/TCSVT.2016.2565978
    [22] YU S Q, TAN D L, TAN T N. A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition[C]//18th International Conference on Pattern Recognition. Piscataway: IEEE Press, 2006: 441-444.
    [23] WANG L, TAN T N, NING H Z, et al. Silhouette analysis-based gait recognition for human identification[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003, 25(12): 1505-1518. doi: 10.1109/TPAMI.2003.1251144
    [24] CAO Z, HIDALGO G, SIMON T, et al. OpenPose: Realtime multi-person 2D pose estimation using part affinity fields[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, 43(1): 172-186. doi: 10.1109/TPAMI.2019.2929257
    [25] ZAHEER M, KOTTUR S, RAVANBAKHSH S, et al. Deep sets[EB/OL]. (2018-04-14) [2021-10-01].https://arxiv.org/abs/1703.06114v1,
    [26] FU Y, WEI Y C, ZHOU Y Q, et al. Horizontal pyramid matching for person re-identification[C]//Proceedings of the AAAI Conference on Artificial Intelligence. Washington, D. C.:AAAI Press, 2019: 8295-8302.
    [27] GOLDBERG A V. An efficient implementation of a scaling minimum-cost flow algorithm[J]. Journal of Algorithms, 1997, 22(1): 1-29. doi: 10.1006/jagm.1995.0805
    [28] KOMODAKIS N, PARAGIOS N. Beyond pairwise energies: Efficient optimization for higher-order MRFs[C]//2009 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2009: 2985-2992.
    [29] WAN J Q, CHEN X, BU S C, et al. Distributed data association in smart camera network via dual decomposition[J]. Information Fusion, 2018, 39: 120-138. doi: 10.1016/j.inffus.2017.04.007
    [30] Multi-camera object tracking (MCT) challenge[DB/OL]. (2017)[2017-10-18]. http://www.mct2014.com.
    [31] LIANG X D, GONG K, SHEN X H, et al. Look into person: Joint body parsing & pose estimation network and a new benchmark[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019, 41(4): 871-885.
    [32] LI P, BROGAN J, FLYNN P J. Toward facial re-identification: Experiments with data from an operational surveillance camera plant[C]//2016 IEEE 8th International Conference on Biometrics Theory, Applications and Systems. Piscataway: IEEE Press, 2016: 1-8.
    [33] LI P, PRIETO M L, FLYNN P J, et al. Learning face similarity for re-identification from real surveillance video: A deep metric solution[C]//2017 IEEE International Joint Conference on Biometrics. Piscataway: IEEE Press, 2018: 243-252.
    [34] ZHENG J K, LIU X C, YAN C G, et al. TraND: Transferable neighborhood discovery for unsupervised cross-domain gait recognition[C]//2021 IEEE International Symposium on Circuits and Systems. Piscataway: IEEE Press, 2021: 1-5.
    [35] QUISPE R, PEDRINI H. Improved person re-identification based on saliency and semantic parsing with deep neural network models[J]. Image and Vision Computing, 2019, 92: 103809. doi: 10.1016/j.imavis.2019.07.009
    [36] LI X, ZHAO L M, WEI L N, et al. DeepSaliency: Multi-task deep neural network model for salient object detection[J]. IEEE Transactions on Image Processing, 2016, 25(8): 3919-3930. doi: 10.1109/TIP.2016.2579306
    [37] LEE Y G, TANG Z, HWANG J N. Online-learning-based human tracking across non-overlapping cameras[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2018, 28(10): 2870-2883. doi: 10.1109/TCSVT.2017.2707399
    [38] CHEN W H, CAO L J, CHEN X T, et al. A novel solution for multi-camera object tracking[C]//2014 IEEE International Conference on Image Processing. Piscataway: IEEE Press, 2015: 2329-2333.
    [39] HE Y H, WEI X, HONG X P, et al. Multi-target multi-camera tracking by tracklet-to-target assignment[J]. IEEE Transactions on Image Processing, 2020, 29: 5191-5205. doi: 10.1109/TIP.2020.2980070
    [40] TEEPE T, KHAN A, GILG J, et al. GaitGraph: Graph convolutional network for skeleton-based gait recognition[EB/OL]. (2021-06-09) [2021-10-01].https://arxiv.org/abs/2101.11228.
    [41] BARNICH O, VAN DROOGENBROECK M. ViBe: A universal background subtraction algorithm for video sequences[J]. IEEE Transactions on Image Processing, 2011, 20(6): 1709-1724. doi: 10.1109/TIP.2010.2101613
  • 加载中
图(15) / 表(8)
计量
  • 文章访问数:  164
  • HTML全文浏览量:  49
  • PDF下载量:  11
  • 被引次数: 0
出版历程
  • 收稿日期:  2021-10-18
  • 录用日期:  2021-12-10
  • 网络出版日期:  2022-01-26
  • 整期出版日期:  2023-08-31

目录

    /

    返回文章
    返回
    常见问答