2.765

2022影响因子

(CJCR)

  • 中文核心
  • EI
  • 中国科技核心
  • Scopus
  • CSCD
  • 英国科学文摘

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于最大互信息区域跟踪的人体行为检测算法

王泰青 王生进 丁晓青

王泰青, 王生进, 丁晓青. 基于最大互信息区域跟踪的人体行为检测算法. 自动化学报, 2012, 38(12): 2023-2031. doi: 10.3724/SP.J.1004.2012.02023
引用本文: 王泰青, 王生进, 丁晓青. 基于最大互信息区域跟踪的人体行为检测算法. 自动化学报, 2012, 38(12): 2023-2031. doi: 10.3724/SP.J.1004.2012.02023
WANG Tai-Qing, WANG Sheng-Jin, DING Xiao-Qing. Human Action Detection Based on Tracking Region of Maximum Mutual Information. ACTA AUTOMATICA SINICA, 2012, 38(12): 2023-2031. doi: 10.3724/SP.J.1004.2012.02023
Citation: WANG Tai-Qing, WANG Sheng-Jin, DING Xiao-Qing. Human Action Detection Based on Tracking Region of Maximum Mutual Information. ACTA AUTOMATICA SINICA, 2012, 38(12): 2023-2031. doi: 10.3724/SP.J.1004.2012.02023

基于最大互信息区域跟踪的人体行为检测算法

doi: 10.3724/SP.J.1004.2012.02023
详细信息
    通讯作者:

    王生进

Human Action Detection Based on Tracking Region of Maximum Mutual Information

  • 摘要: 人体行为检测问题不仅需要判断行为的类别,而且需要估计行为发生的时间和位置,有重要的现实应用意义. 人体行为检测的主要难点在于参数空间维度高以及背景运动干扰. 针对上述难点,本文提出了一种基于最大互信息区域跟踪的人体行为检测算法. 该算法将行为区域定义为最大互信息矩形区域,采用稠密轨迹作为底层特征,利用随机森林学习轨迹特征与行为类别的互信息函数,利用轨迹的时间连续性对行为区域进行大时间跨度的预测和跟踪. 实验结果表明,该算法不仅能够有效地识别不同类别的行为,而且能够适应现实场景中背景运动的干扰,从而准确地检测和跟踪行为区域.
  • [1] Turaga P, Chellappa R, Subrahmanian V S, Udrea O. Machine recognition of human activities: a survey. IEEE Transactions on Circuits and Systems for Video Technology, 2008, 18(11): 1473-1488[2] Gu J X, Ding X Q, Wang S J, Wu Y S. Action and gait recognition from recovered 3-D human joints. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 2010, 40(4): 1021-1033[3] Gu Jun-Xia, Ding Xiao-Qing, Wang Sheng-Jin. Human 3D model-based 2D action recognition. Acta Automatica Sinica, 2010, 36(1): 46-53(谷军霞, 丁晓青, 王生进. 基于人体行为3D模型的2D行为识别. 自动化学报, 2010, 36(1): 46-53)[4] Du You-Tian, Chen Feng, Xu Wen-Li. Approach to human activity multi-scale analysis and recognition based on multi-layer dynamic Bayesian network. Acta Automatica Sinica, 2009, 35(3): 225-232(杜友田, 陈峰, 徐文立. 基于多层动态贝叶斯网络的人的行为多尺度分析及识别方法. 自动化学报, 2009, 35(3): 225-232)[5] Yu G, Goussies N A, Yuan J S, Liu Z C. Fast action detection via discriminative random forest voting and top-K subvolume search. IEEE Transactions on Multimedia, 2011, 13(3): 507-517[6] Ryoo M S, Aggarwal J K. Spatio-temporal relationship match: video structure comparison for recognition of complex human activities. In: Proceedings of the 12th International Conference on Computer Vision. Kyoto, Japan, Brazil: IEEE, 2009. 1593-1600[7] Yuan J S, Liu Z C, Wu Y. Discriminative video pattern search for efficient action detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 33(9): 1728-1743[8] Gall J, Yao A, Razavi N, Van Gool L, Lempitsky V. Hough forests for object detection, tracking, and action recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2011, 33(11): 2188-2202[9] Yao A, Gall J, Van Gool L. A Hough transform-based voting framework for action recognition. In: Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition. San Francisco, CA, USA: IEEE, 2010. 2061-2068[10] Rodriguez M D, Ahmed J, Shah M. Action MACH: a spatio-temporal maximum average correlation height filter for action recognition. In: Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition. Anchorage, Alaska, USA: IEEE, 2008. 1-8[11] Derpanis K G, Sizintsev M, Cannons K, Wildes R P. Efficient action spotting based on a spacetime oriented structure representation. In: Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition. San Francisco, CA, USA: IEEE, 2010. 1990-1997[12] Oikonomopoulos A, Patras I, Pantic M. Spatiotemporal localization and categorization of human actions in unsegmented image sequences. IEEE Transactions on Image Processing, 2011, 20(4): 1126-1140[13] Cao L L, Liu Z C, Huang T S. Cross-dataset action detection. In: Proceedings of the 2010 IEEE Conference on Computer Vision and Pattern Recognition. San Francisco, CA, USA: IEEE, 2010. 1998-2005[14] Laptev I, Cedex R. On space-time interest points. International Journal of Computer Vision, 2005, 64(2-3): 107-123[15] Wang H, Ullah M M, Klser A, Laptev I, Schmid C. Evaluation of local spatio-temporal features for action recognition. In: Proceedings of the 2009 British Machine Vision Conference. London, UK: Springer-Verlag, 2009. 1-11[16] Wang H, Klaser A, Schmid C, Liu C L. Action recognition by dense trajectories. In: Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition. Colorado Springs, CO, USA: IEEE, 2011. 3169-3176[17] Lampert C H, Blaschko M B, Hofmann T. Beyond sliding windows: object localization by efficient subwindow search. In: Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition. Anchorage, Alaska, USA: IEEE, 2008. 1-8[18] Schuldt C, Laptev I, Caputo B. Recognizing human actions: a local SVM approach. In: Proceedings of the 17th International Conference on Pattern Recognition. Cambridge, England, UK: IEEE, 2004. 32-36[19] Chen C C, Aggarwal J K. Modeling human activities as speech. In: Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition. Providence, RI: IEEE, 2011. 3425-3432[20] Yan K, Sukthankar R, Hebert M. Event detection in crowded videos. In: Proceedings of the 11th International Conference on Computer Vision. Rio de Janeiro, Brazil: IEEE, 2007. 1-8
  • 加载中
计量
  • 文章访问数:  1958
  • HTML全文浏览量:  42
  • PDF下载量:  985
  • 被引次数: 0
出版历程
  • 收稿日期:  2012-01-04
  • 修回日期:  2012-07-15
  • 刊出日期:  2012-12-20

目录

    /

    返回文章
    返回