论文:2016,Vol:34,Issue(5):886-892
引用本文:
丁松涛, 曲仕茹. 基于改进时空兴趣点检测的人体行为识别算法[J]. 西北工业大学学报
Ding Songtao, Qu Shiru. An Improved Interest Point Detector for Human Action Recognition[J]. Northwestern polytechnical university

基于改进时空兴趣点检测的人体行为识别算法
丁松涛, 曲仕茹
西北工业大学 自动化学院, 陕西 西安 710129
摘要:
提出了一种基于改进时空兴趣点检测的人体行为识别算法。旨在针对复杂环境的时空特性,在传统兴趣点检测算法的基础上,加入背景点抑制和时空兴趣点约束,以减少无用兴趣点对有效兴趣点信息的干扰。为此,首先对Harris-Laplace算法进行改进,以克服兴趣点检测过程中遇到的多尺度问题和冗余点过多问题,提取筛选后的有效兴趣点作为目标的运动坐标信息。然后基于Bag-of-words模型思想,使用HOG算子对兴趣点进行特征提取,建立视觉词典,使用AIB算法合并词义相近的视觉词汇,作为单词表中的基础词汇。最后使用SVM进行人体行为分类并实现复杂环境下的人体行为识别。为了验证新算法的有效性,分别在现有的公开人体行为基准数据库和一些复杂场景下进行实验。试验结果表明,通过对无用兴趣点的抑制,能够有效降低单帧图像的计算复杂度,减少特征提取时间,提高行为识别准确度。
关键词:    人体行为识别    时空抑制    HOG算子    AIB算法   
An Improved Interest Point Detector for Human Action Recognition
Ding Songtao, Qu Shiru
School of Automation, Northwestern Polytechnical University, Xi'an 710072, China
Abstract:
In this paper, we present a human action recognition algorithm based on interest points in spatial and temporal constraints. In order to overcome the problems of available information interference cause by complex background cenes, We proposed the improved Spatio-Temporal Interest Points (STIP) detection approach by surround suppression combined with local and temporal constraints. Firstly, the improved Harris-Laplace algorithm is proposed to solve the multi-scale problems. Then, based on the bag of words model, the HOG descriptor is used to extract feature vectors and Agglomerative Information Bottleneck(AIB) algorithm to combine the visual vocabulary. The Support Vector Machine (SVM) is trained for action classification and prediction. In order to validate the effectiveness of the proposed method, experiments were carried out under the existing disclosure benchmark datasets of human action and other more complex scenes. Experiment results demonstrate that the proposed human action recognition algorithm is both effective and efficient in a great variety of complex scenes.
Key words:    human action recognition    Spatio-Temporal suppression    HOG    AIB   
收稿日期: 2016-04-05     修回日期:
DOI:
基金项目: 教育部高等学校博士学科点专项科研基金(20096102110027)、航天科技创新基金(CASC201104)与航空科学基金(2012ZC53043)资助
通讯作者:     Email:
作者简介: 丁松涛(1984-),西北工业大学博士研究生,主要从事图像处理及深度学习的研究。
相关功能
PDF(3655KB) Free
打印本文
把本文推荐给朋友
作者相关文章
丁松涛  在本刊中的所有文章
曲仕茹  在本刊中的所有文章

参考文献:
[1] Robertson N, Reid I. Behaviour Understanding in Video:A Combined Method[C]//10th IEEE International Conference on computer Vision, 2005:808-815
[2] Parameswaran V, Chellappa R. View Invariance for Human Action Recognition[J]. International Journal of Computer Vision, 2006, 66(1):83-101
[3] Li X. HMM Based Action Recognition Using Oriented Histograms of Optical Flow Field[J]. Electronics Letters, 2007, 43(10):560-561
[4] Laptev I. On Space-Time Interest Points[J]. International Journal of Computer Vision, 2005, 64(2/3):432-439
[5] Dollar P, Rabaud V, Cottrell G, et al. Behavior Recognition via Sparse Spatio-Temporal Features[C]//2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance, 2005:65-72
[6] Park S, Aggarwal J K. A Hierarchical Bayesian Network for Event Recognition of Human Actions and Interactions[J]. Multimedia Systems, 2004, 10(2):164-179
[7] Bregonzio M, Gong S, Xiang T. Recognising Action as Clouds of Space-Time Interest Points[C]//IEEE Conference on Computer Vision & Pattern Recognition, 2009:1948-1955
[8] Yaron O, Sidi M. A Combined Corner and Edge Detector[C]//Proc of Fourth Alvey Vision Conference, 1988:147-151
[9] Lindeberg T. Feature Detection with Automatic Scale Selection[J]. International Journal of Computer Vision, 1998, 30(2):77-116
[10] Chakraborty B, Holte M B, Moeslund T B, et al. A Selective Spatio-Temporal Interest Point Detector for Human Action Recognition in Complex Scenes[C]//2011 International Conference on Computer Vision, 2011:1776-1783
[11] Slonim N, Tishby N. Agglomerative Information Bottleneck[C]//Advances in Neural Information Processing Systems, 1999:617-623
[12] Schuldt C, Laptev I, Caputo B. Recognizing Human Actions:A Local SVM Approach[C]//17th International Conference on Pattern Recognition, 2004:32-36