基于视频的运动人体异常行为分析识别研究
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
智能视频监控是利用计算机技术实现自动视频分析的技术。作为一种有效安防的手段,智能视频监控系统正越来越受到人们的青睐。对于视频序列中行人异常行为的分析识别是智能视频监控领域中日益受到重视的一个研究方向。基于异常行为分析的智能视频监控系统不仅能忽略大量监控系统中对安防无用的信息,从而高效地完成保障安全的任务,还能节省大量的人力物力,给社会带来很大的经济收益。并且能够实现即时报警,解决了传统监控系统的事后性。本文从理论和实际应用的角度,对以视频为输入的异常行为识别进行了一些新的探索,以下所提出的算法均在实践中进行了应用。本文的主要贡献如下:
     1)提出了基于模板匹配的最小标准方差行为分类算法,解决了监控中不同人的类似动作但不完全相同的情况,有效解决了非固定速度的行为分类问题,能比较好地执行行为的分类任务。围绕着人的动作是时间序列上姿态的集合这样的事实,利用人的动作行为的周期特性,先将图像序列转换为一组静态形状模式,然后在识别过程中和预先存储的行为标本相比较。在建立标准库的过程中,提出了在最不相似的标准模板的生成算法,利用Procrustes中值形状距离作为图象的相似度标准,找出在一个视频测试周期内与标准库内动作标准方差最小的行为,实验证明最小标准方差算法解决了非连续性识别的要求,克服了模板匹配对运动时间间隔变化敏感的问题;提出了基于模板匹配的平均值与最小方差加权的行为分类算法,利用Zernike速度矩距离作为图象的相似度标准,找出在一个视频测试周期内与标准库内标准动作平均值与最小方差加权最小值,从根本上解决了在同一段时间内不同动作变速动作的分类,解决了某个图象集合中标准方差最小,但可能两个图象集最不相似的问题,增加了动作分类的准确率。
     2)提出了基于模糊联想记忆网络的行为分类方法,解决了动作姿态之间联系的问题,使得计算机识别过程向人的思维靠拢。把每个静态姿势作为一个状态,这些状态之间通过某种概率联系起来。任何运动序列可以看作为这些静态姿势的不同状态之间的一次遍历过程,在这些遍历期间计算联合概率,其最大值被选择作为分类行为的标准。人的运动轮廓特性被用作学习和识别的低级特征;学习是利用HMMs来为每个类别产生行为矩阵,作为模糊联想网络的知识,通过对实时视频流的认知,利用神经网络最终判断行为的种类,在不同情况下模糊联想网络的知识还可以随时学习更新。实验结果表明,算法得到了较好的识别结果,并具有一定的抗噪性。
     3)提出了基于模糊理论的判断异常行为识别方法,解决了判断前需要先定义异常行为的问题,实现了对异常行为的直接判断。首先为模型化人体结构设计了简化的人体关节模型图;其次根据行人躯干和四肢轮廓角度的变化,设计了用于模糊化的函数式;再次提出了利用躯干和四肢的模糊隶属度通过计算来得到整个人异常度的一种基于模糊理论异常行为判别的算法;最后在系统实现中,提出了利用质心轨迹和模糊判别的联合方法来甄别行人是否异常的方法,模糊判别可以实现在视频监控范围内对行人行为的主动分析,从而能够对行人异常的动作做出识别并进行报警处理。通过实验证明该方法具有较高的识别率。
     4)提出了可变场所的异常行为识别方法,解决了在监控应用中不同场景具有不同的异常判断规则的问题,使异常判断算法能应用在多种场所。采用双层词包模型判断在不同场景中行为是否为异常,把视频信息放在第一层包中,把场景动作文本词放在第二层包中。视频由一系列时空兴趣点组成的时空词典表示,动作性质由在指定场景下的动作文本词集合来确定。使用潜在语义分析概率模型(pLSA)不但能自动学习时空词的概率分布,找到与之对应的动作类别,也能在监督情况下学习在规定场景下运动文本词概率分布并区分出对应异常或正常行动结果。经过训练学习后,该算法可以识别新视频在相应场景下行为的异常或正常。
     5)提出一种能自动选择人体最大特性区域的覆盖比算法,解决了目前行人跟踪方法中,跟踪区域需要人工事先设定的问题。人体运动属于非刚性物体的时变和空变问题,在运动中的轮廓也是不断变化的。利用覆盖比算法能自动找到行人在初始时刻所需跟踪的最佳区域,然后选取特性区域内的加权颜色直方图作为跟踪特征,利用Bhattacharyya距离描述颜色模型的相似性,作为粒子权值的有力依据,最后在粒子滤波理论框架下实现自动地对行人进行实时跟踪。
Intelligent Video Surveillance (IVS) is a kind of technology to achieve automatic video analysis with computer techniques. As an effective means of defense and security, IVS systems are being more and more popular. The analysis and recognition of pedestrian’s abnormal behavior in the video sequence, a research objective of IVS, has gradually drawn the attention in the field of IVS. The IVS system based on the analysis of abnormal behavior can not only ignore a large number of useless information, which guarantees the high efficiency in the security protection, but also save a lot of human and material resources, which brings great economic benefits to the whole society. In addition, it is also able to achieve real-time alarming to eliminate the lags in tradition monitoring systems. This thesis, in both the theoretical and the practical perspective, probes into abnormal gait recognition with the videos as input. The following proposed algorithm and methods are carried out in practice application. The main contributions of this thesis are summarized as follows:
     1) The classification method of the minimal standard deviation based on template is presented, which can solve the problem that different people have their special gait, and thus implement the classification actions better. Around the fact that people’s actions are a posture set in time sequence, the image sequences are firstly converted to a set of static shape mode in use of the periodicity of the human movement. And then in the recognition process the mode is compared with behavior of pre-stored samples. In the establishment of standard library stage, the minimal similar degree template generation algorithms is presented, which can solve the template representation problem. Take the Procrustes medium value shape distance as the similarity standard of the image, we can find out the behavior which has the minimal standard deviation from the standard database in one single test cycle. Experiments prove that the minimal standard deviation algorithm can meet the non-continuous identification requirements, and overcome sensitivity of the time interval changes in template matching. The weighted average of the maximal mean value and the minimal standard deviation algorithm is proposed, which take the Zernike speed matrix distance as the similarity standard of the image, can solve the uncertain speed motion and increase the precision of movement classification fundamentally.
     2) A Fuzzy Associative Memory (FAM) networks using behavior classification is proposed, which solve the problem between their movements and appearance, so that the computer recognition process can get closer to the people's thought processes. Each static posture is treated as a state, these states are linked through some kind of probability. Any movement sequences can be treated as a traversal process of a static posture between the different states. The joint probability is calculated during the traversal, and the maximum value is chosen as the behavior classification standards. Pedestrian contour is used as a feature with study and identify in low-level. Behavior matrix for each category is generated with HMMs study. The motion classification is deduced by the knowledge of FAM network. In different circumstances, FAM network knowledge can be updated at any time through learning. This thesis uses four layers fuzzy neural network model, which is a system with multiple input and single output. It has nine input unit, which are eight standard deviations of each action and one centroid. Each input is a membership function. The first layer of the system is the input layer. The second layer is the membership function, whose effect is to turn the input into the membership degree. In the third layer, each node represents a rule that comes from the study algorithm. The relationship between present nodes and previous nodes relies on the rules; the node function relies on the application of the rules. The fourth layer is the output layer. In the system with multiple input and single output, the beginning weight is the membership degree of the rules. Then behavior breed has been obtained through the iteration algorithm. Experimental results show that the algorithm has given a good recognition results, and has a certain degree of noise immunity.
     3) Abnormal behavior of pedestrian detection based on Fuzzy theory is proposed, which solve the problem that the definition of abnormal behavior should be defined before determine, thus reach the direct judgments about abnormal behavior. Subject to certain scenes, scholars in and abroad have proposed methods based on statistical techniques, physical parameters, time-pace movement and model separately. The method of statistical techniques is robust and has fast calculation speed. The method using physical parameters is understandable and observation angle independent, while it depends on the parameter of the recovered scene. The time-space movement method can reveal the character of the time and space, but can easily be disturbed by the noise. The model method has the problem that we can hardly get the precise model from the video. It also calls for massive amount of processing. To detect and tract the particular moving targets in specific environment, our method is unrelated with the scenes. Firstly a simplified human joints model has been established to model the human body. Then a fuzzification function is designed with the variety of body’s trunk and limbs contour angles. Thirdly an abnormal behavior discrimination algorithm based on fuzzy theory is proposed, which applies fuzzy membership of the pedestrian’s trunk and limbs to get the overall degree of the anomaly. Finally in the reality of the system, a combined method of center of mass and fuzzy discriminant is presented. Fuzzy discriminant can detect irregularities and implements initiative analysis to body behavior in the visual surveillance. Therefore, abnormal behaviors can be recognized and alarmed. The results show that the new algorithm has a high recognition rate.
     4) A method of abnormal action recognition in variable scenarios is proposed, which eliminate the ambiguity that one single action of the same person can lead to different comprehension under different circumstances. In the monitoring application, different scenarios have different exceptions to determine the rules of the algorithm and can be applied to determine abnormalities in a variety of places. There are different understanding results in different scenarios even if the same person’s action in visual analysis. In order to determine whether the behavior is abnormal in different scenarios, a double-layer bag-of-words model is proposed to solve the problem in our surveillance system. The video information is processed in the first layer of bag-of-words, and the information of scenario-action text words is included in the second one. A video sequence is represented as a collection of spatial-temporal codebook by extracting space-time interest points. The behavior characteristic is represented as a collection of behavior text words in special scenarios. Probabilistic Latent Semantic Analysis (pLSA) model is adopted to automatically learn the probability distributions of spatial-temporal words and the topics correspond to human action categories. PLSA can also learn the probability distributions of the motion text words in a scenario with supervisor and the topics correspond to anomalous or normal actions. The algorithm can categorize the human anomalous or normal action contained in the special occasion to a novel video sequence after being trained.
     5) A method of automatically selecting characteristics for pedestrian tracking is proposed. The paper presents an algorithm of cover ration which can select the largest feature region of human body automatically to improve the way that the tracking area is pre-designated in current pedestrian tracking methods. Since human motion belongs to the time-varying and space-varying problem of a non-rigid object, the outline of a pedestrian in motion is changing constantly. The algorithm of cover ration can find the best area of the pedestrian which can be tracked at any time automatically. Then it can select a weighted color histogram within the feature region as tracking features and take the similarity of color model described by Bhattacharyya distance as a strong evidence of particle weight. Finally, in the framework of particle filter theory, the real-time tracking of pedestrians can be achieved automatically.
引文
[1] Collins R., et al. A system for video surveillance and monitoring: VSAM final report. Carnegie Mellon University: Technical Report: CMU-RI-TR-00-12, 2000.
    [2] Remagnino P., Tan T., Baker K. Multi-agent visual surveillance of dynamic scenes. Image and Vision Computing. 1998, 16 (8): 529-532.
    [3] Collision P.A. The Application of Camera Based Traffic Monitoring Systems. IEE Seminar on CCTV and Road Surveillance. 1999, 8/1-8/6.
    [4] Remagnino P., Tan T., Baker K. Agent orientated annotation in model based visual surveillance. In: Proc IEEE International Conference on Computer Vision, Bombay, India. 1998, 857-862.
    [5]贾云得.机器视觉.北京:科学出版社, 2004. 1-4.
    [6]章毓晋.图象处理和分析.清华大学出版社, 2001. 3-6.
    [7] Gavrila D M, Giebel J, Munder S. Vision-based pedestrian detection: the protector system. In: Proceedings of IEEE Intelligent Vehicles Symposium. Parma, Italy. 2004, 13-18.
    [8] E. Montseny, J. Frau. Computer vision: specialized processors for real-time image analysis. Workshop proceedings, Barcelona, Spain, September 1991.
    [9] Gavrila D. The visual analysis of human movement: a survey. Computer Vision and Image Understanding. 1999, 73(1): 82-98.
    [10] Elgammal A, Duralswami R, Harwood D, et al. Background and foreground modeling using nonparametric kernel density estimation for visual surveillance. Proceedings of the IEEE, 2002, 90(7):1153-1163.
    [11] Collins R, Lipton A and Kanade T. Introduction to the special section on video surveillance. IEEE Trans Pattern Analysis and Machine Intelligence, 2000, 22 (8): 745-746.
    [12] Maybank S and Tan T. Introduction—special section on visual surveillance. International Journal of Computer Vision. 2000, 37(2): 173-173.
    [13] Yang J and Waibel A. A real-time face tracker. In: Proc IEEE Workshop on Applications of Computer Vision, Sarasota, USA. 1996, 142-147.
    [14] Lakany H., Haycs G, Hazlewood M and Hillman S. Human walking: tracking and analysis. In: Proc IEE Colloquium on Motion Analysis and Tracking, Savoy Place, London. 1999, 5/1-5/14.
    [15] I. Haritaoglu, D. Harwood, and L. S. Davis.: W4 real-time surveillance of peopleand their activities. IEEE Trans. Pattern Analysis and Machine Intelligence. 2000, Aug, 22(8): 809–830.
    [16] Roberto Brunelli and Tomaso Poggio. Face Recognition: Features versus Template, IEEE Transaction Pattern Analysis and Machine Intelligence. 1993, 15(10): 1042-1052.
    [17] Canny J. A computational approach to edge detection. IEEE Trans on Pattern Analysis and Machine Intelligence, 1986, 8(6): 679.
    [18] Shen D, Horace H S Ip. Discriminative Wavelet Shape Descriptors for Recognition of 2-D Patterns. Pattern Recognition. 1999, 32 (2): 151-165.
    [19] Grimson WEL, et al. Using adaptive tracking to Classify and monitor activities in a site. In Proceedings of CVPR, Santa Barbara, CA, 1998.
    [20] Stauffer C, etal. adaptive background mixture models for real-time tracking. IEEE conference on Computer Vision and Pattern Recognition. 1999, 246.
    [21] Weiming Hu, Tieniu Tan, Liang Wang and Steve Maybank, A survey on visual surveillance of object motion and behaviors. IEEE Trans. System, Man, and Cybernetics. 2004, 34(3): 334-352.
    [22] D ING GG, GUO BL. New Fast Motion Estimation Algorithm Based on Line Search. Journal of xipan JiaoTong University, 2004, 38(2): 36 - 39.
    [23] JA IN J, JA IN A. Displacement measurement and its application in inter frame imagecoding. IEEE Trans Communication. 1981, 29(12): 1799-1808.
    [24] CHEUNG SCS, KAMATH C. Robust techniques for background subtraction in urban traffic video. Proceedings of Electronic Imaging: Visual Communications and Image Processing 2004 (Part One). San Jose, California. Bellingham, WA: SPIE, 2004, (5308): 881 - 892.
    [25]候志强,韩崇昭.视觉跟踪技术综述.自动化学报, 2006, 32(4): 603-618.
    [26] C. Shen. Robust visual tracking in image sequences. PhD thesis, School of Computer Science, The University of Adelaide, 2005.
    [27] D. Comaniciu, V. Ramesh and P. Meer. Kernel-based object tracking. IEEE Transaction on Pattern Analysis and Machine Intelligence, 2003, 25(5): 564-577.
    [28] P. Perez, C. Hue, J. Vermaak, et al. Color-based probabilistic tracking. In: Proceedings of 7th European Conference on Computer Vision. 2002: 661-675.
    [29] K. Nummiaro, E. Koller-Meier and L. Van Gool. an adaptive color-based particle filter. Image and Vision Computing, 2003, 21(1): 99-110.
    [30] S. T. Birchfield and S. Rangarajan. Spatiograms versus histograms for region-based tracking. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, United States. 2005, 2:1158-1163.
    [31] N. Peterfreund. Robust tracking of position and velocity with Kalman snakes. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1999, 21(6): 564-569.
    [32] P. H. Li, T. W. Zhang and A. E. C. Pece. Visual contour tracking based on particle filters .Image and Vision Computing. 2003, 21(1): 111-123.
    [33] C. Shen, A. van den Hengel and A. Dick. Probabilistic multiple cue integration for particle filter based tracking. In: International Conference on Digital Image Computing: Techniques and Applications, Sydney. 2003, 1: 399–408.
    [34] A. Li, Z. Jing and S. Hu. Particle filter based visual tracking with multi-cue adaptive fusion. Chinese Optics Letters. 2005, 3(6): 326-329.
    [35] P. Perez, J. Vermaak and A. Blake. Data fusion for visual tracking with particles. Proceedings of the IEEE. 2004, 92(3): 495-513.
    [36] E. Maggio, F.Smeraldi and A. Cavallaro. Combining colour and orientation for adaptive particle filter-based tracking. In: Proceedings of British Machine Vision Conference Oxford, UK 2005.
    [37] B. Han, C. Yang, R. Duraiswami, et al. Bayesian filtering and integral image for visual tracking. In: Proceedings of 6th International Worshop on Image Analysis for Multimedia Interactive Services Montreux, Switzerland, 2005.
    [38] C. Yang, R. Duraiswami and L. Davis. Fast multiple object tracking via a hierarchical particle filter. In: Proceedings of IEEE International Conference on Computer Vision, Beijing, China, 2005, I: 212-219.
    [39] D. DeCarlo and D. Metaxas. optical flow constraints on deformable models with applications to face tracking. International Journal of Computer Vision, 2000, 38(2): 99-127.
    [40] A. D. Jepson, D. J. Fleet and T. F. El-Maraghi. Robust online appearance models for visual tracking. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2003, 25(10): 1296-1311.
    [41] T. Ojala, M. Pietikainen and T. Maenpaa. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2002, 24(7):971-987.
    [42] C. Wen-Yan, C. Chu-Song and H. Yi-Ping. Discriminative descriptor-based observation model for visual tracking. In: Proceedings of 18th International Conference on Pattern Recognition, Hong Kong. 2006, 3: 83-86.
    [43] C.Yang, R. Duraiswami and L. Davis. Efficient mean-shift tracking via a new similarity measure. In: Proceedings of IEEE Computer Society Conference onComputer Vision and Pattern Recognition, San Diego, CA, United States. 2005, I: 176-183.
    [44] H. Wang, D. Suter and K. Schindler. Effective appearance model and similarity measure forparticle filtering and visual tracking. In: Proceedings of 9th European Conference on Computer Vision, Graz, Austria. 2006, (3): 606-618.
    [45] J. Lim, D. Ross, R. S. Lin, et al. Incremental learning for visual tracking, In: L. Saul, et al. (Eds.), Advances in Neural Information Processing Systems, MIT Press, 2005.
    [46] J. Ho, K.-C. Lee, M.-H. Yang, et al. visual tracking using learned linear subspaces. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, United States. 2004, 1, 782-789.
    [47] C. Shen, A. Van Den Hengel and M. J. Brooks. Visual tracking via efficient kernel discriminant subspace learning. In: Proceedings of International Conference on Image Processing, Genova, Italy. 2005, 2, 590-593.
    [48] K.-C. Lee and D. Kriegman. Online learning of probabilistic appearance manifolds for video-based recognition and tracking. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, United States. 2005, I, 852-859.
    [49] C. Sminchisescu and A. Jepon. Generative modeling for continuous non-linearly embedded visual inference. In: Proceeding of International Conference on Machine Learning, Banff, Canada, 2004.
    [50] R.Van der Merwe, A. Doucet, N.de Freitas, et al. The unscented particle filter, Technical report CUED/F-INFENG/TR380, Cambridge University Engineering Department, August, 2000.
    [51] E. Wan and R. van der Merwe. The unscented Kalman Filter, In: S. Haykin(Eds.), Kalman Filtering and Neural Networks, John Wiley&Sons Inc, 2001.
    [52] D. Crisan and A. Doucet. A Survey of convergence results on particle filtering methods for practictioners. IEEE Trans. Speech and Audio Proc. 2002, 10(3): 173-185.
    [53] I. Tierney, A. Mira. Some adaptive monte carlo methods for bayesian inference. Statistics in Medicine. 1999, Vol 18: 2507-2015.
    [54] G. O. Roberts, J. S. Rosenthal. Markov chain monte carlo: some practical implications of theoretical results. Can. J. Stat. 1998, Vol. 25: 5-31.
    [55] F. Liang. Dynamically weighted importance sampling in Monte Carlo computation. Journal of the A merican Statistical Association. 2002, Vol. 97.
    [56] J. S. Liu. Metropolized independent sampling with comparisons to rejectionsampling and importance sampling. Statistical Computation. 1996, Vol. 6, 113-119.
    [57] A. Doucet, N. de Freitas, N. Gordon. Sequential monte carlo methodsin practice. Springer-Veriag, 2001.
    [58] Z. Chen. Bayesian filtering: from kalman filters to particle filters, and beyond. Manuscript. Technical Report,2003
    [59] R. Bailey and M. Srinath. Orthogonal Moment Features for Use with Parametric and Non-Parametric Classifiers. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1996, vol. 18, 389-399.
    [60] D. Comaniciu,V.Ramesh and P. Meer. Kernel-based Object Tracking. IEEE Transaction on Pattern Analysis and Machine Intelligence, 2003, 25(5): 564-577.
    [61] T. Kailath, The Divergence and Bhattacharyya Distance Measures in Signal Selection, IEEE Transactions on Communication Technology COM-15(1). 1967, 52-60.
    [62] Blank, M., Gorelick, L., Shechtman, E., Irani, M et al. Actions as space-time shapes. In Proceedings of the tenth IEEE international conference on computer vision, 2005, 1395–1402.
    [63] Gonzalez J, Varona J ,Roca F X, etal. A Spaces: Action spaces for recognition and synthesis of human actions. Lecture Notes in Computer Science, 2002, 2492:189-200.
    [64] Ju S X, Black M J, Yacoob Y. Cardboard people: a parameterized model of articulated image motion. In Proc. Automatic Face and Gesture Recognition.Vermont: IEEE Computer Society Press, 1996.38-44.
    [65] Arie J B, Wang Z, Pandit P, Rajaram S. Human activity recognition using multidimensional indexing. IEEE Trans. Pattern Analysis 2002, 24 (8): 1091-1104.
    [66] DuYou-tian, ChenFeng, Xu wen-li, Li Yong-bin. A Survey on the Vision-based Human Motion Recogntion. ACTA ELECTRONICA SINICA, 2007, 35(1):84-90.
    [67] J.K.Aggarwal, Q.Cai. Human Motion Analysis: A Review. Computer Vision and Image Understanding, 1999, 73(3): 428-440.
    [68] M.k.Leung. Y.H. Yang. First Sight: A Human Body Outline Labeling System. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1995, 17(4): 359-377.
    [69] J.Deutscher, A.Blake, I.Reid, Articulated body motion capture by annealed particle filtering. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head, USA. 2000, volume 2, 126-133.
    [70] H.Sidenbladh, M.J.Black, D.J.Fleet, Stochastic tracking of 3D human figures using 2D image motion, Proceedings of the 6th European Conference on Computer Vision, Dublin, Ireland. 2000, 702-718.
    [71] H.Sidenbladh, Probabilistic Tracking and Reconstruction of 3D Human Motion inMonocular Video Sequences, PHD Thesis, Department of Numerical Analysis and Computer Science, KTH, Sweden, 2001.
    [72] Fujiyoshi H, Lipton A. Real-time human motion analysis by image skeletonization. In: Proc IEEE Workshop on Applications of Computer Vision, Princeton, NJ, 1998. 15-21.
    [74] Dollár, P., Rabaud, V., Cottrell et al. Behavior recognition via sparse spatio-temporal features. In Proc. 2nd joint IEEE international workshop on visual surveillance and performance evaluation of tracking and surveillance, Beijing, China. 2005, 65—72,
    [75] Zadeh, L.A. Fuzzy sets. Information and Control. 1965. 8 (3): 338–353.
    [76] Hu MK. Visual Pattern Recognition by Moment Invariants. IRE Trans. Information theory, 1962, IT (8):179-187.
    [77] J. D. Shutler and M. S. Nixon, Zernike velocity moments for sequence-based description of moving features. Image and Vision Computing. 2006, 24: 343-356.
    [78] Oren Boiman and Michal Irani. Detecting irregularities in images and in video. Proc. IEEE int. conf. Computer Vision. 2005, 462~469.
    [79] T. Takagi and M. Sugeno, K. Fuzzy identification of systems and its applications to modeling and control. IEEE Trans. Syst., Man, Cybern., 1985, SMC-15(1): 116–132.
    [80] Weina Wang, Yunjie Zhang, Yi Li and Xiaona Zhang, K. The Global Fuzzy C-Means Clustering Algorithm. Proceedings of the 6th World Congress on Intelligent Control and Automation, June 21 - 23, 2006, Dalian, China, pp. 3604-3607
    [81] Scott M. Thede, Mary P. Harper. A second-order Hidden Markov Model for part-of-speech tagging. Proceedings of the 37th annual meeting of the Association for Computational Linguistics on Computational Linguistics. 1999, 175– 182.
    [82] B. Kosko. Neural Networks and Fuzzy Systems. A Dynamical Systems Approach to Machine Intelligence. 1987. Englewood Cliffs, NJ: Prentice Hall.
    [83] Hsien-cheng Chang, Hui-Chuan Chen, and Jen-Ho Fang. Lithology Determination from Well Logs with Fuzzy Associative Memory Neural Network. IEEE Trans. On Geoscience and Remote Sensing. 1997, 35(3): 773-780.
    [84] P. D. Wasserman. Theory and Practice. Neural Computing. 1989 New York: Van Nostrand Reinhold.
    [85] Polana R and Nelson R. Low level recognition of human motion. In: Proc IEEE Workshop on Motion of Non-Rigid and Articulated Objects, Austin, TX. 1994, 77-82.
    [86] Starner T and Pentland A. Real-time American Sign Language recognition from video using hidden Markov models. In: Proc International Symposium on ComputerVision, Coral Gables, Florida. 1995, 265-270.
    [87] McKenna S et al, Tracking groups of people. Computer Vision and Image Understanding. 2000, 80 (1): 42-56.
    [88] Feng X, Perona P. Human action recognition sequence of movelet codewords. In Proc. 3DPVT. Italy: IEEE Computer Society Press. 2002. 717-723.
    [89] Ziheng Zhou, Adam Prugel-Bennett, Robert I. Damper. A Bayesian Framework for Extracting Human Gait Using Strong Prior Knowledge. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2006. 1738-1752
    [90] McLachlan, Skye Arblaster, Jaimal Liu, etc. A multi-stage shared control method for an intelligent mobility assistant. Proceedings of the 2005 IEEE 9th International Conference on Rehabilitation Robotics, Chicago, 2005. 426-429.
    [91] LONG Hui ping, XI Sheng feng , HOU Xin hua. The Algorithm and Analysis from Experiment Data by Means of Minimum Double Multiplication. Computing Technology and Automation. 2008, 27(3):20-23.
    [92] Rafael C.Gonzalez, Richard E.Woods. Digital Image Processing. Publishing House of Electronics Industry, 2002.
    [93] Sun H, Feng T and Tan T. Robust extraction of moving objects from image sequences. In: Proc the Fourth Asian Conference on Computer Vision, Taiwan, 2000. 961-964.
    [94] J. C. Niebles, H.Wang, L. Fei-Fei. Unsupervised learning of human action categories using spatialtemporal words. International Journal of Computer Vision. 2008, 79(3): 299–318
    [95] T. Hofmann. Probabilistic latent semantic indexing. In Proc. Twenty-Second Annual International Conference on Research and Development in Information Retrieval, 1999, 50—57, California, USA.
    [96] HU Zhi-Lan, JIANG Fan, WANG Gui-Jin, etc. Anomaly Detection Based on Motion Direction. ACTA AUTOMATICA SINICA, 2008, 34(11):1348-1357.
    [97] Cbsr. Center for Biometrics and Security Research [EB/OL], 2008, Base http://www.cbsr.ia.ac.cn/china/Action%20Databases%20CH.asp
    [98]孔晓东智能视频监控技术研究上海交通大学博士论文2008.3
    [99]王亮,胡卫明,谭铁牛.人运动的视觉分析综述.计算机学报. 2002, 25(3).
    [100] Maggioni C and Kammerer B. Gesture Computer: history, design, and applications. Computer Vision for Human-Machine Interaction, Cambridge Univ. Press, 1998.
    [101] Freeman W and Weissman C. Television control by hand gestures. In: ProcInternational Conference on Automatic Face and Gesture Recognition, Zurich, Switzerland, 1995, 179-183.
    [102] Lipton A, Fujiyoshi H and Patil R. Moving target classification and tracking from real-time video. In: Proc IEEE Workshop on Applications of Computer Vision, Princeton, NJ, 1998, 8-14.
    [103] Anderson C, Bert P and Vander Wal G. Change detection and tracking using pyramids transformation techniques. In: Proc SPIE Conference on Intelligent Robots and Computer Vision, Cambridge, MA, 1985, 579: 72-78.
    [104] Meyer D, Denzler J and Niemann H. Model based extraction of articulated objects in image sequences for gait analysis. In: Proc IEEE International Conference on Image Processing, Santa Barbara, California 1997, 78-81
    [105] Barron J, Fleet D and Beauchemin S. Performance of optical flow techniques. International Journal of Computer Vision. 1994, 12 (1): 42-77.
    [106] Verri A, Uras S and DeMicheli E. Motion Segmentation from optical flow. In: Proc the 5th Alvey Vision Conference, Brighton, UK, 1989, 209-214.
    [107] Welch G and Bishop G. An introduction to the Kalman filter. In: http://www.cs.unc.edu, UNC-ChapelHill, TR95-041, 2000.
    [108] Isard M and Blake A. Condensation—conditional density propagation for visual tracking. International Journal of Computer Vision, 1998, 29 (1): 5-28.
    [109] Pavlovi? V, Rehg J, Cham T-J and Murphy K. A dynamic Bayesian network approach to figure tracking using learned dynamic models. In: Proc IEEE International Conference on Computer Vision, Corfu, Greece, 1999, 94-101.
    [110] Myers C, Rabinier L and Rosenberg A. Performance tradeoffs in dynamic time warping algorithms for isolated word recognition. IEEE Trans Acoustics, Speech, and Signal Processing. 1980, 28 (6): 623-635.
    [111] Bobick A and Wilson A. A state-based technique for the summarization and recognition of gesture. In: Proc International Conference on Computer Vision, Cambridge. 1995, 382-388.
    [112] Poritz A. Hidden Markov Models: a guided tour. In: Proc IEEE International Conference on Acoustics, Speech and Signal Processing, New York City, NY, 1988, 7-13.
    [113] Brand M, Oliver N and Pentland A. Coupled hidden Markov models for complex action recognition. In: Proc IEEE Conference Computer Vision and Pattern Recognition, Puerto Rico, 1997, 994-999.
    [114] Guo Y, Xu G and Tsuji S. Understanding human motion patterns. In: ProcInternational Conference on Pattern Recognition, Jerusalem, Israel, 1994, 325-329
    [115] Schmid C , Mohr R. Local Grayvalue Invariants for Image Retrieval. IEEE Trans on Pattern Anal Mach Intell , 1997, 19 (5): 530-535.
    [116] Kitchen L , Ro senfeld A. Gray-Level Corner Detection. Pattern Reccognition Letters, 1982, 1:95-102
    [117] Dorthe Meyer, Josef and Heinrich Niemann. Gait classification with HMMs for Trajectories of body parts extracted by mixture densities. British Machine Vision Conference, England. 1998, 459-468.
    [118] Tassone, E., West, G., Venkatesh, S. Temporal PDMs for gait classification. Pattern Recognition, 2002. Proceedings. 16th International Conference. 2002, 1065–1068.
    [119] Rita Cucchiara, Costantino Grana, Andrea Prati, etc. Probabilistic Posture Classification for Human-Behavior Analysis. IEEE Trans. System, Man, and Cybernetics. 2005, 35(1): 42-54.
    [120] Q. Meng, B. Li, H. Holstein. Recognition of human periodic movements from unstructured information using a motion-based frequency domain approach. Image and Vision Computing. 24, 795–809.
    [121] Chee Seng Chan, Honghai Liu and David J. Brown. Recognition of Human Motion From Qualitative Normalised Templates. Journal of Intelligent and Robotic Systems. 2007, 48, 79-95.
    [122] Chia-Feng Juang, and Chia-Ming Chang. Human Body Posture Classification by a Neural Fuzzy Network and Home Care System Application. IEEE Trans. System, Man, and Cybernetics. 2007, 37(6): 984-994.
    [123] Preben Fihl, Thomas B. Moeslund. Classification of Gait Types Based on the Duty-Factor. Advanced Video and Signal Based Surveillance, AVSS 2007, IEEE Conference. 2007, 318-323.
    [124] Liang Wang, David Suter. Visual learning and recognition of sequential data manifolds with applications to human movement analysis. Computer Vision and Image Understanding. 2008, 110: 153–172.
    [125] Liang Wang, Tieniu Tan, etc. Automatic Gait Recognition Based on Statistical Shape Analysis. IEEE Trans. Image Processing. 2003, 12(93): 1120~1131.
    [126] Horn B K P, Schunk B G. Determining optical flow. Artificial Intelligence, 1981, 17(1-3): 185-203.
    [127] CORTES C, VAPNIK V. Support vector machine. Machine Learning, 1995, 20(3):273-297.
    [128] E.T. Lee and S.C. Lee. Fuzzy sets and neural networks. Cybernetics. 1974, 4(2):83–101.
    [129]龙辉平,习胜丰,侯新华.实验数据的最小二乘拟合算法与分析.计算技术与自动化, 2008, 27(3): 20-23.
    [130]印勇,张毅,刘丹平.基于改进Hu矩的异常行为识别.计算机技术与发展, 2009, 19(9): 90-92.
    [131]李和平,胡占义,吴毅红,吴福朝.基于半监督学习的行为建模与异常检测.软件学报. 2007, 18(3):527?537.
    [132]陈宜稳,王威,王润生.基于视频区域特征的行人异常行为检测.计算机应用. 2007, 27(10): 2610-1612.
    [133] Lao Weilun, Han Jungong, de With Peter H.N. Automatic video-based human motion analyzer for consumer surveillance system. IEEE Transactions on Consumer Electronics. 2009, 55(2): 591-598.
    [134] Haering Niels, Venetianer Péter L, Lipton, Alan. The evolution of video surveillance: An overview. Machine Vision and Applications. 2008, 19(5-6): 279-290.
    [135] Hsieh Jun-Wei, Hsu Yung-Tai, Liao Hong-Yuan Mark, Chen Chih-Chiang. Video-based human movement analysis and its application to surveillance systems. IEEE Transactions on Multimedia. 2008, 10(3): 372-384.
    [136] Morris Brendan Tran, Trivedi. Mohan Manubhai. A survey of vision-based trajectory learning and analysis for surveillance. IEEE Transactions on Circuits and Systems for Video Technology. 2008, 18(8): 1114-1127.
    [137] Chung Pau-Choo, Liu Chin-De. A daily behavior enabled hidden Markov model for human behavior understanding. Pattern Recognition. 2008, 41(5): 1589-1597.
    [138] Anjum Nadeem, Cavallaro Andrea. Multifeature object trajectory clustering for video analysis. IEEE Transactions on Circuits and Systems for Video Technology. 2008, 18(11): 1555-1564.
    [139] Xiang Tao, Gong, Shaogang. Video behavior profiling for anomaly detection. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2008, 30(5): 893-908.
    [140] Lin H T, Lin C J. A study on sigmoid kernels for SVM and the training of non-PSD kernels by SM()type methods [EB/OI]. [2003].h ttp://www.csie.ntu. edu. tw/~cjlin/papers/tanh.pdf.