Abstract
Recognition of human activities has emerged as a critical research area due to its potential applications in many automated monitoring applications. However, it is still a challenging problem due to inter- and intra-class variations in human activities, varying illumination conditions, viewpoint changes, etc. This work presents human activity recognition framework for motion activities recorded in realistic and multi-view environments. To represent complex motion activities, we designed a novel feature descriptor based on the integration of motion information and the appearance of moving human objects. For this purpose, first, we employed an object segmentation technique capable of dealing with camera motion, varying lighting conditions, and scale of the human object. Then we used the optical flow technique to compute each moving pixel’s velocity vector and orientation by avoiding background noise. Then histogram of oriented gradients of velocity and orientation information is computed to get the relative distribution of motion information for samples of each activity category. Finally, a feature fusion strategy integrates a local-oriented histogram of velocity and orientation information to construct the final feature vector. Support vector machine is used to compute the class score of activity categories. In order to give an empirical justification, we conducted several extensive experiments on three publically available datasets, namely IXMAS, UT Interaction, and CASIA. To demonstrate the effectiveness of the proposed methods, we compared their results with several state-of-the-art methods. The recognition results demonstrate the supremacy of the proposed method over the other state-of-the-art methods considered for comparison.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Kushwaha A, Khare A, Srivastava P (2021) On integration of multiple features for human activity recognition in video sequences. Multimedia Tools Appl 1–28
Ladjailia A, Bouchrika I, Merouani HF, Harrati N, Mahfouf Z (2020) Human activity recognition via optical flow: decomposing activities into basic actions. Neural Comput Appl 32(21):16387–16400
Khare M, Binh NT, Srivastava RK (2014) Human object classification using dual tree complex wavelet transform and Zernike moment. In: Transactions on large-scale data and knowledge-centered systems, vol XVI. Springer, Berlin, Heidelberg, pp 87–101
Srivastava P, Khare A (2018) Utilizing multiscale local binary pattern for content-based image retrieval. Multimedia Tools Appl 77(10):12377–12403
Sivic J, Zisserman A (2003) Video Google: a text retrieval approach to object matching in videos. In: Proceedings ninth IEEE international conference on computer vision, Nice, France, vol 1, pp 1470–1477. https://doi.org/10.1109/ICCV.2003.1238663
Sánchez J, Perronnin F, Mensink T, Verbeek J (2013) Image classification with the fisher vector: theory and practice. Int J Comput Vis 105(3):222–245
Souly N, Shah M (2016) Visual saliency detection using group lasso regularization in videos of natural scenes. Int J Comput Vis 117(1):93–110
Kushwaha A, Khare A, Prakash O, Khare M (2020) Dense optical flow based background subtraction technique for object segmentation in moving camera environment. IET Image Proc 14(14):3393–3404
Kushwaha A, Prakash O, Srivastava RK, Khare A (2019) Dense flow-based video object segmentation in dynamic scenario. In: Recent trends in communication, computing, and electronics. Springer, Singapore, pp 271–278
Al-Faris M, Chiverton J, Yang L, Ndzi D (2017) Appearance and motion information based human activity recognition. In: IET 3rd international conference on intelligent signal processing (ISP 2017). IET, pp 1–6
Farnebäck G (2003) Two-frame motion estimation based on polynomial expansion. In: Scandinavian conference on image analysis. Springer, Berlin, Heidelberg, pp 363–370
Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: Proceedings of IEEE computer society conference on computer vision and pattern recognition, vol 1, pp 886–893
Li X (2007) HMM based action recognition using oriented histograms of optical flow field. Electron Lett 43(10):560–561
Kim SJ, Kim SW, Sandhan T, Choi JY (2014) View invariant action recognition using generalized 4D features. Pattern Recogn Lett 49:40–47
Ryoo MS, Aggarwal JK (2009) Spatio-temporal relationship match: video structure comparison for recognition of complex human activities. In: 2009 IEEE 12th international conference on computer vision. IEEE, pp 1593–1600
Wang Y, Huang K, Tan T (2007) Human activity recognition based on r transform. In: 2007 IEEE conference on computer vision and pattern recognition, pp 1–8
Singh R, Dhillon JK, Kushwaha AK, Srivastava R (2019) Depth based enlarged temporal dimension of 3D deep convolutional network for activity recognition. Multimedia Tools Appl 78(21):30599–30614
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
Kushwaha A, Khare A (2021) Human activity recognition by utilizing local ternary pattern and histogram of oriented gradients. In: Proceedings of international conference on big data, machine learning and their applications. Springer, Singapore, pp 315–324
Kushwaha A, Khare A, Khare M (2021) Human activity recognition algorithm in video sequences based on integration of magnitude and orientation information of optical flow. Int J Image Graph 22:2250009
Yeffet L, Wolf L (2009) Local trinary patterns for human action recognition. In: 2009 IEEE 12th international conference on computer vision, pp 492–497
Nigam S, Khare A (2016) Integration of moment invariants and uniform local binary patterns for human activity recognition in video sequences. Multimedia Tools Appl 75(24):17303–17332
Seemanthini K, Manjunath SS (2018) Human detection and tracking using HOG for action recognition. Procedia Comput Sci 132:1317–1326
Boser BE, Guyon IM, Vapnik VN (1992) A training algorithm for optimal margin classifiers. In: Proceedings of the fifth annual workshop on Computational learning theory, pp 144–152
Aly S, Sayed A (2019) Human action recognition using bag of global and local Zernike moment features. Multimedia Tools Appl 78(17):24923–24953
Acknowledgements
This work was supported by the Science and Engineering Research Board (SERB), Department of Science and Technology (DST), New Delhi, India, under Grant No. CRG/2020/001982.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Kushwaha, A., Khare, A. (2023). Human Activity Recognition in Video Sequences Based on the Integration of Optical Flow and Appearance of Human Objects. In: Muthusamy, H., Botzheim, J., Nayak, R. (eds) Robotics, Control and Computer Vision. Lecture Notes in Electrical Engineering, vol 1009. Springer, Singapore. https://doi.org/10.1007/978-981-99-0236-1_9
Download citation
DOI: https://doi.org/10.1007/978-981-99-0236-1_9
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0235-4
Online ISBN: 978-981-99-0236-1
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)