Skip to main content

Human Activity Recognition in Video Sequences Based on the Integration of Optical Flow and Appearance of Human Objects

  • Conference paper
  • First Online:
Robotics, Control and Computer Vision

Part of the book series: Lecture Notes in Electrical Engineering ((LNEE,volume 1009))

  • 313 Accesses

Abstract

Recognition of human activities has emerged as a critical research area due to its potential applications in many automated monitoring applications. However, it is still a challenging problem due to inter- and intra-class variations in human activities, varying illumination conditions, viewpoint changes, etc. This work presents human activity recognition framework for motion activities recorded in realistic and multi-view environments. To represent complex motion activities, we designed a novel feature descriptor based on the integration of motion information and the appearance of moving human objects. For this purpose, first, we employed an object segmentation technique capable of dealing with camera motion, varying lighting conditions, and scale of the human object. Then we used the optical flow technique to compute each moving pixel’s velocity vector and orientation by avoiding background noise. Then histogram of oriented gradients of velocity and orientation information is computed to get the relative distribution of motion information for samples of each activity category. Finally, a feature fusion strategy integrates a local-oriented histogram of velocity and orientation information to construct the final feature vector. Support vector machine is used to compute the class score of activity categories. In order to give an empirical justification, we conducted several extensive experiments on three publically available datasets, namely IXMAS, UT Interaction, and CASIA. To demonstrate the effectiveness of the proposed methods, we compared their results with several state-of-the-art methods. The recognition results demonstrate the supremacy of the proposed method over the other state-of-the-art methods considered for comparison.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 249.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Kushwaha A, Khare A, Srivastava P (2021) On integration of multiple features for human activity recognition in video sequences. Multimedia Tools Appl 1–28

    Google Scholar 

  2. Ladjailia A, Bouchrika I, Merouani HF, Harrati N, Mahfouf Z (2020) Human activity recognition via optical flow: decomposing activities into basic actions. Neural Comput Appl 32(21):16387–16400

    Article  Google Scholar 

  3. Khare M, Binh NT, Srivastava RK (2014) Human object classification using dual tree complex wavelet transform and Zernike moment. In: Transactions on large-scale data and knowledge-centered systems, vol XVI. Springer, Berlin, Heidelberg, pp 87–101

    Google Scholar 

  4. Srivastava P, Khare A (2018) Utilizing multiscale local binary pattern for content-based image retrieval. Multimedia Tools Appl 77(10):12377–12403

    Article  Google Scholar 

  5. Sivic J, Zisserman A (2003) Video Google: a text retrieval approach to object matching in videos. In: Proceedings ninth IEEE international conference on computer vision, Nice, France, vol 1, pp 1470–1477. https://doi.org/10.1109/ICCV.2003.1238663

  6. Sánchez J, Perronnin F, Mensink T, Verbeek J (2013) Image classification with the fisher vector: theory and practice. Int J Comput Vis 105(3):222–245

    Article  MathSciNet  MATH  Google Scholar 

  7. Souly N, Shah M (2016) Visual saliency detection using group lasso regularization in videos of natural scenes. Int J Comput Vis 117(1):93–110

    Article  MathSciNet  Google Scholar 

  8. Kushwaha A, Khare A, Prakash O, Khare M (2020) Dense optical flow based background subtraction technique for object segmentation in moving camera environment. IET Image Proc 14(14):3393–3404

    Article  Google Scholar 

  9. Kushwaha A, Prakash O, Srivastava RK, Khare A (2019) Dense flow-based video object segmentation in dynamic scenario. In: Recent trends in communication, computing, and electronics. Springer, Singapore, pp 271–278

    Google Scholar 

  10. Al-Faris M, Chiverton J, Yang L, Ndzi D (2017) Appearance and motion information based human activity recognition. In: IET 3rd international conference on intelligent signal processing (ISP 2017). IET, pp 1–6

    Google Scholar 

  11. Farnebäck G (2003) Two-frame motion estimation based on polynomial expansion. In: Scandinavian conference on image analysis. Springer, Berlin, Heidelberg, pp 363–370

    Google Scholar 

  12. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: Proceedings of IEEE computer society conference on computer vision and pattern recognition, vol 1, pp 886–893

    Google Scholar 

  13. Li X (2007) HMM based action recognition using oriented histograms of optical flow field. Electron Lett 43(10):560–561

    Article  Google Scholar 

  14. Kim SJ, Kim SW, Sandhan T, Choi JY (2014) View invariant action recognition using generalized 4D features. Pattern Recogn Lett 49:40–47

    Article  Google Scholar 

  15. Ryoo MS, Aggarwal JK (2009) Spatio-temporal relationship match: video structure comparison for recognition of complex human activities. In: 2009 IEEE 12th international conference on computer vision. IEEE, pp 1593–1600

    Google Scholar 

  16. Wang Y, Huang K, Tan T (2007) Human activity recognition based on r transform. In: 2007 IEEE conference on computer vision and pattern recognition, pp 1–8

    Google Scholar 

  17. Singh R, Dhillon JK, Kushwaha AK, Srivastava R (2019) Depth based enlarged temporal dimension of 3D deep convolutional network for activity recognition. Multimedia Tools Appl 78(21):30599–30614

    Article  Google Scholar 

  18. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556

  19. Kushwaha A, Khare A (2021) Human activity recognition by utilizing local ternary pattern and histogram of oriented gradients. In: Proceedings of international conference on big data, machine learning and their applications. Springer, Singapore, pp 315–324

    Google Scholar 

  20. Kushwaha A, Khare A, Khare M (2021) Human activity recognition algorithm in video sequences based on integration of magnitude and orientation information of optical flow. Int J Image Graph 22:2250009

    Article  Google Scholar 

  21. Yeffet L, Wolf L (2009) Local trinary patterns for human action recognition. In: 2009 IEEE 12th international conference on computer vision, pp 492–497

    Google Scholar 

  22. Nigam S, Khare A (2016) Integration of moment invariants and uniform local binary patterns for human activity recognition in video sequences. Multimedia Tools Appl 75(24):17303–17332

    Article  Google Scholar 

  23. Seemanthini K, Manjunath SS (2018) Human detection and tracking using HOG for action recognition. Procedia Comput Sci 132:1317–1326

    Article  Google Scholar 

  24. Boser BE, Guyon IM, Vapnik VN (1992) A training algorithm for optimal margin classifiers. In: Proceedings of the fifth annual workshop on Computational learning theory, pp 144–152

    Google Scholar 

  25. Aly S, Sayed A (2019) Human action recognition using bag of global and local Zernike moment features. Multimedia Tools Appl 78(17):24923–24953

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by the Science and Engineering Research Board (SERB), Department of Science and Technology (DST), New Delhi, India, under Grant No. CRG/2020/001982.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ashish Khare .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Kushwaha, A., Khare, A. (2023). Human Activity Recognition in Video Sequences Based on the Integration of Optical Flow and Appearance of Human Objects. In: Muthusamy, H., Botzheim, J., Nayak, R. (eds) Robotics, Control and Computer Vision. Lecture Notes in Electrical Engineering, vol 1009. Springer, Singapore. https://doi.org/10.1007/978-981-99-0236-1_9

Download citation

Publish with us

Policies and ethics