Abstract
We present a regression method for the estimation of hand orientation using an uncalibrated camera. For training the system, we use a depth camera to capture a large dataset of hand color images and orientation angles. Each color image is segmented producing a silhouette image from which contour distance features are extracted. The orientation angles are captured by robustly fitting a plane to the depth image of the hand, providing a surface normal encoding the hand orientation in 3D space. We then train multiple Random Forest regressors to learn the non-linear mapping from the space of silhouette images to orientation angles. For online testing of the system, we only require a standard 2D image to infer the 3D hand orientation. Experimental results show the approach is computationally efficient, does not require any camera calibration, and is robust to inter-person shape variation.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
References
Allison, S.: Wearable tech - the future, or just a fad? (February 2014) (Online; posted February 13, 2014)
Olsson, M.I., Martin, M.W., Hebenstreit, J.J., Cazalet, P.M.: Wearable device with input and output structures. US Patent App. 14/037, 788 (2013)
Erol, A., Bebis, G., Nicolescu, M., Boyle, R.D., Twombly, X.: Vision-based hand pose estimation: A review. Computer Vision and Image Understanding 108(1), 52–73 (2007)
Wu, Y., Huang, T.S.: Capturing articulated human hand motion: A divide-and-conquer approach. In: The Proceedings of the Seventh IEEE International Conference on Computer Vision (ICCV), vol. 1, pp. 606–611. IEEE (1999)
Rosales, R., Athitsos, V., Sigal, L., Sclaroff, S.: 3d hand pose reconstruction using specialized mappings. In: Proceedings of Eighth IEEE International Conference on Computer Vision (ICCV), vol. 1, pp. 378–385. IEEE (2001)
de La Gorce, M., Fleet, D.J., Paragios, N.: Model-based 3d hand pose estimation from monocular video. IEEE Transactions on Pattern Analysis and Machine Intelligence 33(9), 1793–1805 (2011)
Lee, T., Hollerer, T.: Handy ar: Markerless inspection of augmented reality objects using fingertip tracking. In: 11th IEEE International Symposium on Wearable Computers, pp. 83–90. IEEE (2007)
Lee, T., Hollerer, T.: Hybrid feature tracking and user interaction for markerless augmented reality. In: IEEE Virtual Reality Conference (VR 2008), pp. 145–152. IEEE (2008)
Lee, T., Hollerer, T.: Multithreaded hybrid feature tracking for markerless augmented reality. IEEE Transactions on Visualization and Computer Graphics 15(3), 355–368 (2009)
Kato, H., Kato, T.: A marker-less augmented reality based on fast fingertip detection for smart phones. In: IEEE International Conference on Consumer Electronics (ICCE), pp. 127–128. IEEE (2011)
Agarwal, A., Triggs, B.: Recovering 3d human pose from monocular images. IEEE Transactions on Pattern Analysis and Machine Intelligence 28(1), 44–58 (2006)
Elgammal, A., Lee, C.-S.: Inferring 3d body pose from silhouettes using activity manifold learning. In: Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), vol. 2, pp. II-681. IEEE (2004)
Franco, J.-S., Boyer, E.: Fusion of multiview silhouette cues using a space occupancy grid. In: Tenth IEEE International Conference on Computer Vision (ICCV), vol. 2, pp. 1747–1753. IEEE (2005)
Causo, A., Ueda, E., Kurita, Y., Matsumoto, Y., Ogasawara, T.: Model-based hand pose estimation using multiple viewpoint silhouette images and unscented kalman filter. In: The 17th IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN, pp. 291–296. IEEE (2008)
Breiman, L.: Random forests. Machine Learning 45(1), 5–32 (2001)
Yoruk, E., Konukoglu, E., Sankur, B., Darbon, J.: Shape-based hand recognition. IEEE Transactions on Image Processing 15(7), 1803–1815 (2006)
Leventon, M.E., Grimson, W.E.L., Faugeras, O.: Statistical shape influence in geodesic active contours. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 1, pp. 316–323. IEEE (2000)
Šarić, M.: Libhand: A library for hand articulation, Version 0.9 (2011)
Shotton, J., Girshick, R., Fitzgibbon, A., Sharp, T., Cook, M., Finocchio, M., Moore, R., Kohli, P., Criminisi, A., Kipman, A., et al.: Efficient human pose estimation from single depth images. IEEE Transactions on Pattern Analysis and Machine Intelligence 35(12), 2821–2840 (2013)
Keskin, C., Kıraç, F., Kara, Y.E., Akarun, L.: Real time hand pose estimation using depth sensors. In: Consumer Depth Cameras for Computer Vision, pp. 119–137. Springer (2013)
Sharp, T.: Implementing decision trees and forests on a gpu. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008, Part IV. LNCS, vol. 5305, pp. 595–608. Springer, Heidelberg (2008)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Asad, M., Slabaugh, G. (2014). Hand Orientation Regression Using Random Forest for Augmented Reality. In: De Paolis, L., Mongelli, A. (eds) Augmented and Virtual Reality. AVR 2014. Lecture Notes in Computer Science(), vol 8853. Springer, Cham. https://doi.org/10.1007/978-3-319-13969-2_13
Download citation
DOI: https://doi.org/10.1007/978-3-319-13969-2_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-13968-5
Online ISBN: 978-3-319-13969-2
eBook Packages: Computer ScienceComputer Science (R0)