Head and Hand Detection using Kinect Camera 360

Download Full Text
Mostafa Karbasi, Ahmad Waqas, Parham Nooralishahi, Seyed Mohammad Reza Mazloomnezhad
Published Date:
June 05, 2015
Volume 4, Issue 6
1 - 7

kinect, head detection, hand detection, indoor
Mostafa Karbasi, Ahmad Waqas, Parham Nooralishahi, Seyed Mohammad Reza Mazloomnezhad, "Head and Hand Detection using Kinect Camera 360". International Journal of Research in Computer Science, 4 (6): pp. 1-7, June 2015. Other Formats


Using head and hand blobs as an input to the computer are very crucial for human-computer interaction (HCI) applications. These blobs play an important role in bridging the information gap between a human and computer. One of the famous technologies that play a crucial role as an advanced input device for HCI is the Kinect camera developed by Microsoft. Kinect camera (codenamed Project Nathal) has a distinct advantage over other 3D cameras because it obtains more accurate depth information of a subject easily and very fast. By using Kinect, one can track up to six people concurrently and also obtain motion analysis with feature extraction. Being extremely useful in indoor HCI applications, it cannot be used in outdoor applications because its infrared depth sensor makes it extremely sensitive to sunlight

  1. Lee, S.-H. and S.-H. Oh, A Kinect Sensor based Windows Control Interface. International Journal of Control & Automation, 2014. 7(3).
  2. Kean, S., J. Hall, and P. Perry, Meet the Kinect: An Introduction to Programming Natural User Interfaces. 2011: Apress.
  3. Hsu, H.-m.J., The potential of kinect in education. International Journal of Information and Education Technology, 2011. 1(5): p. 365-370.
  4. Wang, J.-G. and E. Sung, EM enhancement of 3D head pose estimated by point at infinity. Image and Vision Computing, 2007. 25(12): p. 1864-1874.
  5. Leyvand, T., et al., Kinect identity: Technology and experience. Computer, 2011. 44(4): p. 94-96.
  6. Schramm, M., Kinect: The company behind the tech explains how it works. AOL Tech, http://www. joystiq. com/2010/06/19/kinect-how-it-works-from-the-company-behind-the-tech, 2010.
  7. Jaemin, L., et al. A robust gesture recognition based on depth data. in Frontiers of Computer Vision,(FCV), 2013 19th Korea-Japan Joint Workshop on. 2013. IEEE.
  8. Supplies, P., 3-D-Sensing Technology to “Project Natal” for Xbox 360.
  9. Shao, L., et al., Computer vision for RGB-D sensors: Kinect and its applications. IEEE transactions on cybernetics, 2013. 43(5): p. 1314-1317.
  10. Miranda, L., et al. Real-time gesture recognition from depth data through key poses learning and decision forests. in Graphics, Patterns and Images (SIBGRAPI), 2012 25th SIBGRAPI Conference on. 2012. IEEE.
  11. Rimkus, K., et al. 3D human hand motion recognition system. in Human System Interaction (HSI), 2013 The 6th International Conference on. 2013. IEEE.
  12. Guo, K., P. Ishwar, and J. Konrad. Action recognition in video by covariance matching of silhouette tunnels. in Computer Graphics and Image Processing (SIBGRAPI), 2009 XXII Brazilian Symposium on. 2009. IEEE.
  13. Tuzel, O., F. Porikli, and P. Meer, Region covariance: A fast descriptor for detection and classification, in Computer Vision–ECCV 2006. 2006, Springer. p. 589-600.
  14. Erol, A., et al., Vision-based hand pose estimation: A review. Computer Vision and Image Understanding, 2007. 108(1): p. 52-73.
  15. Athitsos, V. and S. Sclaroff. Estimating 3D hand pose from a cluttered image. in Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on. 2003. IEEE.
  16. Rosales, R., et al. 3D hand pose reconstruction using specialized mappings. in Computer Vision, 2001. ICCV 2001. Proceedings. Eighth IEEE International Conference on. 2001. IEEE.
  17. de La Gorce, M., D.J. Fleet, and N. Paragios, Model-based 3d hand pose estimation from monocular video. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 2011. 33(9): p. 1793-1805.
  18. Park, M., et al. Hand detection and tracking using depth and color information. in Proceedings of the International Conference on Image Processing, Computer Vision, and Pattern Recognition (IPCV’12). 2012.
  19. Scharstein, D. and R. Szeliski. High-accuracy stereo depth maps using structured light. in Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on. 2003. IEEE.
  20. Horn, B.K. and E. Weldon Jr, Direct methods for recovering motion. International Journal of Computer Vision, 1988. 2(1): p. 51-76.
  21. Henry, P., et al. RGB-D mapping: Using depth cameras for dense 3D modeling of indoor environments. in In the 12th International Symposium on Experimental Robotics (ISER). 2010. Citeseer.

    Sorry, there are no citation(s) for this manuscript yet.