Real-Time Tracking of Human Location and Motion using Cameras in a Ubiquitous Smart Home

  • Published : 2009.02.23


The ubiquitous smart home is the home of the future, which exploits context information from both the human and the home environment, providing an automatic home service for the human. Human location and motion are the most important contexts in the ubiquitous smart home. In this paper, we present a real-time human tracker that predicts human location and motion for the ubiquitous smart home. The system uses four network cameras for real-time human tracking. This paper explains the architecture of the real-time human tracker, and proposes an algorithm for predicting human location and motion. To detect human location, three kinds of images are used: $IMAGE_1$ - empty room image, $IMAGE_2$ - image of furniture and home appliances, $IMAGE_3$ - image of $IMAGE_2$ and the human. The real-time human tracker decides which specific furniture or home appliance the human is associated with, via analysis of three images, and predicts human motion using a support vector machine (SVM). The performance experiment of the human's location, which uses three images, lasted an average of 0.037 seconds. The SVM feature of human motion recognition is decided from the pixel number by the array line of the moving object. We evaluated each motion 1,000 times. The average accuracy of all types of motion was 86.5%.



  1. S. K. Das and D. J. Cook, “Guest Editorial - Smart Homes,” IEEE Wireless Communications, vol. 9, no. 6, pp. 62-62, 2002.
  2. Seiie Jang and Woontack Woo, “A unified context-aware application model,” Springer, Lecture Notes in Computer Science, vol. 2680, pp. 178-189, Aug. 2003.
  3. Jonghwa Choi, Dongkyoo Shin, and Dongil Shin, “Research and implementation of the contextaware middleware for controlling home appliances,” IEEE Transactions on Consumer Electronics, vol. 51, no. 1, pp. 301-306, 2005.
  4. Christopher Wren, Ali Azarbayejani, Trevor Darrell, and Alex Pentland, “Pfinder : Real-time Tracking of the human Body,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 7, pp. 780 -785, 1997.
  5. Masafumi Tominaga, Hitoshi Hongo, Hiroyasu Koshimizu, Yoshinori Niwa, and Kazuhiko Yamamoto, “Estimation of human motion from multiple cameras for gesture recognition,” Pattern Recognition, pp. 401-404, 2002.
  6. Hitoshi Hongo, Hiroki Watanabe, Mamoru Yasumoto, Yoshinori Niwa, and Kazuhiko Yamamoto, “Eye Regions Extraction for Omni-directional Gaze Detection Using Multiple Cameras,” In Proc. IASTED Conference on Signal Processing, Pattern Recognition and Applications (SPPRA2001), pp. 241-246, Jul. 2001.
  7. Aaron Bobick and James Davis, “Real-time recognition of activity using Temporal Templates,” In Proc. Third IEEE Workshop on Application of Computer Vision, pp. 1233-1251, 1996.
  8. Guohui Li, Jun Zhang, Hongwen Lin, Tu D, and Maojun Zhang, “A moving object detection approach using integrated background template for smart video sensor,” In Proc. Instrumentation and Measurement Technology Conference, vol. 1, pp. 462-466, 2004.
  9. John Krumm, Steve Harris, Brian Meyers, Barry Brumitt, Michael Hale, and Steve Shafer, “Multi-camera multi-person tracking for EasyLiving,” In Proc. Third IEEE International Workshop on Visual Surveillance, pp. 3-10, 2003.
  10. Rania Y. Khalaf and Stephen S. Intille, “Improving multiple people tracking using temporal consistency,” Dept, of Architecture House_n Project Technical Report, Massachusetts Institute of Technology, Cambridge, MA, 2001.
  11. M. Tominaga, H. Hongo, H. Koshimizu, Y. Niwa, and K. Yamamoto, “Estimation of human motion from multiple cameras for gesture recognition,” In Proc. 16th International Conference on Pattern Recognition, vol. 1, pp. 401-404, Aug. 2002.
  12. C. J.C. Burges, “A tutorial on support vector machines for pattern recognition,” Data Mining Knowledge, pp. 1-47, Dec. 1998.