Automatic Video-based Motion Analysis
Small Business Information
MD, Greenbelt, MD, 20770-1423
AbstractOperations in confined, isolated, and resource-constrained environments can lead to suboptimal human performance. Understanding task performance and crew behavioral health is crucial to mission success and the optimal design, development and operation of next generation space craft. Onboard resources, such as a single conventional video camera, can capture crew motion and interaction. There is a critical need for a software tool which achieves unobtrusive, non-invasive, automatic analysis of crew activity from video footage.Many video-based human motion analysis tools assume a stationary camera and employ segmentation techniques like temporal differencing or background segmentation to detect people. However, these approaches are vulnerable to camera motion and subtle changes in the background. In addition, many existing commercial solutions use simple blob-based video analysis where the entire body is tracked as a single object. Employing such a coarse human body model is appropriate for surveillance applications concerned with motion detection and person counting; however, it is insufficient for understanding precise human actions or gestures. Therefore, a system which is able to detect human body pose automatically, regardless of camera setup, is necessary for addressing these issues.Vecna proposes a video analysis software tool that automatically processes and analyzes complex human motions in conventional 2D video without the use of specialized markers. Unlike many video analytics solutions, Vecna's solution goes beyond simple blob-based video analysis by tracking the geometric configuration of human body parts like the trunk, head, and limbs. This enables our human motion understanding algorithms to model and recognize complex human actions and interactions. The resulting system will represent a substantial breakthrough providing benefits to an array of applications in video surveillance, human-computer interaction, human factors engineering, and robotics.
* information listed above is at the time of submission.