Gesture recognition using depth images
First Claim
1. An apparatus, comprising:
- a face detection engine configured to determine whether a face is present in one or more gray images of respective image frames generated by a depth camera;
a hand tracking engine coupled to the face detection engine, and configured to track a hand in one or more depth images generated by the depth camera, on determination by the face detection engine that a face is present in the one or more gray images; and
a feature extraction and gesture inference engine coupled to the hand tracking engine, and configured to extract features based on results of the tracking by the hand tracking engine, and infer a hand gesture based at least in part on the extract features;
wherein either the face detection engine or the hand tracking engine is further configured to determine a measure of a distance between the face and the camera, using the one or more depth images.
2 Assignments
0 Petitions
Accused Products
Abstract
Methods, apparatuses, and articles associated with gesture recognition using depth images are disclosed herein. In various embodiments, an apparatus may include a face detection engine configured to determine whether a face is present in one or more gray images of respective image frames generated by a depth camera, and a hand tracking engine configured to track a hand in one or more depth images generated by the depth camera. The apparatus may further include a feature extraction and gesture inference engine configured to extract features based on results of the tracking by the hand tracking engine, and infer a hand gesture based at least in part on the extracted features. Other embodiments may also be disclosed and claimed.
15 Citations
26 Claims
-
1. An apparatus, comprising:
-
a face detection engine configured to determine whether a face is present in one or more gray images of respective image frames generated by a depth camera; a hand tracking engine coupled to the face detection engine, and configured to track a hand in one or more depth images generated by the depth camera, on determination by the face detection engine that a face is present in the one or more gray images; and a feature extraction and gesture inference engine coupled to the hand tracking engine, and configured to extract features based on results of the tracking by the hand tracking engine, and infer a hand gesture based at least in part on the extract features; wherein either the face detection engine or the hand tracking engine is further configured to determine a measure of a distance between the face and the camera, using the one or more depth images. - View Dependent Claims (2, 4, 5, 6, 7, 8, 9, 10, 11)
-
-
3. A method comprising:
-
determining by a computing apparatus, whether a face is present in one or more gray images of respective image frames generated by a depth camera; tracking, by the computing apparatus, a hand in selected respective regions of one or more depth images generated by the depth camera, on determination that a face is present in the one or more gray images, wherein the selected respective regions are size-wise smaller than the respective one or more depth images; and inferring a hand gesture, by the computing device, based at least in part on a result of the tracking; wherein tracking comprises determining location measures of the hand for the depth images. - View Dependent Claims (12, 13, 14, 15)
-
-
16. A method comprising:
-
determining by a computing apparatus, whether a face is present in one or more gray images of respective image frames generated by a depth camera; tracking, by the computing apparatus, a hand in selected respective regions of one or more depth images generated by the depth camera, on determination that a face is present in the one or more gray images, wherein the selected respective regions are size-wise smaller than the respective one or more depth images; extracting, by the computing apparatus, one or more features from respective regions of the depth images; and inferring a hand gesture, by the computing apparatus, based at least in part on the one or more features extracted from the depth images; wherein extracting one or more features comprises extracting one or more of an eccentricity measure, a compactness measure, an orientation measure, a rectangularity measure, a horizontal center measure, a vertical center measure, a minimum bounding box angle measure, a minimum bounding box width-to-height ratio measure, a difference between left-and-right measure, or a difference between up-and-down measure. - View Dependent Claims (17, 18)
-
-
19. A computer-readable non-transitory storage medium, comprising:
-
a plurality of programming instructions stored in the storage medium, and configured to cause an apparatus, in response to execution of the programming instructions by the apparatus, to perform operations including; determining whether a face is present in one or more gray images of respective image frames generated by a depth camera; tracking a hand in selected respective regions of one or more depth images generated by the depth camera, on determination that a face is present in the one or more gray images, wherein the selected respective regions are size-wise smaller than the respective one or more depth images; and inferring a hand gesture, based at least in part on a result of the tracking; wherein tracking comprises determining location measures of the hand for the depth images. - View Dependent Claims (20, 21, 22, 23, 24, 25, 26)
-
Specification