Architecture for controlling a computer using hand gestures
First Claim
1. A system that facilitates a user interface, comprising:
- a tracking component that detects at least one of a plurality of objects within a scene and tracks the respective object, detection of the object based at least in palt upon image comparison of a plurality of images relative to a course mapping of the images;
a seeding component that iteratively seeds the tracking component with object hypotheses based upon the presence of the object and the image comparison; and
a filtering component that selectively removes the tracked object from the object hypotheses and/or at least one object hypothesis from the set of object hypotheses, the tracked object removed based at least in part upon a region-based approach in determining depth to cursors and move windows.
3 Assignments
0 Petitions
Accused Products
Abstract
Architecture for implementing a perceptual user interface. The architecture comprises alternative modalities for controlling computer application programs and manipulating on-screen objects through hand gestures or a combination of hand gestures and verbal commands. The perceptual user interface system includes a tracking component that detects object characteristics of at least one of a plurality of objects within a scene, and tracks the respective object. Detection of object characteristics is based at least in part upon image comparison of a plurality of images relative to a course mapping of the images. A seeding component iteratively seeds the tracking component with object hypotheses based upon the presence of the object characteristics and the image comparison. A filtering component selectively removes the tracked object from the object hypotheses and/or at least one object hypothesis from the set of object hypotheses based upon predetermined removal criteria.
444 Citations
70 Claims
-
1. A system that facilitates a user interface, comprising:
-
a tracking component that detects at least one of a plurality of objects within a scene and tracks the respective object, detection of the object based at least in palt upon image comparison of a plurality of images relative to a course mapping of the images;
a seeding component that iteratively seeds the tracking component with object hypotheses based upon the presence of the object and the image comparison; and
a filtering component that selectively removes the tracked object from the object hypotheses and/or at least one object hypothesis from the set of object hypotheses, the tracked object removed based at least in part upon a region-based approach in determining depth to cursors and move windows. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
-
-
17. A system that facilitates a user interface, comprising:
-
means for tracking and detecting at least one of a plurality of objects within a scene based at least in part upon image comparison of a plurality of images relative to a course mapping of the images;
means for iteratively seeding the tracking component with object hypotheses based upon the presence of the object characteristics and the image comparison; and
means for filtering that selectively removes the tracked object from the object hypotheses and/or at least one object hypothesis from the set of object hypotheses, the tracked object removed based at least in part upon a region-based approach in determining depth to cursors and move windows.
-
-
18. A system that facilitates a user interface, comprising:
-
a detecting component that detects at least one of a plurality of objects within a scene;
a tracking component that tracks the detected object;
a seeding component that iteratively seeds the tracking component with object hypotheses based upon the detected objects that are tracked;
a filtering component that selectively removes the tracked object from the object hypotheses or at least one object hypothesis from the set of object hypotheses; and
an interpreting component that interprets an object characteristic of the tracked object and executes a command in response thereto. - View Dependent Claims (19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49)
-
-
50. A computer system that facilitates interaction with a user, comprising:
-
an object processing system for processing object information of one or more objects in a scene, the object processing system including;
a tracking component that at least tracks the one or more objects;
a seeding component that iteratively seeds the tracking component with object hypotheses based upon the objects that are tracked;
a filtering component that selectively removes the tracked object and corresponding object hypothesis from a set of object hypotheses in accordance with predetermined criteria; and
an interpreting component that interprets object characteristics of the tracked object and causes a command to be executed in response thereto;
an input system for receiving user input separately or in combination with the object processing system; and
a presentation system for presenting information to the user in response to at least one of the command being executed and receiving user input via the input system. - View Dependent Claims (51, 52, 53, 54, 55, 56, 57)
-
-
58. A method of facilitating a human-computer interface, comprising:
-
acquiring gesture characteristics of at least one of a plurality of gestures within a volume of space with an acquisition component, the gesture characteristics acquired based at least in part upon image comparison of a plurality of images relative to a course mapping of the images;
iteratively seeding the acquisition component with at least one gesture hypothesis based upon the presence of the gesture characteristics in the volume of space and the image comparison; and
automatically controlling a graphical representation of a graphical interface in response to acquiring the at least one gesture. - View Dependent Claims (59, 60, 61, 62, 63, 64, 65, 66, 67)
-
-
68. A method of facilitating a user interface, comprising:
-
capturing an image of moving objects with a video source;
determining whether one or more of the moving objects exist within the image;
analyzing whether the one or more moving objects are within an engagement volume;
calculating a distance from the video source to the moving objects within the engagement volume;
selecting a closest moving object from one or more moving objects within the engagement volume;
tracking the closest moving object;
evaluating whether the closest moving object remains within the engagement volume;
determining whether the closest moving object remains in motion;
interpreting the motion of the closest moving object;
determining whether the closest moving objects motion is a recognized command; and
executing the recognized command to control a graphical representation of a graphical user interface. - View Dependent Claims (69, 70)
-
Specification