Controlling objects via gesturing
First Claim
1. A system for controlling an object via gesturing comprising:
- at least one camera for receiving light from a device;
a host computer coupled to the at least one camera for processing the received light to produce image data to determine a 3-dimensional location of a gesture, wherein receiving light comprises the at least one camera periodically capturing images associated with the light, and wherein the host computer processes the image data against a coordinate system to determine the 3-dimensional location of the gesture; and
a storage device coupled to the host computer for storing a plurality of prototype sequences, wherein the host computer processes the image data to determine that the gesture has been performed by matching the image data against the plurality of prototype sequences, and wherein if the image data matches one of the prototype sequences a control action corresponding to the gesture is performed on the object, wherein the storage device further comprises a list that includes one or more modified prototype sequences corresponding to each of the plurality of prototype sequences, wherein the one or more modified prototype sequences are different versions of a corresponding prototype sequence.
1 Assignment
0 Petitions
Accused Products
Abstract
The present invention is directed toward a system and process that controls a group of networked electronic components using a multimodal integration scheme in which inputs from a speech recognition subsystem, gesture recognition subsystem employing a wireless pointing device and pointing analysis subsystem also employing the pointing device, are combined to determine what component a user wants to control and what control action is desired. In this multimodal integration scheme, the desired action concerning an electronic component is decomposed into a command and a referent pair. The referent can be identified using the pointing device to identify the component by pointing at the component or an object associated with it, by using speech recognition, or both. The command may be specified by pressing a button on the pointing device, by a gesture performed with the pointing device, by a speech recognition event, or by any combination of these inputs.
368 Citations
24 Claims
-
1. A system for controlling an object via gesturing comprising:
-
at least one camera for receiving light from a device; a host computer coupled to the at least one camera for processing the received light to produce image data to determine a 3-dimensional location of a gesture, wherein receiving light comprises the at least one camera periodically capturing images associated with the light, and wherein the host computer processes the image data against a coordinate system to determine the 3-dimensional location of the gesture; and a storage device coupled to the host computer for storing a plurality of prototype sequences, wherein the host computer processes the image data to determine that the gesture has been performed by matching the image data against the plurality of prototype sequences, and wherein if the image data matches one of the prototype sequences a control action corresponding to the gesture is performed on the object, wherein the storage device further comprises a list that includes one or more modified prototype sequences corresponding to each of the plurality of prototype sequences, wherein the one or more modified prototype sequences are different versions of a corresponding prototype sequence. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A method for controlling an object via gesturing comprising:
-
receiving light with at least one camera; capturing images caused by the light with the at least one camera; producing image data from the light, said image data comprising a user interacting with the object in a scene; generating a list that includes one or more modified prototype sequences for each of a plurality of prototype sequences, wherein the modified prototype sequences are different versions of their corresponding prototype sequence; and matching the image data against the plurality of prototype sequences and their corresponding modified prototype sequences; determining whether the image data matches one of the prototype sequences; determining that a gesture has been performed by the user, if the image data matches the one of the prototype sequences, said gesture comprising a movement performed by the user in a defined direction to interact with the object; and performing a control action on the object corresponding to the gesture. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17)
-
-
18. One or more computer-readable storage media in a computing device having computer-useable instructions embodied thereon for performing a method for controlling an object, the method comprising:
-
receiving infrared light with at least one infrared camera; capturing images associated with the infrared light with the at least one infrared camera; producing infrared image data from the infrared light, said infrared image data comprising a user interacting with the object in a scene; generating a list that includes one or more modified prototype sequences for each of a plurality of prototype sequences, wherein the modified prototype sequences are different versions of their corresponding prototype sequence; and matching the infrared image data against the plurality of prototype sequences and their corresponding modified prototype sequences; determining whether the infrared image data matches one of the prototype sequences; determining that a gesture has been performed by the user, if the infrared image data matches the one of the prototype sequences, said gesture comprising at least one of a movement performed by the user in a defined direction to interact with the object or speech uttered by the user to interact with the object; and performing a control action on the object corresponding to the gesture. - View Dependent Claims (19, 20, 21, 22, 23, 24)
-
Specification