GESTURE BASED USER INTERFACE
First Claim
Patent Images
1. A gesture based user interface, comprising:
- a movement monitor configured to monitor a user'"'"'s hand and to provide a signal based on movements of the hand;
a display; and
a processor configured to move a cursor responsive to the signal from the movement monitor, within a predetermined region on the display, formed of a plurality of sub-regions associated with respective control commands, and to provide the commands to a controlled application responsively to the movements of the hand,wherein the sub-regions include at least one quick-access sub-region, for which the processor provides the command associated with the sub-region responsively to entrance of the cursor into the quick-access sub-region without additional user hand movements, and at least one regular sub-region, for which the processor provides the command associated with the sub-region responsively to identifying a predetermined hand gesture while the cursor is in the sub-region.
3 Assignments
0 Petitions
Accused Products
Abstract
A gesture based user interface includes a movement monitor configured to monitor a user'"'"'s hand and to provide a signal based on movements of the hand. A processor is configured to provide at least one interface state in which a cursor is confined to movement within a single dimension region responsive to the signal from the movement monitor, and to actuate different commands responsive to the signal from the movement monitor and the location of the cursor in the single dimension region.
-
Citations
20 Claims
-
1. A gesture based user interface, comprising:
-
a movement monitor configured to monitor a user'"'"'s hand and to provide a signal based on movements of the hand; a display; and a processor configured to move a cursor responsive to the signal from the movement monitor, within a predetermined region on the display, formed of a plurality of sub-regions associated with respective control commands, and to provide the commands to a controlled application responsively to the movements of the hand, wherein the sub-regions include at least one quick-access sub-region, for which the processor provides the command associated with the sub-region responsively to entrance of the cursor into the quick-access sub-region without additional user hand movements, and at least one regular sub-region, for which the processor provides the command associated with the sub-region responsively to identifying a predetermined hand gesture while the cursor is in the sub-region. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A gesture based user interface, comprising:
-
a movement monitor configured to monitor a user'"'"'s hand and to provide a signal based on movements of the hand; a display; and a processor configured to move a cursor responsive to the signal from the movement monitor, within a predetermined region on the display, formed of a plurality of sub-regions associated with respective control commands, and to provide the commands to a controlled application responsively to the movements of the hand, wherein the processor is configured to ignore downward movements in controlling the cursor. - View Dependent Claims (11)
-
-
12. A method of receiving user input, comprising:
-
identifying movements of the hand of a user; moving a cursor within a predetermined region, formed of a plurality of sub-regions associated with respective control commands, by a processor, responsive to the identified movements; and actuating a control command associated with a sub-region in which the cursor is located, wherein the sub-regions include at least one quick-access sub-region for which the associated command is actuated responsive to entrance of the cursor into the quick-access sub-region without requiring additional user hand movements and at least one regular sub-region, for which the associated command is actuated responsive to identifying a predetermined hand gesture while the cursor is in the sub-region. - View Dependent Claims (13, 14)
-
-
15. A computer software product, comprising a tangible computer-readable medium in which program instructions are stored, which instructions, when read by a computer, cause the computer to:
-
identify movements of a user; move a cursor within a predetermined region, formed of a plurality of sub-regions associated with respective control commands, responsive to the identified movements; and actuate a control command associated with a sub-region in which the cursor is located, wherein the sub-regions include at least one quick-access sub-region for which the associated command is actuated responsive to entrance of the cursor into the quick-access sub-region without requiring additional user hand movements and at least one regular sub-region, for which the associated command is actuated responsive to identifying a predetermined hand gesture while the cursor is in the sub-region.
-
-
16. A gesture based user interface, comprising:
-
a movement monitor configured to monitor a user'"'"'s hand and to provide a signal based on movements of the hand; a display; and a processor configured to provide at least one interface state in which a cursor is confined to movement within a single dimension region responsive to the signal from the movement monitor, and to actuate different commands responsive to the signal from the movement monitor and the location of the cursor in the single dimension region. - View Dependent Claims (17, 18, 19, 20)
-
Specification