Systems and methods for performing actions in response to user gestures in captured images
First Claim
1. A computer-implemented system comprising:
- an image capture device that captures images;
a memory device that stores instructions; and
at least one processor that executes the instructions to perform operations comprising;
receiving, from the image capture device, at least one image including a gesture made by a user;
analyzing the at least one image to identify the gesture made by the user in the at least one image;
determining, based on the identified gesture, a first action to perform on the at least one image;
determining a selection area for the gesture;
identifying an area of interest in the at least one image based on the determined selection area of the gesture, wherein the area of interest includes non-textual content;
performing the first action on the identified area of interest, wherein the first action comprises;
classifying the non-textual content included in the area of interest into at least one of a plurality of different types of non-textual content into which the non-textual content is classifiable by the computer-implemented system, wherein the computer-implemented system is capable of recognizing each of a face, an object, and a landscape; and
generating a first result that indicates the at least one type of non-textual content into which the non-textual content included in the area of interest was classified;
determining a second action to be performed on the identified area of interest based at least in part on the at least one type of non-textual content into which the non-textual content included in the area of interest was classified; and
performing the second action on the identified area of interest.
2 Assignments
0 Petitions
Accused Products
Abstract
Systems, methods, and computer-readable media are provided for performing actions in response to gestures made by a user in captured images. In accordance with one implementation, a computer-implemented system is provided that includes an image capture device that captures at least one image, a memory device that stores instructions, and at least one processor that executes the instructions stored in the memory device. In some implementations, the processor receives, from the image capture device, at least one image including a gesture made by a user and analyzes the at least one image to identify the gesture made by the user. In some implementations, the processor also determines, based on the identified gesture, one or more actions to perform on the at least one image.
-
Citations
20 Claims
-
1. A computer-implemented system comprising:
-
an image capture device that captures images; a memory device that stores instructions; and at least one processor that executes the instructions to perform operations comprising; receiving, from the image capture device, at least one image including a gesture made by a user; analyzing the at least one image to identify the gesture made by the user in the at least one image; determining, based on the identified gesture, a first action to perform on the at least one image; determining a selection area for the gesture; identifying an area of interest in the at least one image based on the determined selection area of the gesture, wherein the area of interest includes non-textual content; performing the first action on the identified area of interest, wherein the first action comprises; classifying the non-textual content included in the area of interest into at least one of a plurality of different types of non-textual content into which the non-textual content is classifiable by the computer-implemented system, wherein the computer-implemented system is capable of recognizing each of a face, an object, and a landscape; and generating a first result that indicates the at least one type of non-textual content into which the non-textual content included in the area of interest was classified; determining a second action to be performed on the identified area of interest based at least in part on the at least one type of non-textual content into which the non-textual content included in the area of interest was classified; and performing the second action on the identified area of interest. - View Dependent Claims (2, 3, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20)
-
-
4. A non-transitory, computer-readable medium storing instructions, the instructions configured to cause at least one processor to perform operations comprising:
-
receiving at least one image including a gesture made by a user; analyzing the at least one image to identify the gesture made by the user in the at least one image; determining, based on the gesture, a first action to perform on the at least one image; determining a selection area for the gesture; identifying an area of interest in the at least one image based on the determined selection area of the gesture, wherein the area of interest includes non-textual content; and performing the first action on the identified area of interest, wherein the first action comprises; classifying the non-textual content included in the area of interest into at least one of a plurality of different types of non-textual content into which the non-textual content is classifiable by the at least one processor, wherein the instructions are configured to cause the at least one processor to be capable of recognizing each of a face, an object, and a landscape; and generating a first result that indicates the type of non-textual content into which the content included in the area of interest was classified; determining a second action to be performed on the identified area of interest based at least in part on the at least one type of non-textual content into which the non-textual content included in the area of interest was classified; and performing the second action on the identified area of interest. - View Dependent Claims (5, 6)
-
-
7. A method comprising the following operations performed by one or more processors:
-
receiving at least one image including a single gesture made by a user; analyzing the at least one image to identify the single gesture made by the user in the at least one image; determining, based on the single gesture, a first action to perform on the at least one image; determining a selection area indicated by the single gesture, such that both of the first action and the selection area are determined based on the single gesture made by the user in the at least one image; identifying an area of interest in the at least one image based on the determined selection area indicated by the gesture; and performing the first action on the identified area of interest, wherein performing the first action comprises recognizing, by the one or more processors that are capable of recognizing each of a face, an object, or a landscape, at least one of the face, the object, or the landscape within the area of interest. - View Dependent Claims (8, 9)
-
Specification