Image processing device, image processing method, and a computer-readable non-transitory medium
First Claim
1. An image processing device configured to execute a process in accordance with a type of gesture performed by a head of a user equipped with a camera, the image processing device, comprising:
- a memory configured to store;
feature point information to separately identify at least two objects,user motion association information in which the type of gesture and a combination of a moving direction of an object, a moving speed of the object, and a motion vector corresponding to the object are associated with each other, anddisplay information indicating a relationship between the type of gesture and collateral information indicating detail information of a plurality of objects existing around the user; and
a processor coupled to the memory and configured to;
extract a plurality of feature points positioned within a predetermined peripheral region in an image from a plurality of images captured by the user wearing the camera at different timings,specify an object corresponding to the plurality of feature points by referring to the feature point information,determine whether the object moving from the predetermined peripheral region to a central region positioned in an inside of the predetermined peripheral region and a coordinate variation of the object after reaching the central region remains within a predetermined range,determine that the object is a target object indicating a gaze target of the user, in response to determining that the object is moving from the predetermined peripheral region to the central region and that the coordinate variation of the object remains within the predetermined range,calculate a combination of a moving direction of the target object, a moving speed of the target object, and a motion vector corresponding to the target object, by using a coordinate variation of the target object before reaching the central region, on the plurality of images,extract the type of gesture corresponding to the calculated combination from the user motion association information,extract the collateral information of the target object corresponding to the extracted type of gesture from the display information, anddisplay the extracted collateral information in a manner to be superposed on the target object.
1 Assignment
0 Petitions
Accused Products
Abstract
An image processing method includes recognizing a first object that is included in the image data, calculating a positional variation amount of a feature point that corresponds to the first object and is moving in an image center direction so as to calculate a moving speed of the first object on the basis of the positional variation amount, determining whether or not the first object is a gaze target object of the user, in accordance with a behavior of the first object, the behavior being obtained on the basis of the positional variation amount of the first object among the plurality of image data, of which acquisition time is respectively different and whether or not the object continuingly exists in a second region, the second region being positioned in an inside of the first region and including a center point of the image data, for a predetermined period of time.
-
Citations
13 Claims
-
1. An image processing device configured to execute a process in accordance with a type of gesture performed by a head of a user equipped with a camera, the image processing device, comprising:
-
a memory configured to store; feature point information to separately identify at least two objects, user motion association information in which the type of gesture and a combination of a moving direction of an object, a moving speed of the object, and a motion vector corresponding to the object are associated with each other, and display information indicating a relationship between the type of gesture and collateral information indicating detail information of a plurality of objects existing around the user; and a processor coupled to the memory and configured to; extract a plurality of feature points positioned within a predetermined peripheral region in an image from a plurality of images captured by the user wearing the camera at different timings, specify an object corresponding to the plurality of feature points by referring to the feature point information, determine whether the object moving from the predetermined peripheral region to a central region positioned in an inside of the predetermined peripheral region and a coordinate variation of the object after reaching the central region remains within a predetermined range, determine that the object is a target object indicating a gaze target of the user, in response to determining that the object is moving from the predetermined peripheral region to the central region and that the coordinate variation of the object remains within the predetermined range, calculate a combination of a moving direction of the target object, a moving speed of the target object, and a motion vector corresponding to the target object, by using a coordinate variation of the target object before reaching the central region, on the plurality of images, extract the type of gesture corresponding to the calculated combination from the user motion association information, extract the collateral information of the target object corresponding to the extracted type of gesture from the display information, and display the extracted collateral information in a manner to be superposed on the target object. - View Dependent Claims (2, 3)
-
-
4. An image processing method executed by an image processing device, the image processing method comprising:
-
storing feature point information to separately identify at least two objects; storing user motion association information in which a type of gesture performed by a head of a user and a combination of a moving direction of an object, a moving speed of the object, and a motion vector corresponding to the object are associated with each other; storing display information indicating a relationship between the type of gesture and collateral information indicating detail information of a plurality of objects existing around the user; extracting a plurality of feature points positioned within a predetermined peripheral region in an image from a plurality of images captured by the user wearing a camera at different timings; specifying an object corresponding to the plurality of feature points by referring to the feature point information; determining whether the object moving from the predetermined peripheral region to a central region positioned in an inside of the predetermined peripheral region and a coordinate variation of the object after reaching the central region remains within a predetermined range; determining that the object is a target object indicating a gaze target of the user, in response to determining that the object is moving from the predetermined peripheral region to the central region and that the coordinate variation of the object remains within the predetermined range; calculating a combination of a moving direction of the target object, a moving speed of the target object, and a motion vector corresponding to the target object, by using a coordinate variation of the target object before reaching the central region, on the plurality of images; extracting the type of gesture corresponding to the calculated combination from the user motion association information; extracting the collateral information of the target object corresponding to the extracted type of gesture from the display information; and displaying the extracted collateral information in a manner to be superposed on the target object. - View Dependent Claims (5, 6, 7, 8, 9, 10, 11, 12, 13)
-
Specification