Method for driving virtual facial expressions by automatically detecting facial expressions of a face image
First Claim
1. A method for driving virtual facial expressions by automatically detecting facial expressions of a face image, which is applied to a digital image capturing device, comprising:
- learning and correcting a plurality of front face image samples to obtain an average position of each key point of a face, eyes, a nose and a mouth of each of said front face image samples on a standard image for imitating a virtual front face, wherein said standard image is a greyscale figure having a fixed size and using a front face as a center without an inclination;
using a Gabor wavelet algorithm to sample a series of Gabor Jets from said key points of said front face image samples to form a Gabor Jet Bunch;
automatically detecting positions of a face, eyes, a nose and a mouth in a target image captured by said digital image capturing device, and converting said positions onto said standard image;
performing a fitting or regression calculation for said positions of eyes, nose and mouth of said target image and said average positions corresponding thereto to obtain initial positions of key points of eyes, nose and mouth of said target image on said standard image;
calculating exact positions of said key points of said target image on said standard image by using a point within a neighborhood of each of said initial positions as a selecting point, comparing a Gabor Jet of each of said initial positions with each of said Gabor Jets in said Gabor Jet Bunch, selecting said Gabor Jet in said Gabor Jet Bunch having a highest similarity with said selecting point as said exact position of said key point on said standard image corresponding to said key point of said target image, inversely aligning said exact positions of said key points on said standard image onto said target image, and labeling said exact positions as exact positions of said key points on said target image; and
automatically tracking positions of key points of other target image captured by said digital image capturing device later, and correcting said exact position of said key point on said standard image through obtaining a motion parameter (dx, dy) between said key points corresponding to said target image and said other target image by using an optical flow technique, calculating an error ε
(dx, dy) of said motion parameter (dx, dy) according to the following formula, wherein I (x,y) represents gray scales of said target image, J(x,y) represents gray scales of said other target image, and x, y represent coordinates of each of said key points of said target image or said other target image;
ε
(d)=ε
(dx,dy)=Σ
Σ
(I(x,y)−
J(x+dx,y+dy))2,to find dx, dy that minimize said error ε
(dx, dy) and obtain an estimated position of said key point of said other target image based on said position of said key point of said target image corresponding thereto, and performing a fitting or regression calculation for said estimated positions and said average positions to calculate estimated positions of key points of said other target image on said standard image.
5 Assignments
0 Petitions
Accused Products
Abstract
A method for driving virtual facial expressions by automatically detecting facial expressions of a face image is applied to a digital image capturing device. The method includes the steps of detecting a face image captured by the image capturing device and images of a plurality of facial features with different facial expressions to obtain a key point position of each facial feature on the face image; mapping the key point positions to a virtual face as the key point positions of corresponding facial features on the virtual face; dynamically tracking the key point of each facial feature on the face image; estimating the key point positions of each facial feature of the current face image according to the key point positions of each facial feature on a previous face image; and correcting the key point positions of the corresponding facial features on the virtual face.
44 Citations
17 Claims
-
1. A method for driving virtual facial expressions by automatically detecting facial expressions of a face image, which is applied to a digital image capturing device, comprising:
-
learning and correcting a plurality of front face image samples to obtain an average position of each key point of a face, eyes, a nose and a mouth of each of said front face image samples on a standard image for imitating a virtual front face, wherein said standard image is a greyscale figure having a fixed size and using a front face as a center without an inclination; using a Gabor wavelet algorithm to sample a series of Gabor Jets from said key points of said front face image samples to form a Gabor Jet Bunch; automatically detecting positions of a face, eyes, a nose and a mouth in a target image captured by said digital image capturing device, and converting said positions onto said standard image; performing a fitting or regression calculation for said positions of eyes, nose and mouth of said target image and said average positions corresponding thereto to obtain initial positions of key points of eyes, nose and mouth of said target image on said standard image; calculating exact positions of said key points of said target image on said standard image by using a point within a neighborhood of each of said initial positions as a selecting point, comparing a Gabor Jet of each of said initial positions with each of said Gabor Jets in said Gabor Jet Bunch, selecting said Gabor Jet in said Gabor Jet Bunch having a highest similarity with said selecting point as said exact position of said key point on said standard image corresponding to said key point of said target image, inversely aligning said exact positions of said key points on said standard image onto said target image, and labeling said exact positions as exact positions of said key points on said target image; and automatically tracking positions of key points of other target image captured by said digital image capturing device later, and correcting said exact position of said key point on said standard image through obtaining a motion parameter (dx, dy) between said key points corresponding to said target image and said other target image by using an optical flow technique, calculating an error ε
(dx, dy) of said motion parameter (dx, dy) according to the following formula, wherein I (x,y) represents gray scales of said target image, J(x,y) represents gray scales of said other target image, and x, y represent coordinates of each of said key points of said target image or said other target image;
ε
(d)=ε
(dx,dy)=Σ
Σ
(I(x,y)−
J(x+dx,y+dy))2,to find dx, dy that minimize said error ε
(dx, dy) and obtain an estimated position of said key point of said other target image based on said position of said key point of said target image corresponding thereto, and performing a fitting or regression calculation for said estimated positions and said average positions to calculate estimated positions of key points of said other target image on said standard image.- View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A method for driving virtual facial expressions by automatically detecting facial expressions of a face image, which is applied to a digital image capturing device, comprising:
-
obtaining a key point distribution model through learning from a plurality of front face image samples, wherein said key point distribution model is used to represent a distribution of statistic and positions of each key point of a face, eyes, a nose and a mouth of each of said front face image samples on a standard image for imitating a virtual front face; automatically detecting positions of a face, eyes, a nose and a mouth in a target image captured by said digital image capturing device, and converting said positions onto said standard image; determining whether or not said positions of said key points on said standard image deviate from an average position calculated from said statistics and positions of said key point distribution model;
if yes, then said position of said key point on said standard image is pulled back to an accurate position close to said average position;performing a calculation for said positions of said eyes, nose and mouth of said target image and said distribution of statistics and positions corresponding thereto to obtain initial positions of key points of eyes, nose and mouth of said target image on said standard image; using said initial positions as exact positions of said key points of said target image on said standard image, and inversely aligning said exact positions onto said target image, and labeling said exact positions as exact positions of said key points on said target image; and automatically tracking positions of key points of other target image captured by said digital image capturing device later, and correcting said exact position of said key point on said standard image. - View Dependent Claims (11)
-
-
12. A method for driving virtual facial expressions by automatically detecting facial expressions of a face image, which is applied to a digital image capturing device, comprising:
-
obtaining a key point distribution model through learning from a plurality of front face image samples, wherein said key point distribution model is used to represent a distribution of statistic and positions of each key point of a face, eyes, a nose and a mouth of each of said front face image samples on a standard image for imitating a virtual front face, wherein said standard image is a greyscale figure having a fixed size and using a front face as a center without an inclination; automatically detecting positions of a face, eyes, a nose and a mouth in a target image captured by said digital image capturing device, and converting said positions onto said standard image; determining whether or not said positions of said key points on said standard image deviate from an average position calculated from said statistics and positions of said key point distribution model;
if yes, then said position of said key point on said standard image is pulled back to an accurate position close to said average position;performing a calculation for said positions of said eyes, nose and mouth of said target image and said distribution of statistics and positions corresponding thereto to obtain initial positions of key points of eyes, nose and mouth of said target image on said standard image; using said initial positions as exact positions of said key points of said target image on said standard image, and inversely aligning said exact positions of said key points on said standard image onto said target image, and labeling said exact positions as exact positions of said key points on said target image; and automatically tracking positions of key points of other target image captured by said digital image capturing device later, and correcting said exact position of said key point on said standard image. - View Dependent Claims (13)
-
-
14. A method for driving virtual facial expressions by automatically detecting facial expressions of a face image, which is applied to a digital image capturing device, comprising:
-
learning and correcting a plurality of front face image samples to obtain an average position of each key point of a face, eyes, a nose and a mouth of each of said front face image samples on a standard image for imitating a virtual front face, wherein said standard image is a greyscale figure having a fixed size and using a front face as a center without an inclination; obtaining a key point distribution model through learning from said front face image samples, wherein said key point distribution model is used to represent a distribution of statistic and positions of said key points of said front face image samples on said standard image; automatically detecting positions of a face, eyes, a nose and a mouth in a target image captured by said digital image capturing device, and converting said positions onto said standard image; determining whether or not said positions of said key points on said standard image deviate from an average position calculated from said statistics and positions of said key point distribution model;
if yes, then said position of said key point on said standard image is pulled back to an accurate position close to said average position calculated from said statistics and positions of said key point distribution model;performing a fitting or regression calculation for said positions of eyes, nose and mouth of said target image and said average positions calculated from said statistics and positions of said key point distribution model and corresponding thereto to obtain initial positions of key points of eyes, nose and mouth of said target image on said standard image; using said initial positions as exact positions of said key points of said target image on said standard image, and inversely aligning said exact positions of said key points on said standard image onto said target image, and labeling said exact positions as exact positions of said key points on said target image; and automatically tracking positions of key points of other target image captured by said digital image capturing device later, and correcting said exact position of said key point on said standard image. - View Dependent Claims (15)
-
-
16. A method for driving virtual facial expressions by automatically detecting facial expressions of a face image, which is applied to a digital image capturing device, comprising:
-
learning and correcting a plurality of front face image samples to obtain an average position of each key point of a face, eyes, a nose and a mouth of each of said front face image samples on a standard image for imitating a virtual front face, wherein said standard image is a greyscale figure having a fixed size and using a front face as a center without an inclination; obtaining a key point distribution model through learning from said front face image samples, wherein said key point distribution model is used to represent a distribution of statistic and positions of said key points of said front face image samples on said standard image; using a Gabor wavelet algorithm to sample a series of Gabor Jets from said key points of said front face image samples to form a Gabor Jet Bunch; automatically detecting positions of a face, eyes, a nose and a mouth in a target image captured by said digital image capturing device, and converting said positions onto said standard image; determining whether or not said positions of said key points on said standard image deviate from an average position calculated from said statistics and positions of said key point distribution model;
if yes, then said position of said key point on said standard image is pulled back to an accurate position close to said average position calculated from said statistics and positions of said key point distribution model;performing a fitting or regression calculation for said positions of eyes, nose and mouth of said target image and said average positions calculated from said statistics and positions of said key point distribution model and corresponding thereto to obtain initial positions of key points of eyes, nose and mouth of said target image on said standard image; calculating exact positions of said key points of said target image on said standard image by using a point within a neighborhood of each of said initial positions as a selecting point, comparing a Gabor Jet of each of said initial positions with each of said Gabor Jets in said Gabor Jet Bunch, selecting said Gabor Jet in said Gabor Jet Bunch having a highest similarity with said selecting point as said exact position of said key point on said standard image corresponding to said key point of said target image, inversely aligning said exact positions of said key points on said standard image onto said target image, and labeling said exact positions as exact positions of said key points on said target image; and automatically tracking positions of key points of other target image captured by said digital image capturing device later, and correcting said exact position of said key point on said standard image. - View Dependent Claims (17)
-
Specification