Real time eye tracking for human computer interaction
First Claim
1. A method for gaze direction estimation, comprising:
- detecting at least one iris of an eye in an image;
estimating a 3D eyeball center and radius, a fovea position, and an iris center;
determining an estimated eye gaze directional vector between the fovea position and the iris center, based on at least the estimated 3D eyeball center and radius, fovea position, and iris center;
calibrating the estimated eye gaze directional vector with respect to a known condition, by directing the eye to look into at least one camera and toward at least one calibration point;
solving for ta using the following quartic equations;
Sta4+Tta3+Uta2+Vta+W=0,
S=Y2,
T=−
2TZ−
2(A*B)XY,
U=Z2+2(A*B)XY−
r2y2+x2,
V2r2YZ,
W=−
r2Z2,
X=(A×
B×
A)*P,
Y=(A×
B×
A)*B,
Z=(A×
B×
P)*B;
wherein;
r is a 3D eyeball radius,E is an estimate for the 3D eyeball center,E′
is an estimate for the 3D iris center,P is a calibration point having a known 3D location,m1 and m2 are iris centers represented in the image,A and B are normalized vectors between m1 and m2 and the camera,ta and tb represent lengths such that Ata=E and Btb=E′
,wherein ∥
Ata−
Btb∥
=r and point E′
(or Btb) is a point of intersection between C=Ata−
P and B, equivalent to a point of intersection between B and a plane determined by point P and normal N=(A×
B×
C), such that
2 Assignments
0 Petitions
Accused Products
Abstract
A gaze direction determining system and method is provided. A two-camera system may detect the face from a fixed, wide-angle camera, estimates a rough location for the eye region using an eye detector based on topographic features, and directs another active pan-tilt-zoom camera to focus in on this eye region. A eye gaze estimation approach employs point-of-regard (PoG) tracking on a large viewing screen. To allow for greater head pose freedom, a calibration approach is provided to find the 3D eyeball location, eyeball radius, and fovea position. Both the iris center and iris contour points are mapped to the eyeball sphere (creating a 3D iris disk) to get the optical axis; then the fovea rotated accordingly and the final, visual axis gaze direction computed.
-
Citations
29 Claims
-
1. A method for gaze direction estimation, comprising:
-
detecting at least one iris of an eye in an image; estimating a 3D eyeball center and radius, a fovea position, and an iris center; determining an estimated eye gaze directional vector between the fovea position and the iris center, based on at least the estimated 3D eyeball center and radius, fovea position, and iris center; calibrating the estimated eye gaze directional vector with respect to a known condition, by directing the eye to look into at least one camera and toward at least one calibration point; solving for ta using the following quartic equations;
Sta4+Tta3+Uta2+Vta+W=0,
S=Y2,
T=−
2TZ−
2(A*B)XY,
U=Z2+2(A*B)XY−
r2y2+x2,
V2r2YZ,
W=−
r2Z2,
X=(A×
B×
A)*P,
Y=(A×
B×
A)*B,
Z=(A×
B×
P)*B;
wherein; r is a 3D eyeball radius, E is an estimate for the 3D eyeball center, E′
is an estimate for the 3D iris center,P is a calibration point having a known 3D location, m1 and m2 are iris centers represented in the image, A and B are normalized vectors between m1 and m2 and the camera, ta and tb represent lengths such that Ata=E and Btb=E′
,wherein ∥
Ata−
Btb∥
=r and point E′
(or Btb) is a point of intersection between C=Ata−
P and B, equivalent to a point of intersection between B and a plane determined by point P and normal N=(A×
B×
C), such that - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12)
-
-
13. An apparatus for determining an eye gaze direction, comprising:
-
a camera, a display screen; and an automated processor, configured to; detect at least one iris of an eye in an image from the camera, to estimate a 3D eyeball center and radius, a fovea position, and an iris center; estimate an eye gaze directional vector between the fovea position and the iris center; calibrate the eye gaze directional vector estimate of the eye by acquiring at least one image each of the eye directed looking into at least one camera and of the eye directed looking toward at least one calibration point; solve for ta using the following quartic equations;
Sta4+Tta3+Uta2+Vta+W=0,
S=Y2,
T=−
2TZ−
2(A*B)XY,
U=Z2+2(A*B)XY−
r2y2+x2,
V=2r2YZ,
W=−
r2Z2,
X=(A×
B×
A)*P,
Y=(A×
B×
A)*B,
Z=(A×
B×
P)*B;
wherein; r is a 3D eyeball radius, E is an estimate for the 3D eyeball center, E′
is an estimate for the 3D iris center,P is a calibration point having a known 3D location, m1 and m2 are iris centers represented in the image, A and B are normalized vectors between m1 and m2 and the camera, ta and tb represent lengths such that Ata=E and Btb=E′
,wherein ∥
Ata−
Btb∥
=r and point E′
(or Btb) is a point of intersection between C=Ata−
P and B, equivalent to a point of intersection between B and a plane determined by point P and normal N=(A×
B×
C), such that - View Dependent Claims (14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25)
-
-
26. A method for gaze direction estimation, comprising:
-
detecting at least one iris of an eye in an image; estimating a 3D eyeball center and radius, a fovea position, and an iris center; determining an estimated eye gaze directional vector between the fovea position and the iris center, based on at least the estimated 3D eyeball center and radius, the estimated fovea position, and the estimated iris center, wherein the 3D eyeball center and radius, the fovea position, and the iris center upon which the estimated eye gaze directional vector is based are calibrated by imaging the eye while directing the eye to look into at least one camera and toward at least one calibration point; solving for ta using the following quartic equations;
Sta4+Tta3+Uta2+Vta+W=0,
S=Y2,
T=−
2TZ−
2(A*B)XY,
U=Z2+2(A*B)XY−
r2y+x2,
V=2r2YZ,
W=−
r2Z2,
X=(A×
B×
A)*P,
Y=(A×
B×
A)*B,
Z=(A×
B×
P)*B;
wherein; r is a 3D eyeball radius, E is an estimate for the 3D eyeball center, E′
is an estimate for the 3D iris center,P is a calibration point having a known 3D location, m1 and m2 are iris centers represented in the image, A and B are normalized vectors between m1 and m2 and the camera, ta and tb represent lengths such that Ata=E and Btb=E′
,wherein ∥
Ata−
Btb∥
=r and point E′
(or Btb) is a point of intersection between C=Ata−
P and B, equivalent to a point of intersection between B and a plane determined by point P and normal N=(A×
B×
C), such that - View Dependent Claims (27, 28, 29)
-
Specification