Synergistic face detection and pose estimation with energy-based models
First Claim
1. A computer-implemented method of face detection and pose estimation, the method comprising the following steps:
- training, a convolutional neural network to map facial images to points on a face manifold, parameterized by facial pose, and to map non-facial images to points away from the face manifold; and
simultaneously determining, whether an image is a face from its proximity to the face manifold and an estimate of facial pose of that image from its projection to the face manifold.
1 Assignment
0 Petitions
Accused Products
Abstract
A method for human face detection that detects faces independently of their particular poses and simultaneously estimates those poses. Our method exhibits an immunity to variations in skin color, eyeglasses, facial hair, lighting, scale and facial expressions, and others. In operation, we train a convolutional neural network to map face images to points on a face manifold, and non-face images to points far away from that manifold, wherein that manifold is parameterized by facial pose. Conceptually, we view a pose parameter as a latent variable, which may be inferred through an energy-minimization process. To train systems based upon our inventive method, we derive a new type of discriminative loss function that is tailored to such detection tasks. Our method enables a multi-view detector that can detect faces in a variety of poses, for example, looking left or right (yaw axis), up or down (pitch axis), or tilting left or right (roll axis). Systems employing our method are highly-reliable, run at near real time (5 frames per second on conventional hardware), and is robust against variations in yaw (±90°), roll (±45°), and pitch (±60°).
34 Citations
7 Claims
-
1. A computer-implemented method of face detection and pose estimation, the method comprising the following steps:
-
training, a convolutional neural network to map facial images to points on a face manifold, parameterized by facial pose, and to map non-facial images to points away from the face manifold; and
simultaneously determining, whether an image is a face from its proximity to the face manifold and an estimate of facial pose of that image from its projection to the face manifold. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
Specification