GENERATION OF A THREE-DIMENSIONAL REPRESENTATION OF A USER
First Claim
Patent Images
1. A method that facilitates constructing a computer-implemented three-dimensional representation of a head of a user, the method comprising:
- receiving a plurality of RGB frames of the head of the user from a camera, the plurality of RGB frames captured by the camera over a range of time;
receiving a plurality of depth frames from a depth sensor, the depth frames being indicative of distances of respective portions of the head of the user from the depth sensor, the depth frames generated by the depth sensor over the range of time;
identifying at least one feature of the head of the user in the plurality of RGB frames, the at least one feature being one of a center of an eye of the user, a center of a nose of the user, a first nasal alar of the user, or a second nasal alar of the user;
generating a three-dimensional mesh of the head of the user based at least in part upon the plurality of depth frames and the identifying of the at least one feature of the head of the user in the RGB frames; and
texturizing the three-dimensional mesh based at least in part upon the plurality of RGB frames.
3 Assignments
0 Petitions
Accused Products
Abstract
Described herein are technologies pertaining to generating a relatively accurate virtual three-dimensional model of a head/face of a user. Depth frames are received from a depth sensor and color frames are received from a camera, wherein such frames capture a head of a user. Based upon the depth frames and the color frames, the three-dimensional model of the head of the user is generated.
-
Citations
20 Claims
-
1. A method that facilitates constructing a computer-implemented three-dimensional representation of a head of a user, the method comprising:
-
receiving a plurality of RGB frames of the head of the user from a camera, the plurality of RGB frames captured by the camera over a range of time; receiving a plurality of depth frames from a depth sensor, the depth frames being indicative of distances of respective portions of the head of the user from the depth sensor, the depth frames generated by the depth sensor over the range of time; identifying at least one feature of the head of the user in the plurality of RGB frames, the at least one feature being one of a center of an eye of the user, a center of a nose of the user, a first nasal alar of the user, or a second nasal alar of the user; generating a three-dimensional mesh of the head of the user based at least in part upon the plurality of depth frames and the identifying of the at least one feature of the head of the user in the RGB frames; and texturizing the three-dimensional mesh based at least in part upon the plurality of RGB frames. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11)
-
-
12. A system, comprising:
-
a processor; and a memory that comprises a plurality of components that are executed by the processor, the plurality of components comprising; a receiver component that receives; a plurality of RGB frames captured by a RGB camera over a range of time, each RGB frame in the plurality of RGB frames comprising an image of a head of a user; and a plurality of depth frames captured by a depth sensor over the range of time, each depth frame in the plurality of depth frames comprising a depth map of the head of the user; and a model generator component that; identifies facial features of the user in the plurality of RGB frames; aligns depth frames in the plurality of depth frames with one another based at least in part upon the facial features of the user identified in the plurality of RGB frames; and generates an animated three-dimensional model of the head of the user based at least in part upon the depth frames that are in alignment with one another. - View Dependent Claims (13, 14, 15, 16, 17, 18, 19)
-
-
20. A computer-readable medium comprising instructions that, when executed by a processor, cause the processor to perform acts comprising:
-
receiving a plurality of RGB frames from an RGB camera over a range of time, the RGB frames capturing a head of a user; receiving a plurality of depth frames from a depth sensor over the range of time, the depth frames capturing the head of the user; identifying at least one feature of the user in the plurality of RGB frames, the at least one feature being one of a center of an eye of the user, a center of a nose of the user, a first nasal alar of the user, or a second nasal alar of the user; generating a three-dimensional point cloud corresponding to the head of the user based at least in part upon the plurality of depth frames and the identifying of the at least one feature of the user in the plurality of RGB frames; selecting a template head model from a library of template head models based upon the three-dimensional point cloud; refining the head model based at least in part upon the at least one feature of the user in the plurality of RGB frames; and subsequent to refining the head model, texturing the head model based at least in part upon the plurality of RGB frames.
-
Specification