3D geometric modeling and 3D video content creation
First Claim
1. A device for creating a 3D video stream providing 3D motion capture of an object, said device comprising:
- a projector and a sensor having a fixed position in relation to each other;
said projector being configured to project a coded light pattern comprising a repeating structure of a plurality of unique feature types arranged such that each feature type appears at most once along any one of a plurality of distinguishable first epipolar lines; and
said sensor being configured to capture a series of images of reflections, upon the object, of a portion of the coded light pattern, the reflections being oriented at a non-zero angle in relation to second epipolar lines of an epipolar field in each captured image associated with the fixed relative position between said projector and said sensor, anda processor configured to;
i. identify, within the reflections captured in the images, reflections of the feature types projected by said projector;
ii. calculate distances to points on the object based on the identified reflections;
iii. generate points of a series of point clouds using the calculated distances; and
iv. create a video stream providing 3D motion capture of the object based on the generated point cloudswherein said projector is configured to;
i. project the coded light pattern at an angle to the fixed relative position between said projector and said sensor, andii. project the coded light pattern such that;
a. each one of the first epipolar lines has a known location in each captured image and is associated with the fixed relative position between the projector and the sensor, andb. a distance between the second epipolar lines is less than YC, where C is a size of a feature type in each captured image in the shortest dimension of the repeating structure, and Y is a number of feature types in the shortest dimension of the repeating structure.
0 Assignments
0 Petitions
Accused Products
Abstract
A system, apparatus and method of obtaining data from a 2D image in order to determine the 3D shape of objects appearing in said 2D image, said 2D image having distinguishable epipolar lines, said method comprising: (a) providing a predefined set of types of features, giving rise to feature types, each feature type being distinguishable according to a unique bi-dimensional formation; (b) providing a coded light pattern comprising multiple appearances of said feature types; (c) projecting said coded light pattern on said objects such that the distance between epipolar lines associated with substantially identical features is less than the distance between corresponding locations of two neighboring features; (d) capturing a 2D image of said objects having said projected coded light pattern projected thereupon, said 2D image comprising reflected said feature types; and (e) extracting: (i) said reflected feature types according to the unique bi-dimensional formations; and (ii) locations of said reflected feature types on respective epipolar lines in said 2D image.
-
Citations
23 Claims
-
1. A device for creating a 3D video stream providing 3D motion capture of an object, said device comprising:
-
a projector and a sensor having a fixed position in relation to each other; said projector being configured to project a coded light pattern comprising a repeating structure of a plurality of unique feature types arranged such that each feature type appears at most once along any one of a plurality of distinguishable first epipolar lines; and said sensor being configured to capture a series of images of reflections, upon the object, of a portion of the coded light pattern, the reflections being oriented at a non-zero angle in relation to second epipolar lines of an epipolar field in each captured image associated with the fixed relative position between said projector and said sensor, and a processor configured to; i. identify, within the reflections captured in the images, reflections of the feature types projected by said projector; ii. calculate distances to points on the object based on the identified reflections; iii. generate points of a series of point clouds using the calculated distances; and iv. create a video stream providing 3D motion capture of the object based on the generated point clouds wherein said projector is configured to; i. project the coded light pattern at an angle to the fixed relative position between said projector and said sensor, and ii. project the coded light pattern such that; a. each one of the first epipolar lines has a known location in each captured image and is associated with the fixed relative position between the projector and the sensor, and b. a distance between the second epipolar lines is less than YC, where C is a size of a feature type in each captured image in the shortest dimension of the repeating structure, and Y is a number of feature types in the shortest dimension of the repeating structure. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13)
-
-
14. A method for creating a 3D video stream providing 3D motion capture of an object, said method comprising:
-
projecting, using a projector, a coded light pattern comprising a repeating structure of a plurality of unique feature types arranged such that each feature type appears at most once along any one of a plurality of distinguishable first epipolar lines; capturing, using a sensor having a fixed position in relation to the projector, a series of images of reflected reflections, upon the object, of a portion of the coded light pattern, the reflections being oriented at a non-zero angle in relation to second epipolar lines of an epipolar field in each captured image associated with the fixed relative position between said projector and said sensor, using a processor functionally associated with the sensor to; i. identify, within the reflections captured in the images, reflections of the feature types projected by said projector; ii. calculate distances to points on the object based on the identified reflections; iii. generate points of a series of point clouds using the calculated distances; and iv. create a video stream providing 3D motion capture of the object based on the generated point clouds wherein the coded light pattern is projected at an angle to the fixed relative position between said projector and said sensor, and each one of the first epipolar lines has a known location in each captured image and is associated with the fixed relative position between the projector and the sensor, and a distance between the second epipolar lines is less than YC, where C is a size of a feature type in each captured image in the shortest dimension of the repeating structure, and Y is a number of feature types in the shortest dimension of the repeating structure. - View Dependent Claims (15, 16, 17, 18, 19, 20, 21, 22, 23)
-
Specification