Systems and methods for depth-assisted perspective distortion correction
First Claim
Patent Images
1. A camera system, comprising:
- a plurality of cameras configured to capture image data from multiple viewpoints, wherein cameras in the plurality of cameras are situated in various positions corresponding to the multiple viewpoints;
a processor;
a memory containing an image processing application; and
a display;
wherein the image processing application stored in the memory directs the processor to;
obtain image data captured by the plurality of cameras from multiple viewpoints including an initial viewpoint;
generate depth map data indicating distances to faces within a scene from the initial viewpoint using information based on differences among the multiple viewpoints of the image data;
detect a face within the image data and a distance from the initial viewpoint to the face from the depth map data;
segment face image data from background image data using the depth map data;
rerender the face from a synthetic viewpoint by warping the segmented face image data based upon the depth map data to generate warped face image data, where the synthetic viewpoint is a greater distance from the face along an optical axis relative to the distance from the initial viewpoint to the face, and the warping corrects perspective distortion in the segmented face image data resulting from camera optics by;
selecting a desired viewpoint distance that specifies a distance from the synthetic viewpoint to the face;
projecting the segmented face image data to 3D locations based upon distances to pixels within the segmented face image data contained within the depth map data;
re-projecting the 3D locations to new 2D pixel locations based upon the desired viewpoint distance to create warped face image data; and
filling holes in the warped face image data;
combine the warped face image data with the background image data to create perspective distortion corrected image data; and
output the perspective distortion corrected image data to the display.
1 Assignment
0 Petitions
Accused Products
Abstract
Systems and methods for automatically correcting apparent distortions in close range photographs that are captured using an imaging system capable of capturing images and depth maps are disclosed. In many embodiments, faces are automatically detected and segmented from images using a depth-assisted alpha matting. The detected faces can then be re-rendered from a more distant viewpoint and composited with the background to create a new image in which apparent perspective distortion is reduced.
1097 Citations
13 Claims
-
1. A camera system, comprising:
-
a plurality of cameras configured to capture image data from multiple viewpoints, wherein cameras in the plurality of cameras are situated in various positions corresponding to the multiple viewpoints; a processor; a memory containing an image processing application; and a display; wherein the image processing application stored in the memory directs the processor to; obtain image data captured by the plurality of cameras from multiple viewpoints including an initial viewpoint; generate depth map data indicating distances to faces within a scene from the initial viewpoint using information based on differences among the multiple viewpoints of the image data; detect a face within the image data and a distance from the initial viewpoint to the face from the depth map data; segment face image data from background image data using the depth map data; rerender the face from a synthetic viewpoint by warping the segmented face image data based upon the depth map data to generate warped face image data, where the synthetic viewpoint is a greater distance from the face along an optical axis relative to the distance from the initial viewpoint to the face, and the warping corrects perspective distortion in the segmented face image data resulting from camera optics by; selecting a desired viewpoint distance that specifies a distance from the synthetic viewpoint to the face; projecting the segmented face image data to 3D locations based upon distances to pixels within the segmented face image data contained within the depth map data; re-projecting the 3D locations to new 2D pixel locations based upon the desired viewpoint distance to create warped face image data; and filling holes in the warped face image data; combine the warped face image data with the background image data to create perspective distortion corrected image data; and output the perspective distortion corrected image data to the display. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12)
-
-
13. A camera system, comprising:
-
a plurality of cameras configured to capture image data from multiple viewpoints, wherein cameras in the plurality of cameras are situated in various positions corresponding to the multiple viewpoints; a processor; a memory containing an image processing application; and a display; wherein the image processing application stored in the memory directs the processor to; obtain image data captured by the plurality of cameras from multiple viewpoints including an initial viewpoint; generate depth map data indicating distances to faces within a scene from the initial viewpoint using information based on differences among the multiple viewpoints of the image data; detect a face within the image data and a distance from the initial viewpoint to the face from the depth map data; segment face image data from background image data using the depth map data by; identifying a trimap for the image data using the image data and the depth map data; and generating an alpha matte using the trimap; rerender the face from a synthetic viewpoint by warping the segmented face image data based upon the depth map data to generate warped face image data, where the synthetic viewpoint is a greater distance from the face along an optical axis relative to the distance from the initial viewpoint to the face, and the warping corrects perspective distortion in the segmented face image data resulting from camera optics by; selecting a desired viewpoint distance that specifies a distance from the synthetic viewpoint to the face; projecting the segmented face image data to 3D locations based upon distances to pixels within the segmented face image data contained within the depth map data; re-projecting the 3D locations to new 2D pixel locations based upon the desired viewpoint distance to create warped face image data; and filling holes in the warped face image data; inpaint the background image data; combine the warped face image data with the background image data to create perspective distortion corrected image data; and output the perspective distortion corrected image data to the display.
-
Specification