Illumination estimation from a single image
First Claim
1. One or more non-transitory computer-readable media having a plurality of executable instructions embodied thereon, which, when executed by one or more processors, cause the one or more processors to perform a method for estimating illumination of images, the method comprising:
- receiving an input image of an indoor scene, wherein the input image depicts a view of the indoor scene that is less than 360 degrees;
encoding the input image to an intermediate representation; and
utilizing a trained neural network system to generate, from the intermediate representation, a recovery light mask and low-frequency RGB image each representing a panoramic environment encompassing the view depicted in the input image, wherein utilizing the trained neural network system includes;
generating the recovery light mask utilizing a first branch of the neural network system, wherein the recovery light mask indicates a probability of each pixel representing an area within the panoramic environment being a light source, wherein at least some pixels in the recovery light mask represent part of the panoramic environment that is beyond the view depicted in the input image, andgenerating the low-frequency RGB image indicating low-frequency information utilizing a second branch of the neural network system.
2 Assignments
0 Petitions
Accused Products
Abstract
Methods and systems are provided for using a single image of an indoor scene to estimate illumination of an environment that includes the portion captured in the image. A neural network system may be trained to estimate illumination by generating recovery light masks indicating a probability of each pixel within the larger environment being a light source. Additionally, low-frequency RGB images may be generated that indicating low-frequency information for the environment. The neural network system may be trained using training input images that are extracted from known panoramic images. Once trained, the neural network system infers plausible illumination information from a single image to realistically illumination images and objects being manipulated in graphics applications, such as with image compositing, modeling, and reconstruction.
20 Citations
11 Claims
-
1. One or more non-transitory computer-readable media having a plurality of executable instructions embodied thereon, which, when executed by one or more processors, cause the one or more processors to perform a method for estimating illumination of images, the method comprising:
-
receiving an input image of an indoor scene, wherein the input image depicts a view of the indoor scene that is less than 360 degrees; encoding the input image to an intermediate representation; and utilizing a trained neural network system to generate, from the intermediate representation, a recovery light mask and low-frequency RGB image each representing a panoramic environment encompassing the view depicted in the input image, wherein utilizing the trained neural network system includes; generating the recovery light mask utilizing a first branch of the neural network system, wherein the recovery light mask indicates a probability of each pixel representing an area within the panoramic environment being a light source, wherein at least some pixels in the recovery light mask represent part of the panoramic environment that is beyond the view depicted in the input image, and generating the low-frequency RGB image indicating low-frequency information utilizing a second branch of the neural network system. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
-
8. A computer-implemented method for estimating illumination of images, the method comprising:
-
receiving an input image of an indoor scene, wherein the input image depicts a view of the indoor scene that is less than 360 degrees; encoding the input image to an intermediate representation; and utilizing a trained neural network system to generate, from the intermediate representation, a recovery light mask and low-frequency RGB image each representing a panoramic environment encompassing the view depicted in the input image, wherein utilizing the trained neural network system includes; generating the recovery light mask utilizing a first branch of the neural network system, wherein the recovery light mask indicates a probability of each pixel representing an area within the panoramic environment being a light source, and wherein at least some pixels in the recovery light mask represent part of the panoramic environment that is beyond the view depicted in the input image; and generating the low-frequency RGB image indicating low-frequency information utilizing a second branch of the neural network system. - View Dependent Claims (9, 10, 11)
-
Specification