Background modification in video conferencing
First Claim
1. A computer-implemented method for real-time video processing, the method comprising:
- receiving a video including a sequence of images;
identifying at least one object of interest in one or more of the images;
detecting at least one shape unit, at least one action unit, and a position vector of the at least one object of interest, the at least one shape unit representing a parameter of a face of the at least one object of interest, the at least one action unit representing a facial mimic, and the position vector corresponding to a rotation around three axes and a translation along the axes;
generating a virtual face mesh from the at least one shape unit, the at least one action unit, and the position vector, the at least one shape unit controlling a shape of the virtual face mesh, the at least one action unit contributing to the shape of the virtual face mesh;
tracking the at least one object of interest in the video, wherein the tracking comprises aligning the virtual face mesh to the at least one object of interest in one or more of the images based at least in part on one or more of the at least one shape unit, the at least one action unit, and the position vector;
identifying a background in each of the images by separating the at least one object of interest from each image based at least in part on the virtual face mesh;
modifying the background in each of the images, thereby generating a modified background; and
generating a modified video which includes the at least one object of interest and the modified background.
4 Assignments
0 Petitions
Accused Products
Abstract
Methods and systems for real-time video processing can be used in video conferencing to modify image quality of background. One example method includes the steps of receiving a video including a sequence of images, identifying at least one object of interest (e.g., a face) in one or more of the images, detecting feature reference points of the at least one object of interest, and tracking the at least one object of interest in the video. The tracking may comprise aligning a virtual face mesh to the at least one object of interest in one or more of the images. Further, a background is identified in the images by separating the at least one object of interest from each image based on the virtual face mesh. The background is then modified in each of the images by blurring, changing a resolution, colors, or other parameters.
-
Citations
30 Claims
-
1. A computer-implemented method for real-time video processing, the method comprising:
-
receiving a video including a sequence of images; identifying at least one object of interest in one or more of the images; detecting at least one shape unit, at least one action unit, and a position vector of the at least one object of interest, the at least one shape unit representing a parameter of a face of the at least one object of interest, the at least one action unit representing a facial mimic, and the position vector corresponding to a rotation around three axes and a translation along the axes; generating a virtual face mesh from the at least one shape unit, the at least one action unit, and the position vector, the at least one shape unit controlling a shape of the virtual face mesh, the at least one action unit contributing to the shape of the virtual face mesh; tracking the at least one object of interest in the video, wherein the tracking comprises aligning the virtual face mesh to the at least one object of interest in one or more of the images based at least in part on one or more of the at least one shape unit, the at least one action unit, and the position vector; identifying a background in each of the images by separating the at least one object of interest from each image based at least in part on the virtual face mesh; modifying the background in each of the images, thereby generating a modified background; and generating a modified video which includes the at least one object of interest and the modified background. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28)
-
-
29. A system, comprising:
a computing device including at least one processor and a memory storing processor-executable codes, which, when implemented by the at least one processor, cause to perform the steps of; receiving a video including a sequence of images; identifying at least one object of interest in one or more of the images; detecting at least one shape unit, at least one action unit, and a position vector of the at least one object of interest, the at least one shape unit representing a parameter of a face of the at least one object of interest, the at least one action unit representing a facial mimic, and the position vector corresponding to a rotation around three axes and a translation along the axes; generating a virtual face mesh from the at least one shape unit, the at least one action unit, and the position vector, the at least one shape unit controlling a shape of the virtual face mesh, the at least one action unit contributing to the shape of the virtual face mesh; tracking the at least one object of interest in the video, wherein the tracking comprises aligning the virtual face mesh to the at least one object of interest in one or more of the images based at least in part on one or more of the at least one shape unit, the at least one action unit, and the position vector; identifying a background in each of the images by separating the at least one object of interest from each image based at least in part on the virtual face mesh; modifying the background in each of the images, thereby generating a modified background; and generating a modified video which includes the at least one object of interest and the modified background.
-
30. A non-transitory processor-readable medium having instructions stored thereon, which when executed by one or more processors, cause the one or more processors to implement a method, comprising:
-
receiving a video including a sequence of images; identifying at least one object of interest in one or more of the images; detecting at least one shape unit, at least one action unit, and a position vector of the at least one object of interest, the at least one shape unit representing a parameter of a face of the at least one object of interest, the at least one action unit representing a facial mimic, and the position vector corresponding to a rotation around three axes and a translation along the axes; generating a virtual face mesh from the at least one shape unit, the at least one action unit, and the position vector, the at least one shape unit controlling a shape of the virtual face mesh, the at least one action unit contributing to the shape of the virtual face mesh; tracking the at least one object of interest in the video, wherein the tracking comprises aligning the virtual face mesh to the at least one object of interest in one or more of the images based at least in part on one or more of the at least one shape unit, the at least one action unit, and the position vector; identifying a background in each of the images by separating the at least one object of interest from each image based at least in part on the virtual face mesh; modifying the background in each of the images, thereby generating a modified background; and generating a modified video which includes the at least one object of interest and the modified background.
-
Specification