Transforming video data in accordance with human visual system feedback metrics
First Claim
1. A method comprising:
- presenting a user interface that receives input data defining one or more user preferences;
configuring a video processing device to generate three-dimensional (3D) video data in accordance with parameters defining capabilities supported by a 3D display device;
transforming, with the configured video processing device, video data to generate the 3D video data;
forwarding the 3D video data to the 3D display device;
while transforming the video data, automatically determining one or more metrics using a Human Visual System (HVS) model that reflects a quality of 3D visualization of the generated 3D video data with respect to a human visual system,wherein automatically determining one or more metrics comprises;
communicating with the 3D display device to receive one or more of the transformed 3D video data and decoded 3D video data from the 3D display device, wherein the decoded 3D video data comprises a decoded version of the transformed 3D video data forwarded by the device to the 3D display device; and
analyzing one or more of the transformed 3D video data and the decoded 3D video data to determine one or more qualitative metrics using the HVS model, wherein the qualitative metrics describe the quality of 3D visualization of the generated 3D video data with respect to the human visual system,wherein the qualitative metrics comprise visual discomfort measured in term of at least one of a binocular geometric asymmetry, a binocular filter asymmetry, excessive binocular disparity, fast motion in depth, a 3D artifact, and an amount and consistency of blurriness,wherein the binocular geometric asymmetry includes one or more of different lens distortion, different focal length that generates different magnification, and un-rectified images,wherein the binocular filter asymmetry includes one or more of a different luminance, a different contrast, a color difference, a cross-talk, and a different quality related to compression artifacts, andwherein the 3D artifact comprises spatial and temporal inconsistencies in depth cues; and
while transforming the video data, reconfiguring the video processing device to transform the video data into the 3D video data so as to generate the 3D video data based on the one or more determined metrics and the defined user preferences.
1 Assignment
0 Petitions
Accused Products
Abstract
In general, techniques are described for transforming video data in accordance with human visual system feedback metrics. For example, an apparatus comprising a transformation module, a parameter discovery module and a human visual system (HVS) feedback module implements these techniques. The parameter discovery module configures the transformation module to generate three-dimensional (3D) video data in accordance with parameters defining capabilities supported by a 3D display device. The transformation module transforms video data to generate the 3D video data. The HVS feedback module determines, while the transformation module transforms the video data, one or more metrics using an HVS model that reflects a quality of 3D visualization of the generated 3D video data with respect to a human visual system and reconfigures the one or more modules based on the determined one or more metrics to refine the generation of the 3D video data.
28 Citations
36 Claims
-
1. A method comprising:
-
presenting a user interface that receives input data defining one or more user preferences; configuring a video processing device to generate three-dimensional (3D) video data in accordance with parameters defining capabilities supported by a 3D display device; transforming, with the configured video processing device, video data to generate the 3D video data; forwarding the 3D video data to the 3D display device; while transforming the video data, automatically determining one or more metrics using a Human Visual System (HVS) model that reflects a quality of 3D visualization of the generated 3D video data with respect to a human visual system, wherein automatically determining one or more metrics comprises; communicating with the 3D display device to receive one or more of the transformed 3D video data and decoded 3D video data from the 3D display device, wherein the decoded 3D video data comprises a decoded version of the transformed 3D video data forwarded by the device to the 3D display device; and analyzing one or more of the transformed 3D video data and the decoded 3D video data to determine one or more qualitative metrics using the HVS model, wherein the qualitative metrics describe the quality of 3D visualization of the generated 3D video data with respect to the human visual system, wherein the qualitative metrics comprise visual discomfort measured in term of at least one of a binocular geometric asymmetry, a binocular filter asymmetry, excessive binocular disparity, fast motion in depth, a 3D artifact, and an amount and consistency of blurriness, wherein the binocular geometric asymmetry includes one or more of different lens distortion, different focal length that generates different magnification, and un-rectified images, wherein the binocular filter asymmetry includes one or more of a different luminance, a different contrast, a color difference, a cross-talk, and a different quality related to compression artifacts, and wherein the 3D artifact comprises spatial and temporal inconsistencies in depth cues; and while transforming the video data, reconfiguring the video processing device to transform the video data into the 3D video data so as to generate the 3D video data based on the one or more determined metrics and the defined user preferences. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. An apparatus comprising:
-
a user interface module that presents a user interface to receive input data defining one or more user preferences; a transformation module; a parameter discovery module that configures the transformation module to generate three-dimensional (3D) video data in accordance with parameters defining capabilities supported by a 3D display device, wherein the transformation module transforms video data to generate the 3D video data; at least one interface that forwards the 3D video data to the 3D display device; and a human visual system (HVS) feedback module that automatically determines, while the transformation module transforms the video data, one or more metrics using an HVS model that reflects a quality of 3D visualization of the generated 3D video data with respect to a human visual system and, while the transformation module transforms the video data, reconfigures the transformation module to transform the video data into the 3D video data so as to generate the 3D video data based on the one or more determined metrics and the defined user preferences, wherein the HVS feedback module communicates with the 3D display device to receive one or more of the transformed 3D video data and decoded 3D video data from the 3D display device, wherein the decoded 3D video data comprises a decoded version of the transformed 3D video data forwarded by the device to the 3D display device, and analyzes one or more of the transformed 3D video data and the decoded 3D video data to automatically determine one or more qualitative metrics using the HVS model, wherein the qualitative metrics describe the quality of 3D visualization of the generated 3D video data with respect to the human visual system, wherein the qualitative metrics comprise visual discomfort measured in term of at least one of a binocular geometric asymmetry, a binocular filter asymmetry, excessive binocular disparity, fast motion in depth, a 3D artifact, and an amount and consistency of blurriness, wherein the binocular geometric asymmetry includes one or more of different lens distortion, different focal length that generates different magnification, and un-rectified images, wherein the binocular filter asymmetry includes one or more of a different luminance, a different contrast, a color difference, a cross-talk, and a different quality related to compression artifacts, and wherein the 3D artifact comprises spatial and temporal inconsistencies in depth cues. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17, 18)
-
-
19. A non-transitory computer-readable storage medium comprising instructions that cause a processor to:
-
present a user interface that receives input data defining one or more user preferences; configure a video processing device to generate three-dimensional (3D) video data in accordance with parameters defining capabilities supported by a 3D display device; transform, with the configured video processing device, video data to generate the 3D video data; forward the 3D video data to the 3D display device; while transforming the video data, automatically determine one or more metrics using a Human Visual System (HVS) model that reflects a quality of 3D visualization of the generated 3D video data with respect to a human visual system; communicate with the 3D display device to receive one or more of the transformed 3D video data and decoded 3D video data from the 3D display device, wherein the decoded 3D video data comprises a decoded version of the transformed 3D video data forwarded by the device to the 3D display device; analyze one or more of the transformed 3D video data and the decoded 3D video data to automatically determine one or more qualitative metrics using the HVS model, wherein the qualitative metrics describe the quality of 3D visualization of the generated 3D video data with respect to the human visual system, wherein the qualitative metrics comprise visual discomfort measured in term of at least one of a binocular geometric asymmetry, a binocular filter asymmetry, excessive binocular disparity, fast motion in depth, a 3D artifact, and an amount and consistency of blurriness, wherein the binocular geometric asymmetry includes one or more of different lens distortion, different focal length that generates different magnification, and un-rectified images, wherein the binocular filter asymmetry includes one or more of a different luminance, a different contrast, a color difference, a cross-talk, and a different quality related to compression artifacts, and wherein the 3D artifact comprises spatial and temporal inconsistencies in depth cues; and while transforming the video data, reconfigure the video processing device to transform the video data into the 3D video data so as to generate the 3D video data based on the one or more determined metrics and the defined user preferences. - View Dependent Claims (20, 21, 22, 23, 24, 25, 26, 27)
-
-
28. An apparatus comprising:
-
means for presenting a user interface that receives input data defining one or more user preferences; means for configuring a video processing device to generate three-dimensional (3D) video data in accordance with parameters defining capabilities supported by a 3D display device; means for transforming, with the configured video processing device, video data to generate the 3D video data; means for forwarding the 3D video data to the 3D display device; means for automatically determining, while transforming the video data, one or more metrics using a Human Visual System (HVS) model that reflects a quality of 3D visualization of the generated 3D video data with respect to a human visual system, wherein the means for automatically determining one or more metrics comprises; means for communicating with the 3D display device to receive one or more of the transformed 3D video data and decoded 3D video data from the 3D display device, wherein the decoded 3D video data comprises a decoded version of the transformed 3D video data forwarded by the device to the 3D display device; and means for analyzing one or more of the transformed 3D video data and the decoded 3D video data to determine one or more qualitative metrics using the HVS model, wherein the qualitative metrics describe the quality of 3D visualization of the generated 3D video data with respect to the human visual system, wherein the qualitative metrics comprise visual discomfort measured in term of at least one of a binocular geometric asymmetry, a binocular filter asymmetry, excessive binocular disparity, fast motion in depth, a 3D artifact, and an amount and consistency of blurriness, wherein the binocular geometric asymmetry includes one or more of different lens distortion, different focal length that generates different magnification, and un-rectified images, wherein the binocular filter asymmetry includes one or more of a different luminance, a different contrast, a color difference, a cross-talk, and a different quality related to compression artifacts, and wherein the 3D artifact comprises spatial and temporal inconsistencies in depth cues; and means for reconfiguring, while transforming the video data, the one or more modules based on the determined one or more metrics and the defined user preferences to refine the generation of the 3D video data. - View Dependent Claims (29, 30, 31, 32, 33, 34, 35, 36)
-
Specification