SYSTEMS AND METHODS FOR THE AUTONOMOUS PRODUCTION OF VIDEOS FROM MULTI-SENSORED DATA
2 Assignments
0 Petitions
Accused Products
Abstract
An autonomous computer based method and system is described for personalized production of videos such as team sport videos such as basketball videos from multi-sensored data under limited display resolution. Embodiments of the present invention relate to the selection of a view to display from among the multiple video streams captured by the camera network. Technical solutions are provided to provide perceptual comfort as well as an efficient integration of contextual information, which is implemented, for example, by smoothing generated viewpoint/camera sequences to alleviate flickering visual artefacts and discontinuous story-telling artefacts. A design and implementation of the viewpoint selection process is disclosed that has been verified by experiments, which shows that the method and system of the present invention efficiently distribute the processing load across cameras, and effectively selects viewpoints that cover the team action at hand while avoiding major perceptual artefacts.
-
Citations
63 Claims
-
1-43. -43. (canceled)
-
44. A computer based method for autonomous production of an edited video from multiple video streams captured by a plurality of cameras distributed around a scene of interest, the method comprising:
-
detecting objects in the images of the video streams, selecting for each camera, a field of view based on joint processing of positions of multiple objects that have been detected, building the edited video by selecting and concatenating video segments provided by one or more individual cameras, wherein the building is done in a way that maximizes completeness and closeness metrics along the time, while smoothing out the sequence of rendering parameters associated to concatenated segments, wherein completeness measures to which extent objects-of-interest are included and visible within the displayed viewpoint, and closeness refers to the fineness of details when rendering the objects-of-interest, balance completeness and closeness, as a function of individual user preferences. - View Dependent Claims (45, 46, 47, 48, 49, 50, 51, 63)
-
-
52. Computer based system comprising a processing engine and memory for autonomous production of an edited video from multiple video streams captured by a plurality of cameras distributed around a scene of interest, the system comprising:
-
detector for detecting objects in the images of the video streams, first means for selecting one or more camera viewpoints based on joint processing of positions of multiple objects that have been detected, second means for selecting rendering parameters that maximize and smooth out closeness and completeness metrics by concatenating segments in the video streams provided by one or more individual cameras, wherein the building is done in a way that maximizes completeness and closeness metrics along the time, while smoothing out the sequence of rendering parameters associated to concatenated segments, wherein completeness measures to which extent objects-of-interest are included and visible within the displayed viewpoint, and closeness refers to the fineness of details when rendering the objects-of-interest, balance completeness and closeness, as a function of individual user preferences. - View Dependent Claims (53, 54, 55, 56, 57, 58, 59, 60, 61, 62)
-
Specification