Immersive remote conferencing
First Claim
Patent Images
1. A system comprising:
- a service configured to receive video information and depth information corresponding to data captured by camera mechanisms of remote participants;
a view generator coupled to the service, the view generator configured to process data corresponding to the video information and depth information to place visible representations of remote participants into a common scene, wherein the common scene is rendered via a first person point of view;
a tracker using position tracking data to re-render the common scene to compensate for parallax as a user viewing the scene moves among different viewing angles; and
an audio output controller to provide spatial audio based upon the position of the user, or based upon a position of a visible representation of a remote participant placed in the common scene, or both.
2 Assignments
0 Petitions
Accused Products
Abstract
The subject disclosure is directed towards an immersive conference, in which participants in separate locations are brought together into a common virtual environment (scene), such that they appear to each other to be in a common space, with geometry, appearance, and real-time natural interaction (e.g., gestures) preserved. In one aspect, depth data and video data are processed to place remote participants in the common scene from the first person point of view of a local participant. Sound data may be spatially controlled, and parallax computed to provide a realistic experience. The scene may be augmented with various data, videos and other effects/animations.
-
Citations
19 Claims
-
1. A system comprising:
-
a service configured to receive video information and depth information corresponding to data captured by camera mechanisms of remote participants; a view generator coupled to the service, the view generator configured to process data corresponding to the video information and depth information to place visible representations of remote participants into a common scene, wherein the common scene is rendered via a first person point of view; a tracker using position tracking data to re-render the common scene to compensate for parallax as a user viewing the scene moves among different viewing angles; and an audio output controller to provide spatial audio based upon the position of the user, or based upon a position of a visible representation of a remote participant placed in the common scene, or both. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11)
-
-
12. In a computing environment, a method performed at least in part on at least one processor, comprising:
-
receiving a plurality of sets of data, each set of data corresponding to video and depth data associated with a remote participant; generating a photo-realistic representation of each remote participant based upon the video and depth data associated with that remote participant; rendering a common scene via a first-person point of view with the photo-realistic representations of the remote participants placed into the common scene; using position tracking data to re-render the common scene to compensate for parallax as a user viewing the scene moves among different viewing angles; and controlling audio output to provide spatial audio based upon the position of the user, or based upon a position of a visible representation of a remote participant placed in the common scene, or both. - View Dependent Claims (13, 14, 15, 16, 17)
-
-
18. One or more computer-readable media having computer-executable instructions, which when executed perform steps, comprising:
-
receiving a plurality of sets of data, each set of data corresponding to video and depth data associated with a remote participant; generating a photo-realistic representation of each participant based upon the video and depth data associated with that remote participant; rendering a common scene via a first-person point of view with the photo-realistic representations of the remote participants placed in the common scene; using position tracking data to re-render the common scene to compensate for parallax as a user viewing the scene moves among different viewing angles; and controlling audio output to provide spatial audio based upon the position of the user, or based upon a position of a visible representation of a remote participant placed in the common scene, or both. - View Dependent Claims (19)
-
Specification