Low-latency fusing of virtual and real content
First Claim
Patent Images
1. A system for presenting a mixed reality experience to a user, the system comprising:
- a display device including a first set of sensors for sensing data relating to a position of the display device and data relating to positions of real world objects in a field of view of the display device, the display device further including a display unit for displaying a virtual image to the user of the display device;
a processing unit associated with the display device receiving sensor data from the associated display device, the processing unit receiving the data relating to positions of the real world objects in the field of view at a present time within a video frame, and the processing unit extrapolating a position of a virtual three-dimensional object at a future time in the same video frame, the virtual three-dimensional object positioned at least partially behind a position of a real world object, the position extrapolated to the future time to display a proper occlusion of the virtual object by the real world object at the future time when the virtual object is displayed by the display unit.
2 Assignments
0 Petitions
Accused Products
Abstract
A system that includes a head mounted display device and a processing unit connected to the head mounted display device is used to fuse virtual content into real content. In one embodiment, the processing unit is in communication with a hub computing device. The processing unit and hub may collaboratively determine a map of the mixed reality environment. Further, state data may be extrapolated to predict a field of view for a user in the future at a time when the mixed reality is to be displayed to the user. This extrapolation can remove latency from the system.
32 Citations
13 Claims
-
1. A system for presenting a mixed reality experience to a user, the system comprising:
-
a display device including a first set of sensors for sensing data relating to a position of the display device and data relating to positions of real world objects in a field of view of the display device, the display device further including a display unit for displaying a virtual image to the user of the display device; a processing unit associated with the display device receiving sensor data from the associated display device, the processing unit receiving the data relating to positions of the real world objects in the field of view at a present time within a video frame, and the processing unit extrapolating a position of a virtual three-dimensional object at a future time in the same video frame, the virtual three-dimensional object positioned at least partially behind a position of a real world object, the position extrapolated to the future time to display a proper occlusion of the virtual object by the real world object at the future time when the virtual object is displayed by the display unit. - View Dependent Claims (2, 3, 4, 5, 6)
-
-
7. A method of presenting a mixed reality experience to one or more users, the system comprising:
-
(a) determining state information for at least a first period of time from a first video frame, and a second period of time from a second video frame subsequent to the first video frame, the state information relating to a field of view of a user of the one or more users of an environment, the environment including a mixed reality of one or more real world objects and one or more virtual objects; (b) extrapolating the state information relating to the field of view of the user of the environment for a third period of time in the second video frame, the third period of time being a time in the second video frame, after the second period of time, when the one or more virtual objects of the mixed reality are to be displayed to the user; and (c) displaying at least one virtual object of the one or more virtual objects to the user at the third period of time based on the state information relating to the field of view of the user extrapolated in said step (b), the at least one virtual object displayed as being partially occluded behind a detected real world object, the extrapolated state information at the third period of time enabling a seamless display of the partially occluded virtual object behind the real world object. - View Dependent Claims (8, 9, 10, 11, 12, 13)
-
Specification