Methods, systems, and computer readable media for unified scene acquisition and pose tracking in a wearable display
First Claim
1. A system for unified scene acquisition and pose tracking in a wearable display, the system comprising:
- a wearable frame configured to be worn on the head of a user, the frame having;
at least one camera mounted to the wearable frame for acquiring scene information for a real scene proximate to the user, the scene information including images and depth information, the scene information including positions of real objects separate from the user in the real scene local to the user;
at least one sensor mounted to the wearable frame for acquiring images of gestures and body poses of the user;
a pose tracker mounted to the wearable frame for generating, based on the scene information, a 3D model of the scene, generating, based on the images of gestures and body poses of the user acquired by the at least one sensor, a 3D model of the user, and estimating a position and orientation of the user in relation to the 3D model of the scene based on the images and depth information acquired by the at least one camera mounted to the frame and the images of gestures and body poses of the user acquired by the at least one sensor;
a rendering unit mounted to the wearable frame for generating a virtual reality (VR) image based on the scene information acquired by the at least one camera and the estimated position and orientation of the user in relation to the 3D model of the scene, wherein the rendering unit receives, from a location remote from the user, images and depth information of real objects acquired in a remote scene, the images and depth information of real objects including an image and depth information of a virtual participant in a meeting, wherein the rendering unit receives the positions of the real objects in the scene local to the user, and determines, based on the positions of the real objects and a perceived location of the virtual participant, portions of the image of the virtual participant to occlude in the VR image, wherein image of the virtual participant comprises an image of a human participant captured by a camera local to the human participant and remote from the user; and
at least one display mounted to the frame for displaying to the user a combination of the generated VR image and the scene local to the user, wherein the VR image includes the image of the virtual participant with the portions occluded as determined by the rendering unit.
2 Assignments
0 Petitions
Accused Products
Abstract
Methods, systems, and computer readable media for unified scene acquisition and pose tracking in a wearable display are disclosed. According to one aspect, a system for unified scene acquisition and pose tracking in a wearable display includes a wearable frame configured to be worn by a user. Mounted on the frame are: at least one sensor for acquiring scene information for a real scene proximate to the user, the scene information including images and depth information; a pose tracker for estimating the user'"'"'s head pose based on the acquired scene information; a rendering unit for generating a virtual reality (VR) image based on the acquired scene information and estimated head pose; and at least one display for displaying to the user a combination of the generated VR image and the scene proximate to the user.
46 Citations
37 Claims
-
1. A system for unified scene acquisition and pose tracking in a wearable display, the system comprising:
a wearable frame configured to be worn on the head of a user, the frame having; at least one camera mounted to the wearable frame for acquiring scene information for a real scene proximate to the user, the scene information including images and depth information, the scene information including positions of real objects separate from the user in the real scene local to the user; at least one sensor mounted to the wearable frame for acquiring images of gestures and body poses of the user; a pose tracker mounted to the wearable frame for generating, based on the scene information, a 3D model of the scene, generating, based on the images of gestures and body poses of the user acquired by the at least one sensor, a 3D model of the user, and estimating a position and orientation of the user in relation to the 3D model of the scene based on the images and depth information acquired by the at least one camera mounted to the frame and the images of gestures and body poses of the user acquired by the at least one sensor; a rendering unit mounted to the wearable frame for generating a virtual reality (VR) image based on the scene information acquired by the at least one camera and the estimated position and orientation of the user in relation to the 3D model of the scene, wherein the rendering unit receives, from a location remote from the user, images and depth information of real objects acquired in a remote scene, the images and depth information of real objects including an image and depth information of a virtual participant in a meeting, wherein the rendering unit receives the positions of the real objects in the scene local to the user, and determines, based on the positions of the real objects and a perceived location of the virtual participant, portions of the image of the virtual participant to occlude in the VR image, wherein image of the virtual participant comprises an image of a human participant captured by a camera local to the human participant and remote from the user; and at least one display mounted to the frame for displaying to the user a combination of the generated VR image and the scene local to the user, wherein the VR image includes the image of the virtual participant with the portions occluded as determined by the rendering unit. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17)
-
18. A method for unified scene acquisition and pose tracking in a wearable display, the method comprising:
-
acquiring, from at least one camera that is mounted to a wearable display frame configured to be worn on the head of a user, scene information for a scene local to the user, the scene information including images and depth information, the scene information including positions of real objects separate from the user in the real scene local to the user; acquiring, using at least one sensor mounted to the wearable frame, images of gestures and body poses of the user; generating, by a pose tracker that is mounted to the wearable display frame and based on the scene information, a 3D model of the scene, generating, by the pose tracker, from the images of gestures and body poses of the users acquired by the at least one sensor, a 3D model of the user, and estimating, by the pose tracker, a position and orientation of the user in relation to the 3D model of the scene based on the images and depth information acquired by the at least one camera and the images of the gestures and body poses of the user acquired by the at least one sensor; generating, by a rendering unit that is mounted to the wearable display frame, a virtual reality (VR) image based on the scene information acquired by the at least one camera and the estimated position and orientation of the user in relation to the 3D model of the scene, wherein the rendering unit receives, from a location remote from the user, images and depth information of real objects acquired in a remote scene, the images and depth information of real objects including an image and depth information of a virtual participant in a meeting, wherein the rendering unit receives the positions of the real objects in the scene local to the user and determines, based on the positions of the real objects in the scene local to the user and a perceived location of the virtual participant, portions of the image of the virtual participant to occlude in the VR image, wherein image of the virtual participant comprises an image of a human participant captured by a camera local to the human participant and remote from the user; and displaying to the user a combination of the generated VR image and the scene local to the user using at least one display that is mounted to the wearable display frame, wherein the VR image includes the image of the virtual participant with portions occluded as determined by the rendering unit. - View Dependent Claims (19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35)
-
-
36. A non-transitory computer readable medium having stored thereon executable instructions that when executed by the processor of a computer control the computer to perform steps comprising:
-
acquiring, from at least one camera that is mounted to a wearable display frame configured to be worn on the head of a user, scene information for a scene local to a user, the scene information including images and depth information, the scene information including positions of real objects separate from the user in the real scene local to the user; acquiring, using at least one sensor mounted to the wearable frame, gestures and body poses of the user; generating, by a pose tracker that is mounted to the wearable display frame and based on the scene information, a 3D model of the scene, generating, by the pose tracker, from the images of gestures and body poses of the users acquired by the at least one sensor, a 3D model of the user, and estimating, by the pose tracker, a position and orientation of the user in relation to the 3D model of the scene based on the images and depth information acquired by the at least one camera and the images of gestures and body poses of the user acquired by the at least one sensor; generating, by a rendering unit that is mounted to the wearable display frame, a virtual reality (VR) image based on the scene information acquired by the at least one camera and the estimated position and orientation of the user in relation to the 3D model of the scene local to the user, wherein the rendering unit receives, from a location remote from the user, images and depth information of real objects acquired in a remote scene, the images and depth information including an image and depth information of a virtual participant in a meeting, wherein the rendering unit receives the positions of the real objects in the scene local to the user and determines, based on the positions of the real objects and a perceived location of the virtual participant, portions of the image of the virtual participant to occlude in the VR image, wherein image of the virtual participant comprises an image of a human participant captured by a camera local to the human participant and remote from the user; and displaying to the user a combination of the generated VR image and the scene proximate to the user, wherein the VR image includes the image of the virtual participant with the portions occluded as determined by the rendering unit using at least one display that is mounted to the wearable display frame. - View Dependent Claims (37)
-
Specification