Amplifying audio-visual data based on user's head orientation
First Claim
1. A mobile communication device comprising:
- one or more communication interfaces configured to;
communicate with a media presentation device; and
one or more processors configured to;
receive, via the one or more communication interfaces, from the media presentation device, sensory data comprising information describing orientation of the media presentation device, and scene information representative of an image of a three dimensional audio space adjacent to a user,receive a media request comprising selected information related to one or more objects,determine an approximate direction of a gaze of the user in the three dimensional audio space based on the information describing the orientation of the media presentation device,query a map database based on the selected information of the media request to identify a sound source in the three dimensional audio space at which the user gazes by;
identifying an object in the three dimensional audio space based on a location of the user, the approximate direction of the gaze, and information in the map database; and
determining whether the identified object is the sound source at which the user gazes by at least one of;
i) determining whether an image obtained from the scene information matches an image of the identified object obtained from the map database; and
/or ii) determining whether a distance from the identified object to the user is approximately equal to a distance from the sound source at which the user gazes to the user;
obtain audio data based on the identified sound source at which the user gazes,determine an action to perform on the obtained audio data associated with the identified sound source at which the user gazes, andperform the determined action.
7 Assignments
0 Petitions
Accused Products
Abstract
One or more devices may receive, from a media presentation device worn by a user, sensory data comprising at least information describing orientation of the media presentation device. The one or more devices may also determine an approximate direction of a gaze of the user in a three dimensional audio space based on the information describing the orientation of the media presentation device, identify a sound source at which the user gazes based on the approximate direction of the gaze of the user in the three dimensional audio space, obtain audio data associated with the sound source based on an identity of the identified sound source, determine an action to perform on the audio data associated with the sound source, and perform the determined action.
-
Citations
18 Claims
-
1. A mobile communication device comprising:
-
one or more communication interfaces configured to; communicate with a media presentation device; and one or more processors configured to; receive, via the one or more communication interfaces, from the media presentation device, sensory data comprising information describing orientation of the media presentation device, and scene information representative of an image of a three dimensional audio space adjacent to a user, receive a media request comprising selected information related to one or more objects, determine an approximate direction of a gaze of the user in the three dimensional audio space based on the information describing the orientation of the media presentation device, query a map database based on the selected information of the media request to identify a sound source in the three dimensional audio space at which the user gazes by; identifying an object in the three dimensional audio space based on a location of the user, the approximate direction of the gaze, and information in the map database; and determining whether the identified object is the sound source at which the user gazes by at least one of;
i) determining whether an image obtained from the scene information matches an image of the identified object obtained from the map database; and
/or ii) determining whether a distance from the identified object to the user is approximately equal to a distance from the sound source at which the user gazes to the user;obtain audio data based on the identified sound source at which the user gazes, determine an action to perform on the obtained audio data associated with the identified sound source at which the user gazes, and perform the determined action. - View Dependent Claims (2, 3, 4, 5, 6)
-
-
7. A device comprising:
-
sensors configured to; collect sensory data comprising images of eyes of a user and scene information representative of an image of an audio space adjacent to the user; and one or more processors configured to; receive the sensory data from the sensors, receive a media request comprising selected information related to one or more images of sound sources, determine an approximate direction of a gaze of the user based on the images, included in the sensory data, of the eyes of the user, query a map database based on the selected information of the media request to identify a sound source at which the user gazes by; identifying an object in the audio space based on a location of the user, the approximate direction of the gaze, and information in the map database; and determining whether the identified object is the sound source at which the user gazes by at least one of;
i) determining whether an image obtained from the scene information matches an image of the identified object obtained from the map database; and
/or ii) determining whether a distance from the identified object to the user is approximately equal to a distance from the sound source at which the user gazes to the user;obtain audio data associated with the identified sound source at which the user gazes based on an identity of the identified sound source at which the user gazes, determine an action to perform on the obtained audio data associated with the identified sound source at which the user gazes, and perform the determined action. - View Dependent Claims (8, 9)
-
-
10. A method comprising:
-
receiving, from a media presentation device worn by a user, sensory data comprising i) information describing an orientation of the media presentation device, and ii) scene information representative of an image of a three dimensional audio space adjacent to the media presentation device; receiving a media request comprising selected information related to one or more images of sound sources; determining an approximate direction of a gaze of the user in the three dimensional audio space based on the information describing the orientation of the media presentation device; querying a map database based on the selected information of the media request to identify a sound source at which the user gazes by; identifying an object in the three dimensional audio space based on a location of the user, the approximate direction of the gaze, and information in the map database; and determining whether the identified object is the sound source at which the user gazes by at least one of;
i) determining whether an image obtained from the scene information matches an image of the identified object obtained from the map database; and
/or ii) determining whether a distance from the identified object to the user is approximately equal to a distance from the sound source at which the user gazes to the user;obtaining audio data associated with the identified sound source at which the user gazes based on an identity of the identified sound source at which the user gazes; determining an action to perform on the obtained audio data associated with the identified sound source at which the user gazes; and performing the determined action. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17, 18)
-
Specification