Navigating content utilizing speech-based user-selectable elements
First Claim
Patent Images
1. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed by one or more processors, cause the one or more processors to perform acts comprising:
- receiving a designation of content having user-selectable elements;
analyzing the content to identify an audio command corresponding to one or more user-selectable elements of the user-selectable elements, the audio command being identified based at least in part on an acoustic differentiation between the audio command and a different audio command meeting or exceeding a threshold;
receiving a signal associated with an utterance of a user, the signal generated by a microphone associated with a device;
analyzing the signal associated with the utterance to determine the audio command; and
responding to the utterance in accordance with a user-selectable element of the one or more user-selectable elements corresponding to the audio command, the responding includingcausing information associated with the audio command to be visually output via a projector associated with the device.
2 Assignments
0 Petitions
Accused Products
Abstract
In a content browsing environment, a system analyzes content to identify audio commands to be made available to users. The audio commands may be chosen so that they are easily differentiable from each other when using machine-based speech recognition techniques. When the content is displayed, the system monitors a user'"'"'s speech to detect user utterances corresponding to the audio commands and performs content navigation in response to the user utterances.
-
Citations
29 Claims
-
1. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed by one or more processors, cause the one or more processors to perform acts comprising:
-
receiving a designation of content having user-selectable elements; analyzing the content to identify an audio command corresponding to one or more user-selectable elements of the user-selectable elements, the audio command being identified based at least in part on an acoustic differentiation between the audio command and a different audio command meeting or exceeding a threshold; receiving a signal associated with an utterance of a user, the signal generated by a microphone associated with a device; analyzing the signal associated with the utterance to determine the audio command; and responding to the utterance in accordance with a user-selectable element of the one or more user-selectable elements corresponding to the audio command, the responding including causing information associated with the audio command to be visually output via a projector associated with the device. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 27, 28, 29)
-
-
9. A method, comprising:
-
receiving, by one or more computing devices, a request designating content, wherein the content has a user-selectable element; determining, by at least one computing device of the one or more computing devices, an audio command corresponding to the user-selectable element, the audio command being determined based at least in part on an acoustic differentiation between the audio command and a different audio command meeting or exceeding a threshold; associating, by at least one computing device of the one or more computing devices, the audio command with the user-selectable element; and causing information associated with audio command to be visually output by a projector associated with the second device. - View Dependent Claims (10, 11, 12, 13, 14, 15, 16, 17)
-
-
18. A system comprising:
-
one or more processors; one or more computer-readable media storing computer-executable instructions that, when executed by the one or more processors, cause the one or more processors to perform acts comprising; receiving a request that specifies a user utterance with regard to content having user-selectable elements; analyzing the content to identify an audio command corresponding to one or more user-selectable elements of the user-selectable elements, the audio command being identified based at least in part on an acoustic differentiation between the audio command and a different audio command meeting or exceeding a threshold; selecting the audio command based at least in part on the user utterance; and responding to the request in accordance with a user-selectable element of the one or more user-selectable elements corresponding to the audio command, the responding including causing information associated with the audio command to be visually output via a projector associated with a device. - View Dependent Claims (19, 20, 21, 22, 23, 24, 25, 26)
-
Specification