Method and apparatus for voice-enabling an application
First Claim
1. A method of voice-enabling an application comprising a visual browser that lacks the ability to process audible input and/or output, the method comprising:
- dynamically identifying, via the application comprising the visual browser and based on a current state of the application, one or more commands and/or controls that are used to interact with the visual browser via graphic input and that are not responsive to speech;
generating, via the application comprising the visual browser, at least one markup language fragment specifying a voice grammar corresponding to the identified one or more commands and/or controls that are not responsive to speech;
instantiating, by the application and through a voice library of voice markup language functions, an interpreter by calling at least one function in the voice library via a library application programming interface (API) through which the visual browser and the voice library can communicate;
providing the at least one markup language fragment from the application that instantiated the interpreter to the interpreter for use in recognizing speech;
receiving, via the application, a speech input from a user;
receiving, via the application from the interpreter, an event specifying at least one of the identified one or more commands and/or controls, generated as a result of matching, by the interpreter, the speech input with the voice grammar specified by the at least one markup language fragment to resolve the speech input to the at least one of the identified one or more commands and/or controls so that the event includes at least one attribute specifying a semantic interpretation of the speech input thus rendering the at least one of the identified one or more commands and/or controls responsive to speech; and
interpreting the event via the application.
3 Assignments
0 Petitions
Accused Products
Abstract
A method of voice-enabling an application for command and control and content navigation can include the application dynamically generating a markup language fragment specifying a command and control and content navigation grammar for the application, instantiating an interpreter from a voice library, and providing the markup language fragment to the interpreter. The method also can include the interpreter processing a speech input using the command and control and content navigation grammar specified by the markup language fragment and providing an event to the application indicating an instruction representative of the speech input.
62 Citations
14 Claims
-
1. A method of voice-enabling an application comprising a visual browser that lacks the ability to process audible input and/or output, the method comprising:
-
dynamically identifying, via the application comprising the visual browser and based on a current state of the application, one or more commands and/or controls that are used to interact with the visual browser via graphic input and that are not responsive to speech; generating, via the application comprising the visual browser, at least one markup language fragment specifying a voice grammar corresponding to the identified one or more commands and/or controls that are not responsive to speech; instantiating, by the application and through a voice library of voice markup language functions, an interpreter by calling at least one function in the voice library via a library application programming interface (API) through which the visual browser and the voice library can communicate; providing the at least one markup language fragment from the application that instantiated the interpreter to the interpreter for use in recognizing speech; receiving, via the application, a speech input from a user; receiving, via the application from the interpreter, an event specifying at least one of the identified one or more commands and/or controls, generated as a result of matching, by the interpreter, the speech input with the voice grammar specified by the at least one markup language fragment to resolve the speech input to the at least one of the identified one or more commands and/or controls so that the event includes at least one attribute specifying a semantic interpretation of the speech input thus rendering the at least one of the identified one or more commands and/or controls responsive to speech; and interpreting the event via the application. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
-
8. A non-transitory machine readable storage having stored thereon a computer program having a plurality of code sections executable by a machine for causing the machine to perform a method of voice-enabling an application comprising a visual browser that lacks the ability to process audible input and/or output, the method comprising steps of:
-
dynamically identifying, via the application comprising the visual browser and based on a current state of the application, one or more commands and/or controls that are used to interact with the visual browser via graphic input and that are not responsive to speech; generating, via the application comprising the visual browser, at least one markup language fragment specifying a voice grammar corresponding to the identified one or more commands and/or controls that are not responsive to speech; instantiating, by the application and through a voice library of voice markup language functions, an interpreter, by calling at least one function in the voice library via a library application programming interface (API) through which the visual browser and the voice library can communicate; providing the at least one markup language fragment from the application that instantiated the interpreter to the interpreter for use in recognizing speech; receiving, via the application, a speech input from a user; receiving, via the application from the interpreter, an event specifying the at least one of the identified one or more commands and/or controls generated as a result of matching the speech input with the voice grammar specified by the at least one markup and language fragment to resolve the speech input to the at least one of the identified one or more commands and/or controls so that the event includes at least one attribute specifying a semantic interpretation of the speech input thus rendering the at least one of the identified one or more commands and/or controls responsive to speech; and interpreting the event via the application. - View Dependent Claims (9, 10, 11, 12, 13, 14)
-
Specification