Systems and methods for responding to natural language speech utterance
First Claim
1. A system for multi-pass speech recognition, comprising:
- an input device configured to receive a natural language utterance; and
a multi-pass speech recognition module configured to transcribe the natural language utterance, wherein to transcribe the natural language utterance, the multi-pass speech recognition module is further configured to;
use a dictation grammar to transcribe the natural language utterance in response to a platform associated with the multi-pass speech recognition module having the dictation grammar available;
oruse a virtual dictation grammar to transcribe the natural language utterance in response to the platform associated with the multi-pass speech recognition module not having the dictation grammar available.
6 Assignments
0 Petitions
Accused Products
Abstract
Systems and methods are provided for receiving speech and non-speech communications of natural language questions and/or commands, transcribing the speech and non-speech communications to textual messages, and executing the questions and/or commands. The invention applies context, prior information, domain knowledge, and user specific profile data to achieve a natural environment for one or more users presenting questions or commands across multiple domains. The systems and methods creates, stores and uses extensive personal profile information for each user, thereby improving the reliability of determining the context of the speech and non-speech communications and presenting the expected results for a particular question or command.
-
Citations
36 Claims
-
1. A system for multi-pass speech recognition, comprising:
-
an input device configured to receive a natural language utterance; and a multi-pass speech recognition module configured to transcribe the natural language utterance, wherein to transcribe the natural language utterance, the multi-pass speech recognition module is further configured to; use a dictation grammar to transcribe the natural language utterance in response to a platform associated with the multi-pass speech recognition module having the dictation grammar available;
oruse a virtual dictation grammar to transcribe the natural language utterance in response to the platform associated with the multi-pass speech recognition module not having the dictation grammar available. - View Dependent Claims (2, 3, 4)
-
-
5. A system for multi-pass speech recognition, comprising:
-
an input device configured to receive a natural language utterance; and a multi-pass speech recognition module configured to; determine whether a platform associated with the multi-pass speech recognition module has a dictation grammar available or a virtual dictation grammar available; and use the dictation grammar or the virtual dictation grammar to transcribe the natural language utterance based on whether the platform has the dictation grammar available or the virtual dictation grammar available.
-
-
6. A method for multi-pass speech recognition, comprising:
-
receiving a natural language utterance at an input device; and transcribing the natural language utterance with a multi-pass speech recognition module, wherein transcribing the natural language utterance with the multi-pass speech recognition module includes; using a dictation grammar to transcribe the natural language utterance in response to determining that a platform associated with the multi-pass speech recognition module has the dictation grammar available;
orusing a virtual dictation grammar to transcribe the natural language utterance in response to determining that the platform associated with the multi-pass speech recognition module does not have the dictation grammar available. - View Dependent Claims (7, 8, 9)
-
-
10. A method for multi-pass speech recognition, comprising:
-
receiving a natural language utterance at an input device; determining whether a platform associated with a multi-pass speech recognition module has a dictation grammar available or a virtual dictation grammar available; and transcribing the natural language utterance with the multi-pass speech recognition module, wherein the multi-pass speech recognition module uses the dictation grammar or the virtual dictation grammar to transcribe the natural language utterance based on whether the platform has the dictation grammar available or the virtual dictation grammar available.
-
-
11. A system for knowledge-enhanced speech recognition, comprising:
-
a context stack configured to store one or more expected contexts associated with a natural language utterance; and a knowledge-enhanced speech recognition engine, wherein the knowledge-enhanced speech recognition engine includes one or more processors configured to; access the one or more expected contexts stored in the context stack in response to one or more active grammars in a context description grammar failing to completely match information associated with the natural language utterance; compare the information associated with the natural language utterance to one or more context specific matchers to determine a most likely context associated with the natural language utterance from the one or more expected contexts stored in the context stack; and use one or more grammar expression entries in the context description grammar to generate a command or request associated with the most likely context. - View Dependent Claims (12, 13, 14, 15)
-
-
16. A system for knowledge-enhanced speech recognition comprising:
-
a context stack configured to store one or more expected contexts associated with a natural language utterance; a knowledge-enhanced speech recognition engine, wherein the knowledge-enhanced speech recognition engine includes one or more processors configured to; access the one or more expected contexts stored in the context stack in response to one or more active grammars in a context description grammar failing to completely match information associated with the natural language utterance; compare the information associated with the natural language utterance to one or more context specific matchers to determine a most likely context associated with the natural language utterance from the one or more expected contexts stored in the context stack; and use one or more, grammar expression entries in the context description grammar to generate a command or request associated with the most likely context; and an agent configured to; process the generated command or request in the most likely context to generate a response to the natural language utterance; and update an ordered list associated with the one or more expected contexts in the context stack with information associated with one or more of the most likely context, the generated command or request, or the generated response to enable one or more follow-up commands or requests associated with the most likely context, the generated command or request, or the generated response.
-
-
17. A system for knowledge-enhanced speech recognition, comprising:
-
a context stack configured to store one or more expected contexts associated with a natural language utterance; and a knowledge-enhanced speech recognition engine, wherein the knowledge-enhanced speech recognition engine includes one or more processors configured to; access the one or more expected contexts stored in the context stack in response to one or more active grammars in a context description grammar failing to completely match information associated with the natural language utterance; compare the information associated with the natural language utterance to one or more context specific matchers to determine a most likely context associated with the natural language utterance from the one or more expected contexts stored in the context stack; and use one or more grammar expression entries in the context description grammar to generate a command or request associated with the most likely context, wherein the information compared to the one or more context specific matchers includes phonetic information associated with the natural language utterance or text combinations from a transcription associated with the natural language utterance. - View Dependent Claims (18, 19)
-
-
20. A method for knowledge-enhanced speech recognition, comprising:
-
storing one or more expected contexts in a context stack, wherein a knowledge-enhanced speech recognition engine that includes one or more processors accesses the one or more expected contexts in the context stack in response to one or more active grammars in a context description grammar failing to completely match information associated with the natural language utterance; comparing the information associated with the natural language utterance to one or more context specific matchers to determine a most likely context associated with the natural language utterance, wherein the knowledge-enhanced speech recognition engine determines the most likely context from the one or more expected contexts in the context stack; and using one or more grammar expression entries in the context description grammar to generate a command or request associated with the most likely context. - View Dependent Claims (21, 22, 23, 24)
-
-
25. A method for knowledge-enhanced speech recognition, further comprising:
-
storing one or more expected contexts in a context stack, wherein a knowledge-enhanced speech recognition engine that includes one or more processors accesses the one or more expected contexts in the context stack in response to one or more active grammars in a context description grammar failing to completely match information associated with the natural language utterance; comparing the information associated with the natural language utterance to one or more context specific matchers to determine a most likely context associated with the natural language utterance, wherein the knowledge-enhanced speech recognition engine determines the most likely context from the one or more expected contexts in the context stack; using one or more grammar expression entries in the context description grammar to generate a command or request associated with the most likely context; processing the generated command or request with an agent associated with the most likely context, wherein the agent processes the generated command or request to generate a response to the natural language utterance; and updating an ordered list associated with the one or more expected contexts in the context stack with information associated with one or more of the most likely context, the generated command or request, or the generated response, wherein the agent updates the ordered list to enable one or more follow-up commands or requests associated with the most likely context, the generated command or request, or the generated response.
-
-
26. A method for knowledge-enhanced speech recognition, comprising:
-
storing one or more expected contexts in a context stack, wherein a knowledge-enhanced speech recognition engine that includes one or more processors accesses the one or more expected contexts in the context stack in response to one or more active grammars in a context description grammar failing to completely match information associated with the natural language utterance; comparing the information associated with the natural language utterance to one or more context specific matchers to determine a most likely context associated with the natural language utterance, wherein the knowledge-enhanced speech recognition engine determines the most likely context from the one or more expected contexts in the context stack; and using one or more grammar expression entries in the context description grammar to generate a command or request associated with the most likely context, wherein the information compared to the one or more context specific matchers includes phonetic information associated with the natural language utterance or text combinations from a transcription associated with the natural language utterance. - View Dependent Claims (27, 28)
-
-
29. A system for synchronizing context across multiple electronic devices, comprising:
-
one or more processors configured to; subscribe a first electronic device to one or more context events; receive a context change event from a second electronic device; and inform the first electronic device of the context change event to synchronize a context across the first electronic device and the second electronic device; and a registration module configured to; register a library specifically associated with the first electronic device to subscribe the first electronic device to the one or more context events; and remove the library specifically associated with the first electronic device to unsubscribe the first electronic device from the one or more context events. - View Dependent Claims (30, 31, 32)
-
-
33. A method for synchronizing context across multiple electronic devices, comprising:
-
subscribing a first electronic device to one or more context events; receiving a context change event from a second electronic device; informing the first electronic device of the context change event to synchronize a context across the first electronic device and the second electronic device; registering a library specifically associated with the first electronic device to subscribe the first electronic device to the one or more context events; and removing the library specifically associated with the first electronic device to unsubscribe the first electronic device from the one or more context events. - View Dependent Claims (34, 35, 36)
-
Specification