Dialog management context sharing
First Claim
1. A system comprising:
- a computer-readable memory storing executable instructions; and
one or more processors in communication with the computer-readable memory, wherein the one or more processors are programmed by the executable instructions to at least;
obtain first audio data regarding a first utterance of a user, the first utterance associated with a first user request;
perform speech recognition on the first audio data to generate first speech recognition results;
cause a first application to perform a first function responsive to the first user request based on the first speech recognition results, wherein performance of the first function comprises presenting information in a first modality;
receive, from the first application, contextual information associated with the first user request in a format associated with a second application and information regarding use of the contextual information in the format, wherein the contextual information is not provided to the user in response to performing the first function;
store the contextual information and the information regarding use of the contextual information in the format;
subsequently, obtain second audio data regarding a second utterance of the user, the second utterance associated with a second user request;
perform speech recognition on the second audio data to generate second speech recognition results;
determine that the second application is associated with the second speech recognition results;
provide, to the second application, the contextual information associated with the first user request and the information regarding use of the contextual information in the format; and
cause the second application to perform a second function responsive to the second user request based on the contextual information, the information regarding use of the contextual information in the format, and the second speech recognition results, wherein performance of the second function comprises presenting information in a second modality different than the first modality.
1 Assignment
0 Petitions
Accused Products
Abstract
Features are disclosed for performing functions in response to user requests based on contextual data regarding prior user requests. Users may engage in conversations with a computing device in order to initiate some function or obtain some information. A dialog manager may manage the conversations and store contextual data regarding one or more of the conversations. Processing and responding to subsequent conversations may benefit from the previously stored contextual data by, e.g., reducing the amount of information that a user must provide if the user has already provided the information in the context of a prior conversation. Additional information associated with performing functions responsive to user requests may be shared among applications, further improving efficiency and enhancing the user experience.
17 Citations
24 Claims
-
1. A system comprising:
-
a computer-readable memory storing executable instructions; and one or more processors in communication with the computer-readable memory, wherein the one or more processors are programmed by the executable instructions to at least; obtain first audio data regarding a first utterance of a user, the first utterance associated with a first user request; perform speech recognition on the first audio data to generate first speech recognition results; cause a first application to perform a first function responsive to the first user request based on the first speech recognition results, wherein performance of the first function comprises presenting information in a first modality; receive, from the first application, contextual information associated with the first user request in a format associated with a second application and information regarding use of the contextual information in the format, wherein the contextual information is not provided to the user in response to performing the first function; store the contextual information and the information regarding use of the contextual information in the format; subsequently, obtain second audio data regarding a second utterance of the user, the second utterance associated with a second user request; perform speech recognition on the second audio data to generate second speech recognition results; determine that the second application is associated with the second speech recognition results; provide, to the second application, the contextual information associated with the first user request and the information regarding use of the contextual information in the format; and cause the second application to perform a second function responsive to the second user request based on the contextual information, the information regarding use of the contextual information in the format, and the second speech recognition results, wherein performance of the second function comprises presenting information in a second modality different than the first modality. - View Dependent Claims (2, 3, 4, 5)
-
-
6. A computer-implemented method comprising:
under control of one or more computing devices configured with specific computer-executable instructions, obtaining first speech recognition results associated with a first user request; performing, using a first application, a first function responsive to the first user request based on the first speech recognition results, wherein performing the first function comprises presenting information in a first modality; generating, using the first application, contextual information regarding the first user request in a format associated with a second application, wherein the contextual information is not provided to the user in response to the first request; storing the contextual information and information regarding use of the contextual information in the format; obtaining second speech recognition results associated with a second user request; determining to perform a second function responsive to the second user request based on the second speech recognition results; and performing, using the second application, the second function based at least in part on the contextual information regarding the first user request and the information regarding use of the contextual information in the format, wherein performing the second function comprises presenting the contextual information in a second modality different than the first modality. - View Dependent Claims (7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17)
-
18. One or more non-transitory computer readable media comprising executable code that, when executed, causes one or more computing devices to perform a process comprising:
-
obtaining first speech recognition results associated with a first user request; performing, using a first application, a first function responsive to the first user request based on the first speech recognition results, wherein performing the first function comprises presenting information in a first modality; generating, using the first application, contextual information regarding the first user request in a format associated with a second function of a second application, wherein the contextual information is not provided to the user in response to the first request; storing the contextual information and information regarding use of the contextual information in the format; obtaining second speech recognition results associated with a second user request; determining to perform the second function responsive to the second user request based on the second speech recognition results; and performing, using the second application, the second function based at least in part on the contextual information regarding the first user request and the information regarding use of the contextual information in the format, wherein performing the second function comprises presenting the contextual information in a second modality different than the first modality. - View Dependent Claims (19, 20, 21, 22, 23, 24)
-
Specification