Conversation user interface
First Claim
1. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed on one or more processors, cause the one or more processors of a virtual assistant service to perform acts comprising:
- generating data for a conversation graphical user interface (GUI) that represents a virtual assistant;
causing display of the conversation GUI via a computing device to enable a conversation between the virtual assistant and a user of the computing device;
receiving user input that is provided via the conversation GUI, the user input comprising one of audio input, keypad input, or touch input;
parsing the user input with a natural language processing system that employs a language model;
determining, with the natural language processing system, a response based at least in part on (i) the parsed user input and (ii) at least one of content of a service provider, content of the virtual assistant service, or content of the computing device of the user;
identifying an assumption that is used to determine the response, the assumption comprising at least one of the language model that is employed by the natural language processing system, a profile for the user, or a learned behavior of the user;
causing display of a dialog representation in the conversation GUI for the user input;
causing display of a dialog representation in the conversation GUI for the response;
causing display of the assumption in the conversation GUI;
enabling the user to interact with the conversation GUI to modify the assumption;
receiving a modification to the assumption;
determining a revised response based at least in part on the modification to the assumption; and
causing display of the revised response in the conversation GUI.
5 Assignments
0 Petitions
Accused Products
Abstract
A conversation user interface enables users to better understand their interactions with computing devices, particularly when speech input is involved. The conversation user interface conveys a visual representation of a conversation between the computing device, or virtual assistant thereon, and a user. The conversation user interface presents a series of dialog representations that show input from a user (verbal or otherwise) and responses from the device or virtual assistant. Associated with one or more of the dialog representations are one or more graphical elements to convey assumptions made to interpret the user input and derive an associated response. The conversation user interface enables the user to see the assumptions upon which the response was based, and to optionally change the assumption(s). Upon change of an assumption, the conversation GUI is refreshed to present a modified dialog representation of a new response derived from the altered set of assumptions.
-
Citations
47 Claims
-
1. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed on one or more processors, cause the one or more processors of a virtual assistant service to perform acts comprising:
-
generating data for a conversation graphical user interface (GUI) that represents a virtual assistant; causing display of the conversation GUI via a computing device to enable a conversation between the virtual assistant and a user of the computing device; receiving user input that is provided via the conversation GUI, the user input comprising one of audio input, keypad input, or touch input; parsing the user input with a natural language processing system that employs a language model; determining, with the natural language processing system, a response based at least in part on (i) the parsed user input and (ii) at least one of content of a service provider, content of the virtual assistant service, or content of the computing device of the user; identifying an assumption that is used to determine the response, the assumption comprising at least one of the language model that is employed by the natural language processing system, a profile for the user, or a learned behavior of the user; causing display of a dialog representation in the conversation GUI for the user input; causing display of a dialog representation in the conversation GUI for the response; causing display of the assumption in the conversation GUI; enabling the user to interact with the conversation GUI to modify the assumption; receiving a modification to the assumption; determining a revised response based at least in part on the modification to the assumption; and causing display of the revised response in the conversation GUI. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20)
-
-
21. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed on one or more processors, cause the one or more processors to perform acts comprising:
-
parsing user input with natural language processing techniques that utilize a language model, the user input comprising at least one of gesture input, audio input, keypad input, or touch input; determining a response based at least in part on the parsed user input; causing display of a conversation user interface having a first dialog representation associated with the user input and a second dialog representation associated with the response, the conversation user interface being associated with a virtual assistant; providing, for display as part of the conversation user interface, one or more graphical elements to convey at least one assumption made to interpret the user input and determine the response; and enabling, via the conversation user interface, a user to change the assumption, whereupon change of the assumption causes the second dialog representation associated with the response to be removed and a third dialog representation to be displayed, the third dialog representation representing an updated response for the virtual assistant. - View Dependent Claims (22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35)
-
-
36. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed on one or more processors, cause the one or more processors to perform acts comprising:
-
causing display of a conversation user interface that represents a conversation between a user and a virtual assistant; parsing, based at least in part on natural language processing techniques that employ a language model, input from the user that is received during the conversation between the user and the virtual assistant, the input from the user comprising at least one of gesture input, audio input, keypad input, or touch input; determining a response for the virtual assistant based at least in part on the parsed input; adding, for display in the conversation user interface, a first dialog representation associated with the input from the user and a second dialog representation associated with the response from the virtual assistant to visually convey the conversation; and enabling, via interaction with the conversation user interface, the user to select a particular one of the first or second dialog representations within the conversation user interface and restart the conversation from the particular one of the first or second dialog representations, the restarting including removing the particular one of the first or second dialog representations and displaying a third dialog representation associated with an updated response from the virtual assistant. - View Dependent Claims (37)
-
-
38. A computing device, comprising:
-
a memory; at least one processor; a display; a communication module stored in the memory and executed by the at least one processor to; send input from a user of the computing device to be parsed by a natural language processing system that employs a language model, the input including at least one of audio or text; and receive from the natural language processing system a response to the input from the user; and a conversation application stored in the memory and executed by the at least one processor to facilitate a conversation with the user of the computing device, the conversation application providing a conversation graphical user interface (GUI) to be displayed on the display, the conversation GUI comprising; an entry portion to receive, from the user, the input; an area to visually convey the conversation, the area to present a user dialog representation associated with the input from the user and a device dialog representation associated with the response to the input from the user; and a control associated with at least one of the user or a device dialog representation to visually change, upon actuation, the conversation GUI to reveal an assumption that was used to determine the response, the assumption comprising at least one of a value obtained from the conversation, a value that is external to the conversation, or a value that is obtained from a previous conversation with the user. - View Dependent Claims (39)
-
-
40. A method comprising:
-
under control of one or more computing devices configured to implement a virtual assistant, parsing user input with a natural language processing system that employs a language model, the user input comprising at least one of gesture input, audio input, keypad input, or touch input; representing the user input in a conversation user interface; determining, with the natural language processing system, a response to the user input; representing the response to the user input in the conversation user interface; enabling a user to interact with the conversation user interface to ascertain how the response was determined and to modify assumptions used to determine the response, wherein modification of one or more assumptions results in a modified response, and wherein individual assumptions comprise at least one of a value obtained via the conversation user interface, a value that is external to the conversation user interface, or a value that is obtained from a previous conversation between the user and the virtual assistant; and representing the modified response in the conversation user interface. - View Dependent Claims (41, 42, 43, 44)
-
-
45. One or more non-transitory computer-readable media storing computer-executable instructions that, when executed on one or more processors, cause the one or more processors to perform acts comprising:
-
parsing input from a user with natural language processing techniques that utilize a language model, the input from the user comprising at least one of gesture input, audio input, keypad input, or touch input; determining a response to the input from the user based at least in part on the parsed input; serving or rendering a conversation graphical user interface (GUI), the conversation GUI comprising; an entry portion to receive, from the user, the input; an area to visually convey a conversation, the area to present a user dialog representation associated with the input received from the user and a device dialog representation associated with the response to the input from the user; and a control associated with at least one of the user or device dialog representation to visually change, upon actuation, the conversation GUI to reveal an assumption that was used to determine the response, the assumption comprising at least one of a value obtained from the conversation, a value that is external to the conversation, or a value that is obtained from a previous conversation with the user. - View Dependent Claims (46, 47)
-
Specification