Voice and connection platform
First Claim
Patent Images
1. A computer-implemented method comprising:
- detecting an event;
responsive to detecting the event, proactively initiating a dialogue of a voice assistant on a first user device with a user;
responsive to initiating the dialogue with the user, receiving, at the first user device, a first audio input associated with the dialogue from the user requesting a first action;
performing automatic speech recognition on the first audio input;
determining, at the first user device, a first context of the user;
determining a first tuple describing user intent, the first tuple including the first action and an actor associated with the first action, the first tuple determined by performing natural language understanding based on the automatic speech recognition of the first audio input;
initiating the first action on the first user device based on the first tuple;
subsequent to initiating the first action, receiving a second audio input from the user requesting a second action unrelated to the first action;
initiating the second action;
subsequent to initiating the second action, receiving, at a second user device distinct from the first user device, a third audio input from the user continuing the dialogue and requesting a third action related to the first action, the third audio input missing information for completing a third tuple, the third tuple for initiating the third action;
obtaining the missing information using the first context to complete the third tuple associated with the third action; and
initiating the third action on the second user device based on the third tuple.
1 Assignment
0 Petitions
Accused Products
Abstract
A system and method for providing a voice assistant including receiving, at a first device, a first audio input from a user requesting a first action; performing automatic speech recognition on the first audio input; obtaining a context of user; performing natural language understanding based on the speech recognition of the first audio input; and taking the first action based on the context of the user and the natural language understanding.
-
Citations
18 Claims
-
1. A computer-implemented method comprising:
-
detecting an event; responsive to detecting the event, proactively initiating a dialogue of a voice assistant on a first user device with a user; responsive to initiating the dialogue with the user, receiving, at the first user device, a first audio input associated with the dialogue from the user requesting a first action; performing automatic speech recognition on the first audio input; determining, at the first user device, a first context of the user; determining a first tuple describing user intent, the first tuple including the first action and an actor associated with the first action, the first tuple determined by performing natural language understanding based on the automatic speech recognition of the first audio input; initiating the first action on the first user device based on the first tuple; subsequent to initiating the first action, receiving a second audio input from the user requesting a second action unrelated to the first action; initiating the second action; subsequent to initiating the second action, receiving, at a second user device distinct from the first user device, a third audio input from the user continuing the dialogue and requesting a third action related to the first action, the third audio input missing information for completing a third tuple, the third tuple for initiating the third action; obtaining the missing information using the first context to complete the third tuple associated with the third action; and initiating the third action on the second user device based on the third tuple. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A system comprising:
-
one or more processors; and a memory storing instructions that when executed by the one or more processors, cause the system to perform steps including; detect an event; responsive to detecting the event, proactively initiate a dialogue of a voice assistant on a first user device with a user; responsive to initiating the dialogue with the user, receive, at the first user device, a first audio input associated with the dialogue from the user requesting a first action; perform automatic speech recognition on the first audio input; determine, at the first user device, a first context of the user; determine a first tuple describing user intent, the first tuple including the first action and an actor associated with the first action, the first tuple determined by performing natural language understanding based on the automatic speech recognition of the first audio input; initiate the first action on the first user device based on the first tuple; subsequent to initiating the first action, receive a second audio input from the user requesting a second action unrelated to the first action; initiate the second action; subsequent to initiating the second action, receive, at a second user device distinct from the first user device a third audio input from the user continuing the dialogue and requesting a third action related to the first action, the third audio input missing information for completing a third tuple, the third tuple for initiating the third action; obtaining the missing information using the first context to complete the third tuple associated with the third action; and initiating the third action on the second user device based on the third tuple. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17, 18)
-
Specification