REDUCING THE NEED FOR MANUAL START/END-POINTING AND TRIGGER PHRASES
First Claim
1. A method for operating a virtual assistant on an electronic device, the method comprising:
- receiving, at the electronic device, an audio input;
monitoring the audio input to identify a first spoken user input;
identifying the first spoken user input in the audio input;
determining whether to respond to the first spoken user input based on contextual information associated with the first spoken user input;
in response to a determination to respond to the first spoken user input;
generating a response to the first spoken user input; and
monitoring the audio input to identify a second spoken user input; and
in response to a determination not to respond to the first spoken user input, monitoring the audio input to identify the second spoken user input without generating the response to the first spoken user input.
1 Assignment
0 Petitions
Accused Products
Abstract
Systems and processes for selectively processing and responding to a spoken user input are provided. In one example, audio input containing a spoken user input can be received at a user device. The spoken user input can be identified from the audio input by identifying start and end-points of the spoken user input. It can be determined whether or not the spoken user input was intended for a virtual assistant based on contextual information. The determination can be made using a rule-based system or a probabilistic system. If it is determined that the spoken user input was intended for the virtual assistant, the spoken user input can be processed and an appropriate response can be generated. If it is instead determined that the spoken user input was not intended for the virtual assistant, the spoken user input can be ignored and/or no response can be generated.
-
Citations
38 Claims
-
1. A method for operating a virtual assistant on an electronic device, the method comprising:
-
receiving, at the electronic device, an audio input; monitoring the audio input to identify a first spoken user input; identifying the first spoken user input in the audio input; determining whether to respond to the first spoken user input based on contextual information associated with the first spoken user input; in response to a determination to respond to the first spoken user input; generating a response to the first spoken user input; and monitoring the audio input to identify a second spoken user input; and in response to a determination not to respond to the first spoken user input, monitoring the audio input to identify the second spoken user input without generating the response to the first spoken user input. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36)
-
-
37. A non-transitory computer-readable storage medium comprising instructions for:
-
receiving an audio input; monitoring the audio input to identify a first spoken user input; identifying the first spoken user input in the audio input; determining whether to respond to the first spoken user input based on contextual information associated with the first spoken user input; responsive to a determination to respond to the first spoken user input; generating a response to the first spoken user input; and monitoring the audio input to identify a second spoken user input; and responsive to a determination not to respond to the first spoken user input, monitoring the audio input to identify the second spoken user input without generating the response to the first spoken user input.
-
-
38. A system comprising:
-
one or more processors; memory; and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs including instructions for; receiving an audio input; monitoring the audio input to identify a first spoken user input; identifying the first spoken user input in the audio input; determining whether to respond to the first spoken user input based on contextual information associated with the first spoken user input; responsive to a determination to respond to the first spoken user input; generating a response to the first spoken user input; and monitoring the audio input to identify a second spoken user input; and responsive to a determination not to respond to the first spoken user input, monitoring the audio input to identify the second spoken user input without generating the response to the first spoken user input.
-
Specification