×

Enhanced speech endpointing

  • US 10,339,917 B2
  • Filed: 09/03/2015
  • Issued: 07/02/2019
  • Est. Priority Date: 09/03/2015
  • Status: Active Grant
First Claim
Patent Images

1. A computer implemented method, comprising:

  • storing, by a computing device that is configured to set an end of speech condition after a user has stopped speaking for a period of time, (i) a context identifier in association with one or more expected speech recognition results for a first context and (ii) an additional context identifier in association with one or more additional expected speech recognition results for a second context;

    after storing the context identifier in association with the one or more expected speech recognition results for the first context and the additional context identifier in association with the one or more additional expected speech recognition results for the second context, receiving, by the computing device, audio data corresponding to an utterance spoken by the user of the client device;

    while receiving the audio data corresponding to the utterance spoken, receiving, by the computing device, the context identifier that indicates a context associated with (i) the client device or (ii) the user of the client device;

    accessing, by the computing device and from among the one or more expected speech recognition results and the one or more additional expected speech recognition results, the one or more expected speech recognition results based on the one or more expected speech recognition results being stored in association with the context identifier;

    before an automated speech recognizer provides a final speech recognition result for the audio data for output, comparing, by the computing device, an intermediate speech recognition result generated for the audio data by the automated speech recognizer to each of the one or more expected speech recognition results with the context identifier;

    based at least on comparing the intermediate speech recognition result generated for the audio data by the automated speech recognizer to the one or more expected speech recognition results associated with the context identifier, determining, by the computing device, that the intermediate speech recognition result generated for the audio data by the automated speech recognizer matches one of the one or more expected speech recognition results associated with the context identifier; and

    based on determining that the intermediate speech recognition result generated for the audio data by the automated speech recognizer matches the one of the one or more expected speech recognition results associated with the context identifier and before the period of time has elapsed after the user stopped speaking,setting the end of speech condition and providing, for output by the computing device the intermediate speech recognition result that matches the one of the one or more expected speech recognition results as the final speech recognition result based on the audio data.

View all claims
  • 2 Assignments
Timeline View
Assignment View
    ×
    ×