×

Integration of embedded and network speech recognizers

  • US 8,868,428 B2
  • Filed: 08/14/2012
  • Issued: 10/21/2014
  • Est. Priority Date: 01/26/2010
  • Status: Active Grant
First Claim
Patent Images

1. A computer-implemented method comprising:

  • receiving audio data corresponding to a user utterance;

    providing the audio data to a remote speech recognizer and, in response, obtaining several search results that are identified as a result of a search of a remote database using at least a portion of a remotely-generated transcription of the user utterance as a query, the remotely-generated transcription of the user utterance being generated by the remote speech recognizer;

    obtaining a locally-generated transcription of the user utterance and a speech recognition confidence value associated with the locally-generated transcription of the user utterance, the locally-generated transcription of the user utterance and the confidence value being generated by a local speech recognizer;

    in response to determining that the speech recognition confidence value generated by the local speech recognizer fails to meet a threshold value;

    bypassing performing a search of a local database using at least a portion of the locally-generated transcription as a query,providing one or more of the search results that are identified as a result of the search of the remote database for output,receiving data indicative of a selection of a particular search result from among the provided search results that are identified as a result of the search of the remote database, andstoring the locally-generated transcription of the user utterance, generated by the local speech recognizer, in association with data identifying the particular search result;

    receiving, after storing the locally-generated transcription of the user utterance, second audio data corresponding to a second user utterance;

    obtaining a locally-generated transcription of the second user utterance, the locally-generated transcription of the second user utterance being generated by the local speech recognizer;

    determining that the locally-generated transcription of the second user utterance matches the stored locally-generated transcription of the user utterance;

    providing the second audio data to the remote speech recognizer after determining that the locally-generated transcription of the second user utterance matches the stored locally-generated transcription of the user utterance and, in response, obtaining several additional search results that are identified as a result of a search of the remote database using at least a portion of a remotely-generated transcription of the second user utterance as a query, the remotely-generated transcription of the second user utterance being generated by the remote speech recognizer; and

    providing the particular search result and the several additional search results that are identified as a result of the search of the remote database for output, based on the data identifying the particular search result that is stored in association with the locally-generated transcription of the user utterance.

View all claims
  • 2 Assignments
Timeline View
Assignment View
    ×
    ×