SPEECH RECOGNITION USING DOCK CONTEXT
2 Assignments
0 Petitions
Accused Products
Abstract
Methods, systems, and apparatuses, including computer programs encoded on a computer storage medium, for performing speech recognition using dock context. In one aspect, a method includes accessing audio data that includes encoded speech. Information that indicates a docking context of a client device is accessed, the docking context being associated with the audio data. A plurality of language models is identified. At least one of the plurality of language models is selected based on the docking context. Speech recognition is performed on the audio data using the selected language model to identify a transcription for a portion of the audio data.
-
Citations
31 Claims
-
1-5. -5. (canceled)
-
6. A computer-implemented method, comprising:
-
accessing first audio data that includes encoded speech; accessing information that indicates a first docking context of a client device, the first docking context being associated with the first audio data; identifying a plurality of language models; determining that the first docking context indicates docking of the client device with a first docking station of a first type; selecting at least a first language model of the plurality of language models based on determining that the first docking context indicates docking of the client device with the first docking station of the first type; performing speech recognition on the first audio data using the selected first language model to identify a transcription for a portion of the first audio data; accessing second audio data that includes encoded speech; accessing information that indicates a second docking context of the client device, the second docking context being associated with the second audio data; determining that the second docking context indicates docking of the client device with a second docking station of a second type, the second type being different from the first type; selecting at least a second language model of the plurality of language models based on determining that the second docking context indicates docking of the client device with the second docking station of the second type, the second language model being different from the first language model; and performing speech recognition on the second audio data using the second language model to identify a transcription for a portion of the second audio data, wherein docking stations of the first type provide capabilities for one or more first manners of using the client device, and wherein docking stations of the second type provide capabilities for one or more second manners of using the client device that are different from the one or more first manners of using the client device. - View Dependent Claims (7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 26, 27, 28, 29, 30)
-
-
17. A system comprising:
-
one or more processors; and a computer-readable medium coupled to the one or more processors having instructions stored thereon which, when executed by the one or more processors, cause the system to perform operations comprising; accessing first audio data that includes encoded speech; accessing information that indicates a first docking context of a client device, the first docking context being associated with the first audio data; identifying a plurality of language models; determining that the first docking context indicates docking of the client device with a first docking station of a first type; selecting at least a first language model of the plurality of language models based on determining that the first docking context indicates docking of the client device with the first docking station of the first type; performing speech recognition on the first audio data using the first language model to identify a transcription for a portion of the first audio data; accessing second audio data that includes encoded speech; accessing information that indicates a second docking context of the client device, the second docking context being associated with the second audio data; determining that the second docking context indicates docking of the client device with a second docking station of a second type, the second type being different from the first type; selecting at least a second language model of the plurality of language models based on determining that the second docking context indicates docking of the client device with the second docking station of the second type, the second language model being different from the first language model; and performing speech recognition on the second audio data using the second language model to identify a transcription for a portion of the second audio data, wherein docking stations of the first type provide capabilities for one or more first manners of using the client device, and wherein docking stations of the second type provide capabilities for one or more second manners of using the client device that are different from the one or more first manners of using the client device. - View Dependent Claims (18, 19, 20)
-
-
21. A non-transitory computer storage medium encoded with a computer program, the program comprising instructions that when executed by one or more computers cause the one or more computers to perform operations comprising:
-
accessing first audio data that includes encoded speech; accessing information that indicates a first docking context of a client device, the first docking context being associated with the first audio data; identifying a plurality of language models; determining that the first docking context indicates docking of the client device with a first docking station of a first type; selecting at least a first language model of the plurality of language models based on determining that the first docking context indicates docking of the client device with the first docking station of the first type; performing speech recognition on the first audio data using the first language model to identify a transcription for a portion of the first audio data; accessing second audio data that includes encoded speech; accessing information that indicates a second docking context of the client device, the second docking context being associated with the second audio data determining that the second docking context indicates docking of the client device with a second docking station of a second type, the second type being different from the first type; selecting at least a second language model of the plurality of language models based on determining that the second docking context indicates docking of the client device with the second docking station of the second type, the second language model being different from the first language model; and performing speech recognition on the second audio data using the second language model to identify a transcription for a portion of the second audio data wherein docking stations of the first type provide capabilities for one or more first manners of using the client device, and wherein docking stations of the second type provide capabilities for one or more second manners of using the client device that are different from the one or more first manners of using the client device. - View Dependent Claims (22, 23, 24, 25)
-
-
31. A computer-implemented method, comprising:
-
accessing first audio data that includes encoded speech; accessing information that indicates a first docking context of a first client device, the first docking context being associated with the first audio data; identifying a plurality of language models; determining that the first docking context indicates docking of the first client device with a first docking station of a first type; selecting at least a first language model of the plurality of language models based on that the first docking context indicates docking of the first client device with the first docking station of the first type; performing speech recognition on the first audio data using the first language model to identify a transcription for a portion of the first audio data; accessing second audio data that includes encoded speech; accessing information that indicates a second docking context of a second client device, the second docking context being associated with the second audio data, the second client device being different from the first client device; determining that the second docking context indicates docking of the second client device with a second docking station of a second type, the second type being different from the first type; selecting at least a second language model of the plurality of language models based on determining that the second docking context indicates docking of the second client device with the second docking station of the second type, the second language model being different from the first language model; and performing speech recognition on the second audio data using the second language model to identify a transcription for a portion of the second audio data, wherein docking stations of the first type provide capabilities for one or more first manners of using a client device, and wherein docking stations of the second type provide capabilities for one or more second manners of using a client device that are different from the one or more first manners of using a client device.
-
Specification