Method for processing speech data for a distributed recognition system
DCFirst Claim
Patent Images
1. A method of processing speech data from an utterance for a distributed speech query recognition system comprising the steps of:
- establishing a network connection between a server computing system and a client device suitable for transporting a streaming communication;
receiving a continuous speech byte data stream containing speech data processed by a first component of the distributed speech query recognition system situated in the client device;
wherein said speech data is characterized by a form and data content representing only a partial recognition of an utterance;
further wherein said data stream includes NULL data used to identify a silence in speech data from said client device said NULL data being inserted at the client device after other NULL data is removed prior to transmission of the speech byte data stream; and
further processing said speech data at a second component of the distributed speech query recognition system situated at said server computing system to generate additional speech related content and complete recognition of words in said speech data.
1 Assignment
Litigations
0 Petitions
Accused Products
Abstract
Speech signal information is formatted, processed and transported in accordance with a format adapted for TCP/IP protocols used on the Internet and other communications networks. NULL characters are used for indicating the end of a voice segment. The method is useful for distributed speech recognition systems such as a client-server system, typically implemented on an intranet or over the Internet based on user queries at his/her computer, a PDA, or a workstation using a speech input interface.
790 Citations
20 Claims
-
1. A method of processing speech data from an utterance for a distributed speech query recognition system comprising the steps of:
-
establishing a network connection between a server computing system and a client device suitable for transporting a streaming communication; receiving a continuous speech byte data stream containing speech data processed by a first component of the distributed speech query recognition system situated in the client device; wherein said speech data is characterized by a form and data content representing only a partial recognition of an utterance; further wherein said data stream includes NULL data used to identify a silence in speech data from said client device said NULL data being inserted at the client device after other NULL data is removed prior to transmission of the speech byte data stream; and further processing said speech data at a second component of the distributed speech query recognition system situated at said server computing system to generate additional speech related content and complete recognition of words in said speech data. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18)
-
-
19. A method of processing speech data for a distributed speech query recognition system comprising the steps of:
-
establishing a network connection between a server computing system and a client device suitable for transporting a streaming communication; receiving a data stream containing speech vector data from the client device, said speech vector data representing acoustic features of speech data and being characterized by a form and data content insufficient to recognize words; wherein said data stream includes NULL data information used to identify a silence in speech data from said client device, said NULL data being inserted at the client device after other NULL data is removed prior to transmission of the data stream; and further processing said speech vector data at said server computing system to generate additional speech feature related content and identify words in said speech data.
-
-
20. A method of processing speech data for a distributed speech query recognition system comprising the steps of:
-
establishing a network connection suitable for transporting a streaming communication between a server computing system and a client device; configuring speech processing operations to be performed by said client device and server computing system respectively; wherein said speech processing operations are automatically configured based on computing capabilities of said client device and server computing system respectively, and such that said server computing system supports a number of client devices having different computing capabilities; receiving a data stream containing speech vector data from the client device, said speech vector data representing acoustic features of speech data and being characterized by a data content insufficient to recognize words; wherein said data stream includes at least some NULL data used to identify a silence in speech data from said client device said NULL data being inserted at the client device after other NULL data is removed prior to transmission of the data stream; and further processing said speech vector data at said server computing system to generate additional speech feature related content and identify words in said speech data.
-
Specification