Use of a unified language model
First Claim
1. A computer readable medium including instructions readable by a computer which, when implemented execute a method to perform language processing for recognizing language and providing an output signal indicative thereof, the method comprising:
- receiving an input signal indicative of language;
accessing a unified language model to recognize the language and predict non-terminal tokens contained therein, the unified language model comprising a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals, and a N-gram language model having the non-terminal tokens; and
generating hypotheses for the language by exploring each of the terminals in the unified language model associated with the non-terminal tokens predicted based on a probability value for each terminal, wherein at least one terminal has a different probability value than one other terminal in the same context-free grammar.
1 Assignment
0 Petitions
Accused Products
Abstract
A language processing system includes a unified language model. The unified language model comprises a plurality of context-free grammars having non-terminal tokens representing semantic or syntactic concepts and terminals, and an N-gram language model having non-terminal tokens. A language processing module capable of receiving an input signal indicative of language accesses the unified language model to recognize the language. The language processing module generates hypotheses for the received language as a function of words of the unified language model and/or provides an output signal indicative of the language and at least some of the semantic or syntactic concepts contained therein.
-
Citations
18 Claims
-
1. A computer readable medium including instructions readable by a computer which, when implemented execute a method to perform language processing for recognizing language and providing an output signal indicative thereof, the method comprising:
-
receiving an input signal indicative of language;
accessing a unified language model to recognize the language and predict non-terminal tokens contained therein, the unified language model comprising a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals, and a N-gram language model having the non-terminal tokens; and
generating hypotheses for the language by exploring each of the terminals in the unified language model associated with the non-terminal tokens predicted based on a probability value for each terminal, wherein at least one terminal has a different probability value than one other terminal in the same context-free grammar. - View Dependent Claims (2, 3, 4, 14)
-
-
5. A language processing system comprising:
-
a unified language model comprising;
a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals; and
a N-gram language model having the non-terminal tokens; and
a language processing module capable of receiving an input signal indicative of language and accessing the unified language model to recognize the language and predict non-terminal tokens contained therein, the language processing module further adapted to generate hypotheses for the language by exploring each of the terminals in the unified language model associated with the non-terminal tokens predicted based on a probability value for each terminal, wherein at least one terminal has a different probability value than one other terminal in the same context-free grammar. - View Dependent Claims (6, 7, 8)
-
-
9. A method to perform language processing comprising:
-
receiving an input signal indicative of language;
accessing a unified language model to recognize the language and predict non-terminal tokens contained therein, the unified language model comprising a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals wherein each of the terminals of the plurality of context-free grammars include a probability value, and a N-gram language model having the non-terminal tokens;
assigning probability values of at least some of the terminals of the context-free grammars from a terminal-based language model, wherein at least one terminal has a probability value different than one other terminal in the same context-free grammar and normalizing said values using the set of terminals constrained by the context-free grammars;
generating hypotheses for the language as a function of words in the unified language model corresponding to the non-terminal tokens predicted; and
calculating a language model score for each of the hypotheses using the associated probability value for each terminal present therein and obtained from the plurality of context-free grammars. - View Dependent Claims (10)
-
-
11. A language processing system comprising:
-
a unified language model comprising;
a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals, wherein each of the terminals of the plurality of context-free grammars include a probability value; and
a N-gram language model having the non-terminal tokens; and
a language processing module capable of receiving an input signal indicative of language and accessing the unified language model to recognize the language and predict non-terminal tokens contained therein, the language processing module further adapted to assign probability values of at least some of the terminals of the context-free grammars from a terminal-based language model, wherein at least one terminal has a probability value different than one other terminal in the same context-free grammar and normalize said values using the set of terminals constrained by the context-free grammars, and adapted to generate hypotheses for the language as a function of words in the unified language model corresponding to the non-terminal tokens predicted and calculate a language model score for each of the hypotheses using the associated probability value for each terminal present therein and obtained from the plurality of context-free grammars. - View Dependent Claims (12)
-
-
13. A computer readable medium having instructions to process information, the instructions comprising:
-
a unified language model comprising;
a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals, wherein each of the terminals include a probability value assigned by using non-uniform probability values derived from a terminal based language model and normalizing said values using the set of terminals constrained by the plurality of context-free grammars; and
a N-gram language model having the non-terminal tokens; and
a language processing module capable of receiving an input signal indicative of language and accessing the unified language model to recognize the language and predict non-terminal tokens contained therein, the language processing module further generating hypotheses for the received language as a function of words in the unified language model corresponding to the non-terminal tokens predicted and calculating a language model score for each of the hypotheses using the associated probability value for each terminal present therein and obtained from the plurality of context-free grammars.
-
-
15. A method to perform language processing comprising:
-
receiving an input signal indicative of language;
accessing a unified language model to recognize the language and predict non-terminal tokens contained therein, the unified language model comprising a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals, wherein each of the terminals include a probability value assigned by using non-uniform probability values derived from a terminal based language model, said values being normalized using the set of terminals constrained by the plurality of context-free grammars, and a N-gram language model having the non-terminal tokens; and
generating hypotheses for the received language as a function of words in the unified language model corresponding to the non-terminal tokens predicted and calculating a language model score for each of the hypotheses using the associated probability value for each terminal present therein and obtained from the plurality of context-free grammars. - View Dependent Claims (16)
-
-
17. A computer readable medium including instructions readable by a computer which, when implemented execute a method to perform language processing for recognizing language and providing an output signal indicative thereof, the method comprising:
-
receiving an input signal indicative of language;
accessing a unified language model to recognize the language and predict non-terminal tokens contained therein, the unified language model comprising a plurality of context-free grammars comprising non-terminal tokens representing semantic or syntactic concepts and terminals, wherein each of the terminals include a probability value assigned by using non-uniform probability values derived from a terminal based language model, said values being normalized using the set of terminals constrained by the plurality of context-free grammars, and a N-gram language model having the non-terminal tokens; and
generating hypotheses for the received language as a function of words in the unified language model corresponding to the non-terminal tokens predicted and calculating a language model score for each of the hypotheses using the associated probability value for each terminal present therein and obtained from the plurality of context-free grammars. - View Dependent Claims (18)
-
Specification