Method and apparatus for distribution-based language model adaptation
First Claim
Patent Images
1. A method of forming a language model, the method comprising:
- selecting out-of-task training data having n-gram distributions;
selecting task-specific training data having n-gram distributions;
modifying an n-gram distribution in the out-of-task training data to form modified training data by applying a weight to an n-gram in the out-of-task training data, the weight formed as;
where Ptask-specific is the relative frequency of an n-gram in the task-specific training data, Pout-of-task is the relative frequency of the n-gram in the out-of-task training data, and α
is an adaptation coefficient; and
identifying probabilities for the language model based on the modified training data.
2 Assignments
0 Petitions
Accused Products
Abstract
A method and apparatus are provided for adapting a language model to a task-specific domain. Under the method and apparatus, the relative frequency of n-grams in a small training set (i.e. task-specific training data set) and the relative frequency of n-grams in a large training set (i.e. out-of-domain training data set) are used to weight a distribution count of n-grams in the large training set. The weighted distributions are then used to form a modified language model by identifying probabilities for n-grams from the weighted distributions.
-
Citations
10 Claims
-
1. A method of forming a language model, the method comprising:
-
selecting out-of-task training data having n-gram distributions; selecting task-specific training data having n-gram distributions; modifying an n-gram distribution in the out-of-task training data to form modified training data by applying a weight to an n-gram in the out-of-task training data, the weight formed as; where Ptask-specific is the relative frequency of an n-gram in the task-specific training data, Pout-of-task is the relative frequency of the n-gram in the out-of-task training data, and α
is an adaptation coefficient; andidentifying probabilities for the language model based on the modified training data. - View Dependent Claims (2, 3, 4, 5)
-
-
6. A tangible computer-readable medium having computer-executable instructions for forming a language model through steps comprising:
-
determining a distribution of entities in a small set of training data; changing a distribution of entities in a large set of training data based on the distribution of entities in the small set of training data to form a modified distribution of entities by applying a weight to a count of entities in the large set of training data, the weight being a function of; where Psmall-set is a relative frequency of an entity in the small set of training data, Plarge-set is a relative frequency of the entity in the large set of training data, and α
is an adaptation coefficient; andusing the modified distribution of entities to identify probabilities for the language model. - View Dependent Claims (7, 8, 9, 10)
-
Specification