Method and apparatus for distribution-based language model adaptation
First Claim
Patent Images
1. A method of forming a language model, the method comprising:
- selecting out-of-task training data having n-gram counts;
selecting task-specific training data having n-gram counts;
modifying an n-gram count in the out-of-task training data by applying an n-gram-specific weight that is based in part on an n-gram count in the task-specific training data to form modified training data; and
identifying probabilities for the language model based on the modified training data.
1 Assignment
0 Petitions
Accused Products
Abstract
A method and apparatus are provided for adapting a language model to a task-specific domain. Under the method and apparatus, the relative frequency of n-grams in a small training set (i.e. task-specific training data set) and the relative frequency of n-grams in a large training set (i.e. out-of-domain training data set) are used to weight a distribution count of n-grams in the large training set. The weighted distributions are then used to form a modified language model by identifying probabilities for n-grams from the weighted distributions.
49 Citations
12 Claims
-
1. A method of forming a language model, the method comprising:
-
selecting out-of-task training data having n-gram counts;
selecting task-specific training data having n-gram counts;
modifying an n-gram count in the out-of-task training data by applying an n-gram-specific weight that is based in part on an n-gram count in the task-specific training data to form modified training data; and
identifying probabilities for the language model based on the modified training data. - View Dependent Claims (2, 3, 4, 5)
-
-
6. A computer-readable medium having computer-executable instructions for forming a language model through steps comprising:
-
determining a distribution of entities in a small set of training data;
changing a distribution of entities in a large set of training data based on an entity-specific weight that is formed in part from a distribution of entities in the small set of training data to form a modified distribution of entities; and
using the modified distribution of entities to identify probabilities for the language model. - View Dependent Claims (7, 8, 9, 10)
-
-
11. A method of adapting a general language model to cover a task-specific domain, the method comprising:
-
determining a weight based on the relative frequency of an n-gram in the task-specific domain;
multiplying the weight by a count of the n-gram in a distribution domain associated with the general language model to form a modified count, the modified count formed without adding counts of n-grams in the task-specific domain;
using only the modified counts of n-grams to determine a probability that forms part of an adapted language model. - View Dependent Claims (12)
-
Specification