CROSS-LINGUAL DISCRIMINATIVE LEARNING OF SEQUENCE MODELS WITH POSTERIOR REGULARIZATION
First Claim
1. A computer-implemented method, comprising:
- obtaining, at a computing device having one or more processors, (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language;
labeling, at the computing device, a source side of the aligned bi-text using the supervised sequence model to obtain a labeled source side of the aligned bi-text;
projecting, at the computing device, labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text;
filtering, at the computing device, the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text; and
training, at the computing device, the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language.
2 Assignments
0 Petitions
Accused Products
Abstract
A computer-implemented method can include obtaining (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language. The method can include labeling a source side of the aligned bi-text using the supervised sequence model and projecting labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text. The method can include filtering the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text. The method can also include training the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language.
-
Citations
20 Claims
-
1. A computer-implemented method, comprising:
-
obtaining, at a computing device having one or more processors, (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language; labeling, at the computing device, a source side of the aligned bi-text using the supervised sequence model to obtain a labeled source side of the aligned bi-text; projecting, at the computing device, labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text; filtering, at the computing device, the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text; and training, at the computing device, the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A computing device comprising one or more processors configured to perform operations comprising:
-
obtaining (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language; labeling a source side of the aligned bi-text using the supervised sequence model to obtain a labeled source side of the aligned bi-text; projecting labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text; filtering the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text; and training the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17, 18)
-
-
19. A non-transitory, computer-readable medium having instructions stored thereon that, when executed by one or more processors of a computing device, cause the computing device to perform operations comprising:
-
obtaining (i) an aligned bi-text for a source language and a target language, and (ii) a supervised sequence model for the source language; labeling a source side of the aligned bi-text using the supervised sequence model to obtain a labeled source side of the aligned bi-text; projecting labels from the labeled source side to a target side of the aligned bi-text to obtain a labeled target side of the aligned bi-text; filtering the labeled target side based on a task of a natural language processing (NLP) system configured to utilize a sequence model for the target language to obtain a filtered target side of the aligned bi-text; and training the sequence model for the target language using posterior regularization with soft constraints on the filtered target side to obtain a trained sequence model for the target language. - View Dependent Claims (20)
-
Specification