Methods and Systems for Automated Text Correction
First Claim
Patent Images
1. An apparatus, comprising:
- at least one processor and a memory device coupled to the at least one processor, in which the at least one processor is configured;
to identify words of an input utterance;
to place the words in a plurality of first nodes stored in the memory device;
to assign a word-layer tag to each of the plurality of first nodes based, in part, on neighboring nodes of the plurality of first nodes; and
to generate an output sentence by combining words from the plurality of first nodes with punctuation marks selected, in part, on the word-layer tags assigned to each of the first nodes.
0 Assignments
0 Petitions
Accused Products
Abstract
The present embodiments demonstrate systems and methods for automated text correction. In certain embodiments, the methods and systems may be implemented through analysis according to a single text correction model. In a particular embodiment, the single text correction model may be generated through analysis of both a corpus of learner text and a corpus of non-learner text.
-
Citations
74 Claims
-
1. An apparatus, comprising:
at least one processor and a memory device coupled to the at least one processor, in which the at least one processor is configured; to identify words of an input utterance; to place the words in a plurality of first nodes stored in the memory device; to assign a word-layer tag to each of the plurality of first nodes based, in part, on neighboring nodes of the plurality of first nodes; and to generate an output sentence by combining words from the plurality of first nodes with punctuation marks selected, in part, on the word-layer tags assigned to each of the first nodes. - View Dependent Claims (2, 3, 4, 5)
-
6-7. -7. (canceled)
-
8. A computer program product, comprising:
a non-transitory computer-readable medium comprising; code to identify words of an input utterance; code to place the words in a plurality of first nodes stored in the memory device; code to assign a word-layer tag to each of the plurality of first nodes based, in part, on neighboring nodes of the plurality of first nodes; and code to generate an output sentence by combining words from the plurality of first nodes with punctuation marks selected, in part, on the word-layer tags assigned to each of the first nodes. - View Dependent Claims (9, 10, 11, 12)
-
13-33. -33. (canceled)
-
34. An apparatus, comprising:
at least one processor and a memory device coupled to the at least one processor, in which the at least one processor is configured; to receive a natural language text input, the text input comprising a grammatical error in which a portion of the input text comprises a class from a set of classes; to generate a plurality of selection tasks from a corpus of non-learner text that is assumed to be free of grammatical errors, wherein for each selection task a classifier re-predicts a class used in the non-learner text; to generate a plurality of correction tasks from a corpus of learner text, wherein for each correction task a classifier proposes a class used in the learner text; to train a grammar correction model using a set of binary classification problems that include the plurality of selection tasks and the plurality of correction tasks; and to use the trained grammar correction model to predict a class for the text input from the set of possible classes. - View Dependent Claims (35, 36, 37, 39, 41, 42, 43)
-
38. (canceled)
-
40. (canceled)
-
44-55. -55. (canceled)
-
56. An apparatus, comprising at least one processor and a memory device coupled to the at least one processor, in which the at least one processor is configured to correct semantic collection errors by performing the steps of:
-
automatically identifying one or more translation candidates in response to analysis of a corpus of parallel-language text conducted in a processing device; determining, using the processing device, a feature associated with each translation candidate; generating a set of one or more weight values from a corpus of learner text stored in a data storage device; and calculating, using a processing device, a score for each of the one or more translation candidates in response to the feature associated with each translation candidate and the set of one or more weight values. - View Dependent Claims (58, 59, 60, 61)
-
-
57. (canceled)
-
62. A non-transitory tangible computer-readable medium comprising computer-readable code that, when executed by a computer, cause the computer to perform the operation of correcting semantic collocation errors comprising:
-
automatically identifying one or more translation candidates in response to analysis of a corpus of parallel-language text conducted in a processing device; determining, using the processing device, a feature associated with each translation candidate; generating a set of one or more weight values from a corpus of learner text stored in a data storage device; and calculating, using a processing device, a score for each of the one or more translation candidates in response to the feature associated with each translation candidate and the set of one or more weight values. - View Dependent Claims (63, 64, 65, 66)
-
-
67. A non-transitory tangible computer-readable medium comprising computer-readable code that, when executed by a computer, cause the computer:
-
to receive a natural language text input, the text input comprising a grammatical error in which a portion of the input text comprises a class from a set of classes; to generate a plurality of selection tasks from a corpus of non-learner text that is assumed to be free of grammatical errors, wherein for each selection task a classifier re-predicts a class used in the non-learner text; to generate a plurality of correction tasks from a corpus of learner text, wherein for each correction task a classifier proposes a class used in the learner text; to train a grammar correction model using a set of binary classification problems that include the plurality of selection tasks and the plurality of correction tasks; and to use the trained grammar correction model to predict a class for the text input from the set of possible classes. - View Dependent Claims (68, 69, 70, 71, 72, 73, 74)
-
Specification