Systems and methods to perform machine learning with feedback consistency
First Claim
1. A computer-implemented method to perform machine learning, the method comprising:
- obtaining, by one or more computing devices, data descriptive of an encoder model that is configured to receive a first set of inputs and, in response to receipt of the first set of inputs, output a first set of outputs;
obtaining, by the one or more computing devices, data descriptive of a decoder model that is configured to receive the first set of outputs and, in response to receipt of the first set of outputs, output a second set of outputs;
determining, by the one or more computing devices, a loss function that describes a difference between the first set of inputs and the second set of outputs;
backpropagating, by the one or more computing devices, the loss function through the decoder model without modifying the decoder model; and
after backpropagating, by the one or more computing devices, the loss function through the decoder model, continuing to backpropagate, by the one or more computing devices, the loss function through the encoder model to train the encoder model;
wherein continuing to backpropagate, by the one or more computing devices, the loss function through the encoder model to train the encoder model comprises adjusting, by the one or more computing devices, at least one weight included in the encoder model.
2 Assignments
0 Petitions
Accused Products
Abstract
The present disclosure provides systems and methods that enable training of an encoder model based on a decoder model that performs an inverse transformation relative to the encoder model. In one example, an encoder model can receive a first set of inputs and output a first set of outputs. The encoder model can be a neural network. The decoder model can receive the first set of outputs and output a second set of outputs. A loss function can describe a difference between the first set of inputs and the second set of outputs. According to an aspect of the present disclosure, the loss function can be sequentially backpropagated through the decoder model without modifying the decoder model and then through the encoder model while modifying the encoder model, thereby training the encoder model. Thus, an encoder model can be trained to have enforced consistency relative to the inverse decoder model.
-
Citations
20 Claims
-
1. A computer-implemented method to perform machine learning, the method comprising:
-
obtaining, by one or more computing devices, data descriptive of an encoder model that is configured to receive a first set of inputs and, in response to receipt of the first set of inputs, output a first set of outputs; obtaining, by the one or more computing devices, data descriptive of a decoder model that is configured to receive the first set of outputs and, in response to receipt of the first set of outputs, output a second set of outputs; determining, by the one or more computing devices, a loss function that describes a difference between the first set of inputs and the second set of outputs; backpropagating, by the one or more computing devices, the loss function through the decoder model without modifying the decoder model; and after backpropagating, by the one or more computing devices, the loss function through the decoder model, continuing to backpropagate, by the one or more computing devices, the loss function through the encoder model to train the encoder model; wherein continuing to backpropagate, by the one or more computing devices, the loss function through the encoder model to train the encoder model comprises adjusting, by the one or more computing devices, at least one weight included in the encoder model. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11)
-
-
12. A computing system to perform machine learning, the computing system comprising:
-
at least one processor; and at least one tangible, non-transitory computer-readable medium that stores instructions that, when executed by the at least one processor, cause the computing system to; obtain data descriptive of a model that comprises an encoder model and a decoder model, wherein the encoder model is configured to receive a first set of inputs and, in response to receipt of the first set of inputs, output a first set of outputs, and wherein the decoder model is configured to receive the first set of outputs and, in response to receipt of the first set of outputs, output a second set of outputs; determine a loss function that describes a difference between the first set of inputs and the second set of outputs; backpropagate the loss function through the decoder model without modifying the decoder model; and after backpropagating the loss function through the decoder model, continue to backpropagate the loss function through the encoder model while modifying the encoder model to train the encoder model. - View Dependent Claims (13, 14, 15, 16, 17)
-
-
18. A computing system, comprising:
-
at least one processor; and at least one memory that stores a machine-learned encoder model that is configured to receive a first set of inputs and output a first set of outputs, the encoder model having been trained by sequentially backpropagating a loss function through a decoder model without modifying the decoder model and then through the encoder model to modify at least one weight of the encoder model, the decoder model configured to receive the first set of outputs and output a second set of outputs, the loss function descriptive of a difference between the first set of inputs and the second set of outputs. - View Dependent Claims (19, 20)
-
Specification