SYSTEMS AND METHODS FOR LAYERED TRAINING IN MACHINE-LEARNING ARCHITECTURES
First Claim
1. A computer-implemented method for layered training of machine-learning architectures, the method implemented by a training computing device including a processor coupled to a memory, the method comprising:
- receiving a plurality of data elements wherein each data element is associated with a timestampdetermining a training window for each model layer of a layered stack of model layers;
determining a plurality of training data elements for each training window by identifying the data elements with timestamps corresponding to each of the training windows;
identifying a previous checkpoint for each model layer if the previous checkpoint for each model layer exists, wherein the previous checkpoint for each model layer is generated by a parent model layer;
training each model layer with the determined training data elements for each model layer and the identified previous checkpoint, if any, for each model layer;
generating a plurality of current checkpoints, wherein each current checkpoint of the plurality of current checkpoints is associated with a model layer; and
;
storing the plurality of current checkpoints at the memory.
2 Assignments
0 Petitions
Accused Products
Abstract
A computer-implemented method for layered training of machine-learning architectures includes receiving a plurality of data elements wherein each data element is associated with a timestamp, determining a training window for each model layer of a layered stack of model layers, determining a plurality of training data elements for each training window by identifying the data elements with timestamps corresponding to each of the training windows, identifying a previous checkpoint for each model layer wherein the previous checkpoint for each model layer is generated by a parent model layer, training each model layer with the determined training data elements for each model layer and the identified previous checkpoint for each model layer, generating a plurality of current checkpoints wherein each current checkpoint of the plurality of current checkpoints is associated with a model layer, and storing the plurality of current checkpoints at the memory.
36 Citations
20 Claims
-
1. A computer-implemented method for layered training of machine-learning architectures, the method implemented by a training computing device including a processor coupled to a memory, the method comprising:
-
receiving a plurality of data elements wherein each data element is associated with a timestamp determining a training window for each model layer of a layered stack of model layers; determining a plurality of training data elements for each training window by identifying the data elements with timestamps corresponding to each of the training windows; identifying a previous checkpoint for each model layer if the previous checkpoint for each model layer exists, wherein the previous checkpoint for each model layer is generated by a parent model layer; training each model layer with the determined training data elements for each model layer and the identified previous checkpoint, if any, for each model layer; generating a plurality of current checkpoints, wherein each current checkpoint of the plurality of current checkpoints is associated with a model layer; and
;storing the plurality of current checkpoints at the memory. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8)
-
-
9. A training computing device for layered training of machine-learning architectures, the training computing device comprising a memory for storing data, and a processor in communication with the memory, said processor programmed to:
-
receive a plurality of data elements wherein each data element is associated with a timestamp; determine a training window for each model layer of a layered stack of model layers; determine a plurality of training data elements for each training window by identifying the data elements with timestamps corresponding to each of the training windows; identify a previous checkpoint for each model layer if the previous checkpoint for each model layer exists, wherein the previous checkpoint for each model layer is generated by a parent model layer; train each model layer with the determined training data elements for each model layer and the identified previous checkpoint, if any, for each model layer; generate a plurality of current checkpoints, wherein each current checkpoint of the plurality of current checkpoints is associated with a model layer; and store the plurality of current checkpoints at the memory. - View Dependent Claims (10, 11, 12, 13, 14, 15, 16)
-
-
17. A computer-readable storage device, having processor-executable instructions embodied thereon, for layered training of machine-learning architectures, wherein the computer includes at least one processor and a memory coupled to the processor, wherein, when executed by the computer, the processor-executable instructions cause the computer to:
-
receive a plurality of data elements wherein each data element is associated with a timestamp; determine a training window for each model layer of a layered stack of model layers; determine a plurality of training data elements for each training window by identifying the data elements with timestamps corresponding to each of the training windows; identify a previous checkpoint for each model layer if the previous checkpoint for each model layer exists, wherein the previous checkpoint for each model layer is generated by a parent model layer; train each model layer with the determined training data elements for each model layer and the identified previous checkpoint, if any, for each model layer; generate a plurality of current checkpoints, wherein each current checkpoint of the plurality of current checkpoints is associated with a model layer; and store the plurality of current checkpoints at the memory. - View Dependent Claims (18, 19, 20)
-
Specification