Training Method, Apparatus, and Chip for Neural Network Model
First Claim
Patent Images
1. A training method for a neural network model applied to a training system, wherein the training method comprises:
- determining, by each of at least one M processor cores for each layer of L layers of the neural network model, a model training mode of a layer of the L layers based on an estimated data volume in a model parameter set and an estimated data volume of output data of the layer, wherein the training system comprises the M processor cores, and wherein M and L are integers greater than or equal to 1; and
performing, by each of the M processor cores, training to the layer using a determined model training mode, wherein the determined model training mode comprises at least one of a data parallel training mode or a model parallel training mode.
1 Assignment
0 Petitions
Accused Products
Abstract
A training method, apparatus, and chip for a neural network model includes determining a model training mode of each layer based on an estimated data volume in a model parameter set and an estimated data volume of output data of the layer, obtaining second output data that is obtained by m worker modules by training a (j−1)th layer, and directly obtaining by a worker module a global gradient of a model parameter by training the model parameter based on the second output data when a model parallel training mode is used for a jth layer.
-
Citations
20 Claims
-
1. A training method for a neural network model applied to a training system, wherein the training method comprises:
-
determining, by each of at least one M processor cores for each layer of L layers of the neural network model, a model training mode of a layer of the L layers based on an estimated data volume in a model parameter set and an estimated data volume of output data of the layer, wherein the training system comprises the M processor cores, and wherein M and L are integers greater than or equal to 1; and performing, by each of the M processor cores, training to the layer using a determined model training mode, wherein the determined model training mode comprises at least one of a data parallel training mode or a model parallel training mode. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
-
8. A training apparatus for a neural network model, wherein the training apparatus comprises:
-
a memory configured to store instructions; a processor coupled to the memory and configured to execute the instructions, wherein the processor comprises at least one processor core; and a transceiver coupled to the processor and the memory, wherein the training apparatus is applicable to a training system that comprises M processor cores, wherein the neural network model comprises L layers, wherein M and L are integers greater than or equal to 1, wherein for each layer of the L layers, the at least one processor core is used to train the layer, wherein the processor is configured to control the transceiver to transmit data to a second processor core in the M processor cores, and wherein the instructions cause each of the at least one processor core to be configured to; determining, a model training mode of the layer based on an estimated data volume in a model parameter set and an estimated data volume of output data of the layer, wherein the training system comprises at least one M processor cores; and performing, an training to the layer using a determined training mode, wherein the determined model training mode comprises at least one of a data parallel training mode or a model parallel training mode. - View Dependent Claims (9, 10, 11, 12, 13, 14)
-
-
15. A training chip for a neural network model, applicable to a training system that comprises M chips, wherein the neural network model comprises L layers, wherein each of the M chips comprises at least one processor core, and wherein each of the at least one chip is configured to:
-
determine, by each of at least one M processor cores for each layer of L layers of the neural network model, a model training mode of a layer of the L layers based on an estimated data volume in a model parameter set and an estimated data volume of output data of the layer, wherein the training system comprises at least one M processor cores, and wherein M and L are integers greater than or equal to 1; and perform, by each of the M processor cores, an training to the layer using a determined training mode, wherein the determined model training mode comprises at least one of a data parallel training mode or a model parallel training mode. - View Dependent Claims (16, 17, 18, 19, 20)
-
Specification