Communication efficient federated learning
First Claim
1. A computer-implemented method for communication efficient machine learning, the method comprising:
- obtaining, by a client computing device, global values for a set of parameters of a machine-learned model;
training, by the client computing device, the machine-learned model based at least in part on a local dataset to obtain an update matrix that is descriptive of updated values for the set of parameters of the machine-learned model, wherein the update matrix is restricted to be a low-rank matrix, and wherein the local dataset is stored locally by the client computing device; and
communicating, by the client computing device, information descriptive of the update matrix to a server computing device for use by the server computing device in computation of a global update to the machine-learned model, wherein;
training, by the client computing device, the machine-learned model based at least in part on the local dataset to obtain the update matrix comprises;
defining, by the client computing device, the update matrix as a product of a first matrix and a second matrix, wherein the first matrix comprises fixed values and the second matrix comprises optimizable variables, and wherein the fixed values of the first matrix are known to the server computing device; and
training, by the client computing device, machine-learned model based at least in part on the local dataset to obtain the second matrix; and
communicating, by the client computing device, information descriptive of the update matrix to the server computing device comprises communicating, by the client computing device, information descriptive of the second matrix to the server computing device without sending the first matrix from the client computing device to the server computing device.
2 Assignments
0 Petitions
Accused Products
Abstract
The present disclosure provides efficient communication techniques for transmission of model updates within a machine learning framework, such as, for example, a federated learning framework in which a high-quality centralized model is trained on training data distributed overt a large number of clients each with unreliable network connections and low computational power. In an example federated learning setting, in each of a plurality of rounds, each client independently updates the model based on its local data and communicates the updated model back to the server, where all the client-side updates are used to update a global model. The present disclosure provides systems and methods that reduce communication costs. In particular, the present disclosure provides at least: structured update approaches in which the model update is restricted to be small and sketched update approaches in which the model update is compressed before sending to the server.
42 Citations
9 Claims
-
1. A computer-implemented method for communication efficient machine learning, the method comprising:
-
obtaining, by a client computing device, global values for a set of parameters of a machine-learned model; training, by the client computing device, the machine-learned model based at least in part on a local dataset to obtain an update matrix that is descriptive of updated values for the set of parameters of the machine-learned model, wherein the update matrix is restricted to be a low-rank matrix, and wherein the local dataset is stored locally by the client computing device; and communicating, by the client computing device, information descriptive of the update matrix to a server computing device for use by the server computing device in computation of a global update to the machine-learned model, wherein; training, by the client computing device, the machine-learned model based at least in part on the local dataset to obtain the update matrix comprises; defining, by the client computing device, the update matrix as a product of a first matrix and a second matrix, wherein the first matrix comprises fixed values and the second matrix comprises optimizable variables, and wherein the fixed values of the first matrix are known to the server computing device; and training, by the client computing device, machine-learned model based at least in part on the local dataset to obtain the second matrix; and communicating, by the client computing device, information descriptive of the update matrix to the server computing device comprises communicating, by the client computing device, information descriptive of the second matrix to the server computing device without sending the first matrix from the client computing device to the server computing device. - View Dependent Claims (2, 3, 4, 5, 6, 7)
-
-
8. A client computing device, comprising:
-
at least one processor; and at least one non-transitory computer-readable medium that stores instructions that, when executed by the at least one processor, cause the client computing device to perform operations comprising; obtaining global values for a set of parameters of a machine-learned model; training the machine-learned model based at least in part on a local dataset to obtain an update matrix that is descriptive of updated values for the set of parameters of the machine-learned model, wherein the update matrix is restricted to be a low-rank matrix, and wherein the local dataset is stored locally by the client computing device; and communicating information descriptive of the update matrix to a server computing device for use by the server computing device in computation of a global update to the machine-learned model, wherein; training the machine-learned model based at least in part on the local dataset to obtain the update matrix comprises; defining the update matrix as a product of a first matrix and a second matrix, wherein the first matrix comprises fixed values and the second matrix comprises optimizable variables, and wherein the fixed values of the first matrix are known to the server computing device; and training machine-learned model based at least in part on the local dataset to obtain the second matrix; and communicating information descriptive of the update matrix to the server computing device comprises communicating, by the client computing device, information descriptive of the second matrix to the server computing device without sending the first matrix from the client computing device to the server computing device.
-
-
9. At least one non-transitory computer-readable medium that stores instructions that, when executed by a client computing device, cause the client computing device to perform operations comprising:
-
obtaining global values for a set of parameters of a machine-learned model; training the machine-learned model based at least in part on a local dataset to obtain an update matrix that is descriptive of updated values for the set of parameters of the machine-learned model, wherein the update matrix is restricted to be a low-rank matrix, and wherein the local dataset is stored locally by the client computing device; and communicating information descriptive of the update matrix to a server computing device for use by the server computing device in computation of a global update to the machine-learned model, wherein; training the machine-learned model based at least in part on the local dataset to obtain the update matrix comprises; defining the update matrix as a product of a first matrix and a second matrix, wherein the first matrix comprises fixed values and the second matrix comprises optimizable variables, and wherein the fixed values of the first matrix are known to the server computing device; and training machine-learned model based at least in part on the local dataset to obtain the second matrix; and communicating information descriptive of the update matrix to the server computing device comprises communicating, by the client computing device, information descriptive of the second matrix to the server computing device without sending the first matrix from the client computing device to the server computing device.
-
Specification