Ordered iteration for data update management
First Claim
1. A computer-implemented method of managing data in a networked environment, comprising:
- under control of one or more computer systems configured with executable instructions,receiving workload data for a designated host server of a group of host servers selected to process the workload data for a customer, the group of host servers being connected by a network aggregation fabric including layers of network switches, a path across the aggregation fabric to each host server involving a number of connections across the network switches, the group of host servers being selected to process the workload data for the customer and being dispersed across a number of network switches for at least a lowest layer of the aggregation fabric;
routing the workload data to the designated host server and processing the workload data using the designated host server;
measuring one or more transmission patterns of the workload;
determining an ordering of other host servers in the group to which to send updates to the workload data based upon the measured one or more transmission patterns, wherein the determined ordering is selected in order to statistically minimize a likelihood of network congestion based on known transmission patterns of the workload, and each host server in the group capable of having a different ordering; and
in response to determining one or more updates to the workload data to be sent to the other host servers in the group, sending the updates to the other host servers according to the determined ordering,wherein updates to be periodically shared across all the host servers in the group are sent with determined orderings in order to reduce a statistical likelihood of network congestion due to flow convergence.
1 Assignment
0 Petitions
Accused Products
Abstract
Host machines and other devices performing synchronized operations can be dispersed across multiple racks in a data center to provide additional buffer capacity and to reduce the likelihood of congestion. The level of dispersion can depend on factors such as the level of oversubscription, as it can be undesirable in a highly connected network to push excessive host traffic into the aggregation fabric. As oversubscription levels increase, the amount of dispersion can be reduced and two or more host machines can be clustered on a given rack, or otherwise connected through the same edge switch. By clustering a portion of the machines, some of the host traffic can be redirected by the respective edge switch without entering the aggregation fabric. When provisioning hosts for a customer, application, or synchronized operation, for example, the levels of clustering and dispersion can be balanced to minimize the likelihood for congestion throughout the network.
-
Citations
24 Claims
-
1. A computer-implemented method of managing data in a networked environment, comprising:
-
under control of one or more computer systems configured with executable instructions, receiving workload data for a designated host server of a group of host servers selected to process the workload data for a customer, the group of host servers being connected by a network aggregation fabric including layers of network switches, a path across the aggregation fabric to each host server involving a number of connections across the network switches, the group of host servers being selected to process the workload data for the customer and being dispersed across a number of network switches for at least a lowest layer of the aggregation fabric; routing the workload data to the designated host server and processing the workload data using the designated host server; measuring one or more transmission patterns of the workload; determining an ordering of other host servers in the group to which to send updates to the workload data based upon the measured one or more transmission patterns, wherein the determined ordering is selected in order to statistically minimize a likelihood of network congestion based on known transmission patterns of the workload, and each host server in the group capable of having a different ordering; and in response to determining one or more updates to the workload data to be sent to the other host servers in the group, sending the updates to the other host servers according to the determined ordering, wherein updates to be periodically shared across all the host servers in the group are sent with determined orderings in order to reduce a statistical likelihood of network congestion due to flow convergence. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23)
-
-
24. A computer-implemented method of managing data in a networked environment, comprising:
-
under control of one or more computer systems configured with executable instructions, receiving workload data for one of a group of host servers selected to process the workload data for a customer, the group of host servers being connected by a network aggregation fabric including layers of network switches, a path across the aggregation fabric to each host server involving a number of connections across the network switches, the group of host servers being selected to process the workload data for the customer and being dispersed across a number of network switches for at least a lowest layer of the aggregation fabric; determining an absolute ordering of the group of host servers selected to process workload data for the customer; routing the workload data to a currently selected host server in the absolute ordering and processing the workload data using the processing host server; measuring one or more transmission patterns of the workload; determining an ordering of other host servers in the group to which to send updates to the workload data based upon the measured one or more transmission patterns, wherein the determined ordering is selected in order to statistically minimize a likelihood of network congestion based on known transmission patterns of the workload, and each host server in the group capable of having a different ordering; and in response to determining one or more updates to the workload data to be sent to the other host servers in the group, sending the updates to the other host servers according to the determined ordering, wherein updates to be periodically shared across all the host servers in the group are sent with determined orderings in order to reduce a statistical likelihood of network congestion due to flow convergence.
-
Specification