Proxy-based cache content distribution and affinity
First Claim
Patent Images
1. A distribution network for retrieving cached content comprising:
- a load balancer that receives requests from one or more client computing systems and, without being aware of where any content corresponding to the content requests may reside in the distribution network, the load balancer routing said content requests to one or more edge caching servers;
a plurality of edge caching servers each comprising;
an edge caching unit having at least one hardware processor and memory for caching content obtained from one or more origin servers; and
an edge routing unit which receives content requests from the load balancer, the edge routing unit comprising at least one hardware processor and memory containing computer-executable instructions which, when implemented, cause the edge routing unit to process one or more content requests received from the load balancer according to a computer-implemented predictive routing protocol which predicts where cached content corresponding to requested content resides in the distribution network, and wherein the computer-implemented predictive routing protocol comprises acts of;
identifying which of the plurality of edge caching servers is correlated to the requested content based on the predictive routing protocol; and
routing the content request to the identified edge caching server without having to first determine whether the edge routing server contains the requested content.
2 Assignments
0 Petitions
Accused Products
Abstract
A distributed caching hierarchy that includes multiple edge routing servers, at least some of which receiving content requests from client computing systems via a load balancer. When receiving a content request, an edge routing server identifies which of the edge caching servers the requested content would be in if the requested content were to be cached within the edge caching servers, and distributes the content request to the identified edge caching server in a deterministic and predictable manner to increase the likelihood of increasing a cache-hit ratio.
170 Citations
20 Claims
-
1. A distribution network for retrieving cached content comprising:
-
a load balancer that receives requests from one or more client computing systems and, without being aware of where any content corresponding to the content requests may reside in the distribution network, the load balancer routing said content requests to one or more edge caching servers; a plurality of edge caching servers each comprising; an edge caching unit having at least one hardware processor and memory for caching content obtained from one or more origin servers; and an edge routing unit which receives content requests from the load balancer, the edge routing unit comprising at least one hardware processor and memory containing computer-executable instructions which, when implemented, cause the edge routing unit to process one or more content requests received from the load balancer according to a computer-implemented predictive routing protocol which predicts where cached content corresponding to requested content resides in the distribution network, and wherein the computer-implemented predictive routing protocol comprises acts of; identifying which of the plurality of edge caching servers is correlated to the requested content based on the predictive routing protocol; and routing the content request to the identified edge caching server without having to first determine whether the edge routing server contains the requested content. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9)
-
-
10. A computer-implemented method for retrieving cached content from a hierarchy of edge caching servers, the computer-implemented method being performed by one or more processors executing computer executable instructions for the computer-implemented method, and the computer-implemented method comprising:
-
receiving at a load balancer one or more requests from one or more client computing systems; without being aware of where any content corresponding to content requests may reside in the distribution network, the load balancer routing said content requests to one or more edge caching servers; a plurality of the edge caching servers each comprising an edge caching unit having at least one hardware processor and memory for caching content obtained from one or more origin servers, and an edge routing unit, and wherein the computer implemented method further comprises; receiving at one or more edge routing units, content requests from the load balancer; each edge routing unit comprising at least one hardware processor and memory containing computer-executable instructions which, when implemented, cause the edge routing unit to process one or more content requests received from the load balancer according to a computer-implemented predictive routing protocol which predicts where cached content corresponding to requested content resides in the distribution network, and wherein the computer-implemented predictive routing protocol comprises acts of; identifying which of the plurality of edge caching servers is correlated to the requested content based on the predictive routing protocol; and routing the content request to the identified edge caching server without having to first determine whether the edge routing server contains the requested content. - View Dependent Claims (11, 12, 13, 14, 15, 16, 17, 18)
-
-
19. One or more hardware storage devices containing computer-executable instructions which, when executed by one or more processors, cause the one or more processors to perform a computer-implemented method for retrieving cached content from a hierarchy of edge caching servers, and wherein the computer-implemented method comprises:
- receiving at a load balancer one or more requests from one or more client computing systems;
without being aware of where any content corresponding to content requests may reside in the distribution network, the load balancer routing said content requests to one or more edge caching servers;
a plurality of the edge caching servers each comprising an edge caching unit having at least one hardware processor and memory for caching content obtained from one or more origin servers, and an edge routing unit, and wherein the computer implemented method further comprises;
receiving at one or more edge routing units, content requests from the load balancer;
each edge routing unit comprising at least one hardware processor and memory containing computer-executable instructions which, when implemented, cause the edge routing unit to process one or more content requests received from the load balancer according to a computer-implemented predictive routing protocol which predicts where cached content corresponding to requested content resides in the distribution network, and wherein the computer-implemented predictive routing protocol comprises acts of;
identifying which of the plurality of edge caching servers is correlated to the requested content based on the predictive routing protocol; and
routing the content request to the identified edge caching server without having to first determine whether the edge routing server contains the requested content. - View Dependent Claims (20)
- receiving at a load balancer one or more requests from one or more client computing systems;
Specification