Patent classifications
H04L67/2885
CONTENT DELIVERY NETWORK OPTIMIZATION SYSTEM
A CDN traffic is optimized by a client-side system that maps the servers in the CDN system. Content requests from client devices for domain names are forwarded to servers in the CDN system that may be selected from the map to prevent a cache miss in the a server for a particular request for content.
Opportunistic content delivery using delta coding
Systems and methods are described for avoiding redundant data transfers using delta coding techniques when reliably and opportunistically communicating data to multiple user systems. According to embodiments, user systems track received block sequences for locally stored content blocks. An intermediate server intercepts content requests between user systems and target hosts, and deterministically chucks and fingerprints content data received in response to those requests. A fingerprint of a received content block is communicated to the requesting user system, and the user system determines based on the fingerprint whether the corresponding content block matches a content block that is already locally stored. If so, the user system returns a set of fingerprints representing a sequence of next content blocks that were previously stored after the matching content block. The intermediate server can then send only those content data blocks that are not already locally stored at the user system according to the returned set of fingerprints.
Opportunistic content delivery using delta coding
Systems and methods are described for avoiding redundant data transfers using delta coding techniques when reliably and opportunistically communicating data to multiple user systems. According to embodiments, user systems track received block sequences for locally stored content blocks. An intermediate server intercepts content requests between user systems and target hosts, and deterministically chucks and fingerprints content data received in response to those requests. A fingerprint of a received content block is communicated to the requesting user system, and the user system determines based on the fingerprint whether the corresponding content block matches a content block that is already locally stored. If so, the user system returns a set of fingerprints representing a sequence of next content blocks that were previously stored after the matching content block. The intermediate server can then send only those content data blocks that are not already locally stored at the user system according to the returned set of fingerprints.
CACHE MANAGEMENT IN CONTENT DELIVERY SYSTEMS
Examples described herein relate to apparatuses and methods for managing caching for a content delivery system, which may include receiving a content request indicating that the caching agent is requesting content data for a client, filling the content data in a first cache storage of the business logic agent, providing the cached content data to the caching agent, and while a second cache storage of the caching agent is being filled with the content data, maintaining the cached content data in response to receiving additional content requests from the caching agent. The additional content requests may indicate that the caching agent is requesting the same content data for additional clients.
APPARATUSES AND METHODS FOR REDUCING LATENCY IN A CONVEYANCE OF DATA IN NETWORKS
Aspects of the disclosure may include, for example, determining, by a first processing system, that first data is unavailable within a first storage device of the first processing system, resulting in a first determination, determining, based on the first determination, that the first data is available within a second storage device of a second processing system, resulting in a second determination, wherein the first processing system and the second processing system are included in a first tier of a processing system hierarchy, and wherein the processing system hierarchy includes a second tier, the second tier including a third processing system, obtaining, based on the second determination, the first data from the second storage device of the second processing system, and transmitting, based on the obtaining of the first data from the second storage device of the second processing system, the first data to a communication device. Other embodiments are disclosed.
MECHANISM TO ENFORCE CONSISTENT NEXT HOPS IN A MULTI-TIER NETWORK
In general, the disclosure relates to a method for forwarding a packet through a multi-tier network by establishing a routing protocol session with network devices in the multi-tier network, obtaining routing protocol information from network devices of the multi-tier network, determining a group using the routing protocol information, generating an ordered group listing using network device identifiers (NDIs) for the network devices in the group, and programming a network device hardware of a network device of the set of network devices of the multi-tier network using the ordered group listing. The group includes a set of the network devices of the multi-tier network.
Real-Time Message Delivery And Update Service In A Proxy Server Network
This patent document describes technology for providing real-time messaging and entity update services in a distributed proxy server network, such as a CDN. Uses include distributing real-time notifications about updates to data stored in and delivered by the network, with both high efficiency and locality of latency. The technology can be integrated into conventional caching proxy servers providing HTTP services, thereby leveraging their existing footprint in the Internet, their existing overlay network topologies and architectures, and their integration with existing traffic management components.
DETERMINING WHETHER TO RATE LIMIT TRAFFIC
Some embodiments provide a method for a gateway datapath that executes on a gateway device to implement logical routers for a set of logical networks and process traffic between the logical networks and an external network. The method receives a data message at the gateway device. To process the data message, the method executes a set of processing stages that includes a processing stage for a particular logical router. As part of the processing stage for the particular logical router, the method (i) uses an access control list (ACL) table to determine whether the data message is subject to rate limiting controls defined for the particular logical router and (ii) only when the data message is subject to rate limiting controls, determines whether to allow the data message according to a rate limiting mechanism for the particular logical router.
SYSTEM AND METHOD OF MINIMIZING NETWORK BANDWIDTH RETRIEVED FROM AN EXTERNAL NETWORK
A system and method are provided of minimizing network bandwidth used from an external network by client peers in a local network. The method can include the operation of organizing a plurality of clients each having media streamlets and a client parent in a structure within the local network. Mapping information propagates through the structure. The mapping information represents local streamlet locations as stored by the plurality of clients. Another operation is retrieving a local streamlet from a client identified in the mapping information as having the local streamlet for the requesting client.
SYSTEM AND METHOD OF MINIMIZING NETWORK BANDWIDTH RETRIEVED FROM AN EXTERNAL NETWORK
A system and method are provided of minimizing network bandwidth used from an external network by client peers in a local network. The method can include the operation of organizing a plurality of clients each having media streamlets and a client parent in a structure within the local network. Mapping information propagates through the structure. The mapping information represents local streamlet locations as stored by the plurality of clients. Another operation is retrieving a local streamlet from a client identified in the mapping information as having the local streamlet for the requesting client.