Patent classifications
H04L67/1023
System providing faster and more efficient data communication
A system designed for increasing network communication speed for users, while lowering network congestion for content owners and ISPs. The system employs network elements including an acceleration server, clients, agents, and peers, where communication requests generated by applications are intercepted by the client on the same machine. The IP address of the server in the communication request is transmitted to the acceleration server, which provides a list of agents to use for this IP address. The communication request is sent to the agents. One or more of the agents respond with a list of peers that have previously seen some or all of the content which is the response to this request (after checking whether this data is still valid). The client then downloads the data from these peers in parts and in parallel, thereby speeding up the Web transfer, releasing congestion from the Web by fetching the information from multiple sources, and relieving traffic from Web servers by offloading the data transfers from them to nearby peers.
Policy-based forwarding to a load balancer of a load balancing cluster
Some embodiments of the invention provide a method for forwarding data messages between a client and a server (e.g., between client and server machines and/or applications). In some embodiments, the method receives a data message that a load balancer has directed from a particular client to a particular server after selecting the particular server from a set of several candidate servers for the received data message's flow. The method stores an association between an identifier associated with the load balancer and a flow identifier associated with the message flow, and then forwards the received data message to the particular server. The method subsequently uses the load balancer identifier in the stored association to forward to the particular load balancer a data message that is sent by the particular server. The method of some embodiments is implemented by an intervening forwarding element (e.g., a router) between the load balancer set and the server set.
Policy-based forwarding to a load balancer of a load balancing cluster
Some embodiments of the invention provide a method for forwarding data messages between a client and a server (e.g., between client and server machines and/or applications). In some embodiments, the method receives a data message that a load balancer has directed from a particular client to a particular server after selecting the particular server from a set of several candidate servers for the received data message's flow. The method stores an association between an identifier associated with the load balancer and a flow identifier associated with the message flow, and then forwards the received data message to the particular server. The method subsequently uses the load balancer identifier in the stored association to forward to the particular load balancer a data message that is sent by the particular server. The method of some embodiments is implemented by an intervening forwarding element (e.g., a router) between the load balancer set and the server set.
PFCP Session Load Balancer
Importing of a UE address into a VRF of perimeter equipment is facilitate by receiving a VPN update from the perimeter equipment including a route target of the perimeter equipment and a gNodeB address. In addition, session information is obtained by intercepting traffic between the UE address and a UPF. The session information including the UE address and address of a gNodeB to which the UE is connected. By matching the gNodeB addresses from the VPN update and the session information, the route target of the perimeter equipment to which the UE is connected may be determined. The UE address may then be imported exclusively into the VRF of the perimeter equipment.
ENHANCED SELF-ASSEMBLING AND SELF-CONFIGURING MICROSERVICES
A method for managing systems with interrelated microservices with self-assembling and self-configuring microservices includes receiving at a first micro service a service request from a client. A determination is the made whether the first micro service is capable of processing the service request. If the first micro service is capable of processing the service requests, then processing the service request; if the first micro service cannot process the service request then routing the service request to a first stem service. The first stem service determines whether there is a second micro service that can process the service request. If the second micro service that can process the service requests exists, then forwarding the service request to the second micro service for processing. If there is no second micro service that can service the service requests then morphing the first stem service into a micro service that can service the service request.
Method, apparatus and system for transmitting data
A method, apparatus and system for transmitting data. The system includes: the load balance gate wall translating a virtual MAC address of the virtual server in a destination address of a data packet from a client to a real MAC address of the real server according to a pre-configured address mapping relationship between a real server and a virtual server; the load balance gate wall sending a modified data packet to the virtual switch; the virtual switch determining a data packet flowing into the real server after receiving the data packet from the load balance gate wall, then translating a virtual IP and a virtual port of the virtual server in the destination address in the determined data packet to a real IP and a real port of the real server; and the virtual switch sending the modified data packet to the real server.
Information transmission method, network element selector, and controller
An information transmission method includes: receiving, by a network element selector from a user equipment (UE), a first message including identification information of the UE; determining, based on the identification information of the UE, a user group to which the UE belongs; determining a controller corresponding to the user group; and receiving and sending a further message to the controller.
Sharded Database Leader Replica Distributor
A system can a divide database into a group of shards distributed among a group of data centers, wherein the group of shards comprises respective leader replicas. The system can determine respective correlation values between pairs of shards of the group of shards. The system can examine the pairs of shards in a descending order of respective correlation values, comprising, in response to determining that a respective pair of shards of the pairs of shards has a first correlation value greater than a predetermined threshold value, and that at least one shard of the respective pair of shards is unlocked, reassigning leader replicas of the respective pair of shards to be stored in a same data center of the group of data centers, and locking the leader replicas of the respective pair of shards from being reassigned to another data center of the group of data centers during the examining.
Sharded Database Leader Replica Distributor
A system can a divide database into a group of shards distributed among a group of data centers, wherein the group of shards comprises respective leader replicas. The system can determine respective correlation values between pairs of shards of the group of shards. The system can examine the pairs of shards in a descending order of respective correlation values, comprising, in response to determining that a respective pair of shards of the pairs of shards has a first correlation value greater than a predetermined threshold value, and that at least one shard of the respective pair of shards is unlocked, reassigning leader replicas of the respective pair of shards to be stored in a same data center of the group of data centers, and locking the leader replicas of the respective pair of shards from being reassigned to another data center of the group of data centers during the examining.
Image matching server network implementing a score between a server and an image store
Technology is disclosed for feeding source images from image stores of client systems into an image processing system using image feeding servers (“the technology”). The technology includes multiple image feeding servers that can feed a source image to the image processing system, using which the image processing system can generate processed images to be served to end users. The image feeding servers can be installed at various locations, e.g., geographically spread, and more proximate to where the source images are stored. An image feeding server obtains the source image from the image store in original size, downsizes the source image, and transmits the downsized source image to the image processing system. The technology selects an image feeding server based on a score of the image feeding server. The score can be determined as a function of one or more image feeding server parameters, e.g., latency, workload, or computing resource.