Patent classifications
H04L67/1017
VIRTUAL SYSTEMS MANAGEMENT
An apparatus and method for automatic configuration management of a network are provided. The apparatus and method may provide a virtualization system that has an inventory that includes physical resources and virtual assets managed by a virtualization layer running on the physical resources; wherein virtual assets are software instantiations of computer systems, receive a virtual asset provisioning request that specifies parameters to be considered for the virtual asset provisioning request and receive data on inventory available in the virtualization system. The apparatus and method may automatically provision a particular virtual asset to a particular physical resource in the virtualization system that matches the parameters in the virtual asset provisioning request, assign the provisioned virtual asset to the virtual asset provisioning request and provide user access to the provisioned virtual asset.
VIRTUAL SYSTEMS MANAGEMENT
An apparatus and method for automatic configuration management of a network are provided. The apparatus and method may provide a virtualization system that has an inventory that includes physical resources and virtual assets managed by a virtualization layer running on the physical resources; wherein virtual assets are software instantiations of computer systems, receive a virtual asset provisioning request that specifies parameters to be considered for the virtual asset provisioning request and receive data on inventory available in the virtualization system. The apparatus and method may automatically provision a particular virtual asset to a particular physical resource in the virtualization system that matches the parameters in the virtual asset provisioning request, assign the provisioned virtual asset to the virtual asset provisioning request and provide user access to the provisioned virtual asset.
Establishing and using a tunnel from an origin server in a distributed edge compute and routing service
An edge server of a distributed edge compute and routing service receives a tunnel connection request from a tunnel client residing on an origin server, that requests a tunnel be established between the edge server and the tunnel client. The request identifies the hostname that is to be tunneled. An IP address is assigned for the tunnel. DNS record(s) are added or changed that associate the hostname with the assigned IP address. Routing rules are installed in the edge servers of the distributed edge compute and routing service to reach the edge server for the tunneled hostname. The edge server receives a request for a resource of the tunneled hostname from another edge server that received the request from a client, where the other edge server is not connected to the origin server. The request is transmitted from the edge server to the origin server over the tunnel.
Establishing and using a tunnel from an origin server in a distributed edge compute and routing service
An edge server of a distributed edge compute and routing service receives a tunnel connection request from a tunnel client residing on an origin server, that requests a tunnel be established between the edge server and the tunnel client. The request identifies the hostname that is to be tunneled. An IP address is assigned for the tunnel. DNS record(s) are added or changed that associate the hostname with the assigned IP address. Routing rules are installed in the edge servers of the distributed edge compute and routing service to reach the edge server for the tunneled hostname. The edge server receives a request for a resource of the tunneled hostname from another edge server that received the request from a client, where the other edge server is not connected to the origin server. The request is transmitted from the edge server to the origin server over the tunnel.
Packet processing method, related device, and computer storage medium
A packet processing method includes: receiving, by a load balancing apparatus, a data packet sent by a UE; when the data packet includes historical association information, selecting, by the load balancing apparatus based on the historical association information, a service instance used to process the data packet; and sending, by the load balancing apparatus, the data packet to the selected service instance.
Method and System for Achieving High Availability of Service under High-Load Scene in Distributed System
Provided are a method and system for achieving high availability of service under a high-load scene in a distributed system. The method includes: constructing a node selection model at a master node of a distributed cluster; constructing a request selection model in each slave node; wherein the request selection model includes weights of designated requests and trade-off parameters set for requests each having a weight greater than a set value; when the distributed system enters the high-load scene, reading, by the master node, the node selection model, and sequentially selecting the served slave nodes according to a time slice round robin policy; and reading, by each slave node, the request selection model, and sequentially returning data of each request according to the trade-off parameters of each request.
Method and System for Achieving High Availability of Service under High-Load Scene in Distributed System
Provided are a method and system for achieving high availability of service under a high-load scene in a distributed system. The method includes: constructing a node selection model at a master node of a distributed cluster; constructing a request selection model in each slave node; wherein the request selection model includes weights of designated requests and trade-off parameters set for requests each having a weight greater than a set value; when the distributed system enters the high-load scene, reading, by the master node, the node selection model, and sequentially selecting the served slave nodes according to a time slice round robin policy; and reading, by each slave node, the request selection model, and sequentially returning data of each request according to the trade-off parameters of each request.
SYSTEMS AND METHODS FOR STREAMING MEDIA CONTENT DURING UNAVAILABILITY OF CONTENT SERVER
Systems and methods are described herein for streaming during unavailability of a content server. Upon determining that there are conditions indicating buffering issues during delivery of a media asset, a server determines a first group of devices suitable for receiving the media asset from the server and sharing the media asset on a peer-to-peer network. Then, the server determines a second group of devices suitable for receiving the media asset on a peer-to-peer network from a first group device. The server then determines groupings within which to share and receive the media asset. Next, the server transmits instructions to the devices in the first group to maintain in buffer and share certain portions of the media asset with the second group devices within their grouping. Finally, the server updates information detailing the media asset portions the devices are maintaining in buffer and sharing.
SYSTEMS AND METHODS FOR STREAMING MEDIA CONTENT DURING UNAVAILABILITY OF CONTENT SERVER
Systems and methods are described herein for streaming during unavailability of a content server. Upon determining that there are conditions indicating buffering issues during delivery of a media asset, a server determines a first group of devices suitable for receiving the media asset from the server and sharing the media asset on a peer-to-peer network. Then, the server determines a second group of devices suitable for receiving the media asset on a peer-to-peer network from a first group device. The server then determines groupings within which to share and receive the media asset. Next, the server transmits instructions to the devices in the first group to maintain in buffer and share certain portions of the media asset with the second group devices within their grouping. Finally, the server updates information detailing the media asset portions the devices are maintaining in buffer and sharing.
CONTENT MANAGEMENT SYSTEMS PROVIDING ZERO RECOVERY TIME OBJECTIVE
A server system receives an electronic document from a client system to be stored at a content management system. The server system generates multiple copies of the electronic document that are stored in parallel at respective multiple instances of the content management system. The server system receives a request to retrieve a copy of the electronic document. In response to the request, the server system checks, by a load balancer, whether the respective multiple instances of the content management system are operational to retrieve the copy of the electronic document. The server system then performs load balancing by the load balancer in accordance with a round-robin process to select a particular instance and retrieves the copy of the electronic document from the particular instance of the content management system.