Patent classifications
H04L67/1001
SYSTEM PROVIDING FASTER AND MORE EFFICIENT DATA COMMUNICATION
A system designed for increasing network communication speed for users, while lowering network congestion for content owners and ISPs. The system employs network elements including an acceleration server, clients, agents, and peers, where communication requests generated by applications are intercepted by the client on the same machine. The IP address of the server in the communication request is transmitted to the acceleration server, which provides a list of agents to use for this IP address. The communication request is sent to the agents. One or more of the agents respond with a list of peers that have previously seen some or all of the content which is the response to this request (after checking whether this data is still valid). The client then downloads the data from these peers in parts and in parallel, thereby speeding up the Web transfer, releasing congestion from the Web by fetching the information from multiple sources, and relieving traffic from Web servers by offloading the data transfers from them to nearby peers.
SYSTEM PROVIDING FASTER AND MORE EFFICIENT DATA COMMUNICATION
A system designed for increasing network communication speed for users, while lowering network congestion for content owners and ISPs. The system employs network elements including an acceleration server, clients, agents, and peers, where communication requests generated by applications are intercepted by the client on the same machine. The IP address of the server in the communication request is transmitted to the acceleration server, which provides a list of agents to use for this IP address. The communication request is sent to the agents. One or more of the agents respond with a list of peers that have previously seen some or all of the content which is the response to this request (after checking whether this data is still valid). The client then downloads the data from these peers in parts and in parallel, thereby speeding up the Web transfer, releasing congestion from the Web by fetching the information from multiple sources, and relieving traffic from Web servers by offloading the data transfers from them to nearby peers.
SCALABLE SERVER-BASED WEB SCRIPTING WITH USER INPUT
Disclosed are techniques and apparatuses that are configured to receive an indication that a web browsing session executing on an enterprise server needs additional information based on a request for additional information being sent to a client device. The request may include an identifier of the web browsing session and an identifier of an enterprise server that initiated the web browsing session. A globally unique identifier related to the web browsing session and an identifier of the enterprise server is stored in a common data store. The web browsing session may be paused when the web browsing session requests additional information from a client device. The client device may respond with the additional information. The system may provide the identifier of the enterprise server to a load balancing component so the identified web browsing session executing on the enterprise server may continue to be used.
SCALABLE SERVER-BASED WEB SCRIPTING WITH USER INPUT
Disclosed are techniques and apparatuses that are configured to receive an indication that a web browsing session executing on an enterprise server needs additional information based on a request for additional information being sent to a client device. The request may include an identifier of the web browsing session and an identifier of an enterprise server that initiated the web browsing session. A globally unique identifier related to the web browsing session and an identifier of the enterprise server is stored in a common data store. The web browsing session may be paused when the web browsing session requests additional information from a client device. The client device may respond with the additional information. The system may provide the identifier of the enterprise server to a load balancing component so the identified web browsing session executing on the enterprise server may continue to be used.
IMPLEMENTING A FAULT-TOLERANT MULTI-NRF NETWORK TOPOLOGY
A system, apparatus, method, and non-transitory computer readable medium for implementing a fault-tolerant multi-NRF network topology may include a network repository function (NRF) device including: at least one processor configured to execute computer readable instructions to cause the NRF device to, broadcast a NRF query to a NRF cluster of a core network, the NRF cluster including a plurality of NRF devices located in a public land mobile network (PLMN); receive NRF query responses from each NRF device of the NRF cluster, the NRF query responses including network configuration information of each NRF device; determine a status of each NRF device based on the network configuration information; determine a list of network function (NF) devices registered with the NRF cluster; receive a NF query request from a first NF device; and transmit a NF query response to the first NF device.
IMPLEMENTING A FAULT-TOLERANT MULTI-NRF NETWORK TOPOLOGY
A system, apparatus, method, and non-transitory computer readable medium for implementing a fault-tolerant multi-NRF network topology may include a network repository function (NRF) device including: at least one processor configured to execute computer readable instructions to cause the NRF device to, broadcast a NRF query to a NRF cluster of a core network, the NRF cluster including a plurality of NRF devices located in a public land mobile network (PLMN); receive NRF query responses from each NRF device of the NRF cluster, the NRF query responses including network configuration information of each NRF device; determine a status of each NRF device based on the network configuration information; determine a list of network function (NF) devices registered with the NRF cluster; receive a NF query request from a first NF device; and transmit a NF query response to the first NF device.
On-demand access to compute resources
Disclosed are systems, methods and computer-readable media for controlling and managing the identification and provisioning of resources within an on-demand center as well as the transfer of workload to the provisioned resources. One aspect involves creating a virtual private cluster within the on-demand center for the particular workload from a local environment. A method of managing resources between a local compute environment and an on-demand environment includes detecting an event associated with a local compute environment and based on the detected event, identifying information about the local environment, establishing communication with an on-demand compute environment and transmitting the information about the local environment to the on-demand compute environment, provisioning resources within the on-demand compute environment to substantially duplicate the local environment and transferring workload from the local-environment to the on-demand compute environment. The event can be a threshold or a triggering event within or outside of the local environment.
On-demand access to compute resources
Disclosed are systems, methods and computer-readable media for controlling and managing the identification and provisioning of resources within an on-demand center as well as the transfer of workload to the provisioned resources. One aspect involves creating a virtual private cluster within the on-demand center for the particular workload from a local environment. A method of managing resources between a local compute environment and an on-demand environment includes detecting an event associated with a local compute environment and based on the detected event, identifying information about the local environment, establishing communication with an on-demand compute environment and transmitting the information about the local environment to the on-demand compute environment, provisioning resources within the on-demand compute environment to substantially duplicate the local environment and transferring workload from the local-environment to the on-demand compute environment. The event can be a threshold or a triggering event within or outside of the local environment.
Highly redundant and scalable storage area network architecture
A storage area network architecture and a method for storing data in the storage area network architecture are disclosed. For example, the storage area network architecture comprises a first layer of servers, the first layer of servers comprising a plurality of file receiving servers, a second layer of servers, the second layer of servers comprising a plurality of file sharing servers in communication with the first layer of servers and a third layer of servers, the third layer of servers comprising a plurality of storage servers in communication with the second layer of servers.
Highly redundant and scalable storage area network architecture
A storage area network architecture and a method for storing data in the storage area network architecture are disclosed. For example, the storage area network architecture comprises a first layer of servers, the first layer of servers comprising a plurality of file receiving servers, a second layer of servers, the second layer of servers comprising a plurality of file sharing servers in communication with the first layer of servers and a third layer of servers, the third layer of servers comprising a plurality of storage servers in communication with the second layer of servers.