Patent classifications
G06F9/5083
Composable information handling systems in an open network using access control managers
A method for managing composed information handling systems includes obtaining a composition request for a composed information handling system, making a first determination that a first information handling system is not capable of servicing the composition request local, and based on the first determination: allocating, an available resource on the first information handling system to the composed information handling system, sending a resource allocation request to a system control processor manager for access to an additional resource, obtain, in response to the allocation request, a notification for access to a second information handling system of the information handling systems that provides the available resource, setting up management services for available resource and the additional resource to obtain logical hardware resources, and presenting the logical hardware resources to at least one compute resource set as bare metal resources.
Exchanging runtime state information between datacenters with a gateway using a controller bridge
In an embodiment, a computer-implemented method for dynamically exchanging runtime state data between datacenters with a gateway using a controller bridge is disclosed. In an embodiment, the method comprises: receiving one or more first runtime state data from one or more logical sharding central control planes (“CCPs”) controlling one or more logical sharding hosts; receiving one or more second runtime state data from a gateway that is controlled by a CCP that also controls one or more physical sharding hosts; aggregating to aggregated runtime state data, the one or more first runtime state data received from the one or more logical sharding CCPs and the one or more second runtime state data received from the gateway; determining updated runtime state data based on the aggregated runtime state data, the one or more first runtime state data, and the one or more second runtime state data; and transmitting the updated runtime state data to at least one of the one or more logical sharding CCPs and the gateway.
Reducing latency in downloading electronic resources using multiple threads
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for reducing latency in presenting content. In one aspect, a system includes a native application that presents an interactive item and a latency reduction engine. The latency reduction engine detects interaction with the interactive item that links to a first electronic resource that is different from the native application and provided by a first network domain and in response to the detecting, reduces latency in presenting the first electronic resource, including executing a first processing thread and a second processing thread in parallel. The first processing thread requests a second electronic resource from a second network domain and loads the second electronic resource and, in response to the loading, stores a browser cookie for the second network domain. The second processing thread requests the first electronic resource and presents the first electronic resource.
Distributable and customizable load-balancing of data-associated computation via partitions and virtual processes
Methods, systems, computer-readable media, and apparatuses for determining partitions and virtual processes in a simulation are presented. A plurality of partitions of a simulated world may be determined, and each partition may correspond to a different metric for entities in the simulated world. A plurality of virtual processes for the simulated world may also be determined. The system may assign a different virtual process to each partition. An indication of the partitions may be sent to one or more partition enforcer services, and an indication of the virtual processes may be sent to a virtual process manager.
DATABASE SIMULATION MODELING FRAMEWORK
Methods, systems, and computer program products are provided for creating a resource management testing environment. An initial population of databases is established in a database ring, having an in initial count of databases and different types of databases that are determined based on an initial database population model. The initial population model receives ring classification information for the database ring from a ring grouping model. A sequence of database population-change events is generated based on a model, to change the population of the databases over time in the ring. An orchestration framework performs testing of resource manager operations based on the model-defined initial population of databases and the model-defined populations of databases changed over time. Model-defined resource usage metrics for each database are utilized to test the resource manager operations. Resource usage metrics and database add/drop events of a production system are used to train the models.
Automatic scaling of microservices applications
A device may receive information identifying a set of tasks to be executed by a microservices application that includes a plurality of microservices. The device may determine an execution time of the set of tasks based on a set of parameters and a model. The set of parameters may include a first parameter that identifies a first number of instances of a first microservice of the plurality of microservices, and a second parameter that identifies a second number of instances of a second microservice of the plurality of microservices. The device may compare the execution time and a threshold. The threshold may be associated with a service level agreement. The device may selectively adjust the first number of instances or the second number of instances based on comparing the execution time and the threshold.
Using delayed autocorrelation to improve the predictive scaling of computing resources
Techniques are described for filtering and normalizing training data used to build a predictive auto scaling model used by a service provider network to proactively scale users' computing resources. Further described are techniques for identifying collections of computing resources that exhibit suitably predictable usage patterns such that a predictive auto scaling model can be used to forecast future usage patterns with reasonable accuracy and to scale the resources based on such generated forecasts. The filtering of training data and the identification of suitably predictable collections of computing resources are based in part on autocorrelation analyses, and in particular on “delayed” autocorrelation analyses, of time series data, among other techniques described herein.
Computer-implemented methods and nodes implementing performance estimation of algorithms during evaluation of data sets using multiparty computation based random forest
According to an aspect, there is provided a computer-implemented method of operating a first node. The first node has an algorithm for evaluating input data from another node, with the input data having a plurality of different attributes. The method comprises receiving, from a second node, a proposal for the evaluation of a first set of input data by the algorithm; estimating the performance of the algorithm in evaluating the first set of input data based on the proposal; and outputting, to the second node, an indication of the estimated performance of the algorithm. A corresponding first node is also provided.
System service timeout processing method, and apparatus
Embodiments of this application relate to the field of communications technologies, and provide a system service timeout processing method and an apparatus. The method includes: when a target system service thread in at least one system service thread times out, determining, by a terminal, a first application process communicating with the target system service thread, where the timeout of the target system service thread includes at least one of the following: a locked object occupied by the target system service thread is not released within a preset time, and the target system service thread is blocked; and ending, by the terminal, the first application process.
Distributed computing system for intensive video processing
A method and a system for distributing load in a network including a requesting node and a set of external processing nodes are disclosed. The method comprises sending an Internet Control Message Protocol (ICMP) message to each external processing node of the set of external processing nodes. The method further comprises identifying one or more external processing nodes from the set of external processing nodes as responding nodes based on a receipt of response to a corresponding ICMP message thereto. The method further comprises selecting an external processing node from the identified responding nodes as a most suitable external processing node having capability to fulfill a video processing request submitted by the requesting node. The method further comprises sending a transmission package from the requesting node to the selected external processing node for processing thereof. The method further comprises receiving a binary response from the selected external processing node.