Patent classifications
G06F9/5011
A Multi-Tenant Real-Time Process Controller for Edge Cloud Environments
The present disclosure relates to a method performed by a process control node (210) configured to allocate resources shared by a plurality of tenant applications, wherein each tenant application comprises a selection of non real-time processes and real-time processes, the method comprising receiving a first resource request, from a tenant application, indicative of resources requested to be allocated, by the process control node, for one or more real-time processes of the tenant application, evaluating a scheduling test to determine if the set of processing resources can be allocated from the shared resources by determining if resources requested by the first resource request can be allocated, and if it is determined that the requested resources can be allocated from the shared resources, the method further comprises performing the steps starting the one or more real-time processes of the tenant application within a resource partition of the tenant application, calculating updated resource quotas and priorities for non real-time processes comprised by the tenant application, transmitting a first resource response to the tenant application.
Intelligent and automatic load balancing of workloads on replication appliances based on appliance load scores
Various systems and methods are provided in which a replication process is initiated between a primary site and a recovery site, each having plurality of gateway appliances. Replication loads are evaluated for each given gateway appliance of the plurality of gateway appliances. If a determination is made that at least one gateway appliance of the plurality of gateway appliances is not overloaded, the plurality of gateway appliances are sorted based on replication loads respectively associated with each gateway appliance, and a determination is made as to whether a relative difference in replication loads between a gateway appliance having a highest replication load and a gateway appliance having a lowest replication load exceeds a difference threshold to determine whether the replication workloads between the gateway appliances should be rebalanced.
METHOD FOR EXECUTING MULTITHREADED INSTRUCTIONS GROUPED INTO BLOCKS
A method for executing multithreaded instructions grouped into blocks. The method includes receiving an incoming instruction sequence using a global front end; grouping the instructions to form instruction blocks, wherein the instructions of the instruction blocks are interleaved with multiple threads; scheduling the instructions of the instruction block to execute in accordance with the multiple threads; and tracking execution of the multiple threads to enforce fairness in an execution pipeline.
JOB SCHEDULING MANAGEMENT
Resource utilization data for a set of system components of a computing system is collected. The resource utilization data includes performance records for a set of jobs. By analyzing the collected resource utilization data for the set of system components, a resource allocation is identified for a particular job of the set of jobs. A first execution time for the particular job is determined based on the resource allocation for the particular job and the resource utilization data for the set of system components. A location at which to execute the particular job is determined based on how the particular job has been executed at the location previously. The first execution time may be a time when the computer system achieves a resource availability threshold with respect to the resource allocation. Aspects are also directed toward performing the particular job at the first execution time.
Low latency distributed counters for quotas
A system for managing access to a shared resource includes a plurality of nodes is described. The system may determine a quota for each of a plurality of nodes in view of one or more factors associated with a respective node and set a local counter stored in a memory of each of the nodes to the value of the relevant quota. In response to a request to access a resource made by an end device, the system may serve the request with a first node in the nodes, wherein the local counter stored in the memory of the first node is more than zero, and decrement the local counter stored in the memory of the first node by one after serving the request. Synchronization data is sent from each of the nodes, including the value of the local counter stored in the memory of the relevant node.
Function resource configuration method and device
A function resource configuration method includes receiving, by a wearable device, a distribution request sent by a first terminal, where the distribution request is used to request the wearable device to distribute a first function resource of the wearable device to the first terminal, and where the first function resource is already occupied by a second terminal. The wearable device determines, based on the distribution request, whether to allow the first terminal to use the first function resource. If the wearable device allows the first terminal to use the first function resource, the wearable device sends a notification message to the first terminal that instructs the first terminal to use the first function resource.
INTEROPERABILITY-AS-A-SERVICE IN A CLOUD ENVIRONMENT
Methods, devices, and techniques for determining interoperable resources are discussed herein. For example, in one aspect, a resource in a cloud environment may be discovered. Responsive to discovering the resource, an interoperability support matrix associated with the resource can be obtained. The interoperability support matrix may specify another resource that interoperates with the resource. An interoperability record is then stored in an interoperability support matrix repository. The interoperability record can specify that the another resource interoperates with the resource.
System and method of providing system jobs within a compute environment
The disclosure relates to systems, methods and computer-readable media for using system jobs for performing actions outside the constraints of batch compute jobs submitted to a compute environment such as a cluster or a grid. The method for modifying a compute environment from a system job disclosure associating a system job to a queuable object, triggering the system job based on an event and performing arbitrary actions on resources outside of compute nodes in the compute environment. The queuable objects include objects such as batch compute jobs or job reservations. The events that trigger the system job may be time driven, such as ten minutes prior to completion of the batch compute job, or dependent on other actions associated with other system jobs. The system jobs may be utilized also to perform rolling maintenance on a node by node basis.
Dynamically mapping software infrastructure utilization
A computer-based system and method for real-time monitoring of computer resource usage, including obtaining, by a monitoring application executed by a processor, from a plurality of applications, each application executed by a processor, a report upon the accessing of at least one accessed resource by at least one accessing user; and generating, by the monitoring application based on the report, a map of resources accessed by the plurality of applications. If a notification that a resource has been compromised is obtained, a list of all applications that have accessed the resource may be generated based on the map.
Graphics processors and graphics processing units having dot product accumulate instruction for hybrid floating point format
Described herein is a graphics processing unit (GPU) comprising a first processing cluster to perform parallel processing operations, the parallel processing operations including a ray tracing operation and a matrix multiply operation; and a second processing cluster coupled to the first processing cluster, wherein the first processing cluster includes a floating-point unit to perform floating point operations, the floating-point unit is configured to process an instruction using a bfloat16 (BF16) format with a multiplier to multiply second and third source operands while an accumulator adds a first source operand with output from the multiplier.