G06F9/505

Dynamic Profiling of Storage Class Memory for Implementation of Various Enterprise Solutions

Configuration and dynamic profiling of storage class memory (SCM) devices is provided. Information is retrieved that includes historical SCM device configurations, historical SCM device utilization, functional and non-functional properties of a plurality of SCM devices on a host node, current real time utilization of the plurality of SCM devices by an application workload of a customer running on the host node, and relationships between the plurality of SCM devices, needs of the customer, and resource capabilities and real time resource utilization on the host node. A configuration of each respective SCM device is determined based on retrieved information and an artificial intelligence-predicted SCM device future utilization trajectory of the customer. Each respective SCM device is dynamically configured with a set of SCM device partitions according to a corresponding SCM device profile based on the determined configuration of each respective SCM device of the plurality of SCM devices.

REDUCING THE ENVIRONMENTAL IMPACT OF DISTRIBUTED COMPUTING
20230017632 · 2023-01-19 ·

A process includes obtaining a workload and a set of candidate computing resources and predicting amounts of carbon emissions attributable to executing the workload on different members of the set of candidate computing resources. The process also includes predicting measures of computing performance in executing the workload of the different members of the set of candidate computing resources and computing a set of scores based on the amounts of carbon emissions and the measures of computing performance. The process also includes orchestrating the workload based on the scores.

Quiesce notifications for query retries

The subject technology retrieves information related to a set of instances of compute service managers, each instance of a particular compute service manager being associated with a set of virtual warehouses. The subject technology filters the information to determine a set of candidates from the set of instances of compute service managers. The subject technology sorts the set of candidates based at least in part on each workload of each of the set of candidates. The subject technology selects a candidate compute service manager from the set of instances of compute service managers to issue a query restart by randomly selecting an execution node, the execution node being included in a particular virtual warehouse associated with the candidate compute service manager, the selecting facilitating improving utilization of cluster resources and improving query execution on the selected candidate compute service manager.

MULTI-REGION DEPLOYMENT OF JOBS IN A FEDERATED CLOUD INFRASTRUCTURE
20230014635 · 2023-01-19 ·

A system and method for multi-region deployment of application jobs in a federated cloud computing infrastructure. A job is received for execution in two or more regions of the federated cloud computing infrastructure, each of the two or more regions comprising a collection of servers joined in a raft group for separate, regional execution of the job generating a copy of the job for each of the two or more regions. The job is then deployed to the two or more regions, the workload orchestrator deploying the job according to a deployment plan. A state indication is received from each of the two or more regions, the state indication representing a state of completion of the job by each respective region of the multi-cloud computing infrastructure.

DYNAMIC CROSS-ARCHITECTURE APPLICATION ADAPTION
20230014741 · 2023-01-19 · ·

Embodiments described herein are generally directed to improving performance of high-performance computing (HPC) or artificial intelligence (AI) workloads on cluster computer systems. According to one embodiment, a section of a high-performance computing (HPC) or artificial intelligence (AI) workload executing on a cluster computer system is identified as significant to a figure of merit (FOM) of the workload. An alternate placement among multiple heterogeneous compute resources of a node of the cluster computer system is determined for a portion of the section currently executing on a given compute resource of the multiple heterogeneous compute resources. After predicting an improvement to the FOM based on the alternate placement, the portion is relocated to the alternate placement.

Incremental precision networks using residual inference and fine-grain quantization

One embodiment provides for a computing device comprising a parallel processor compute unit to perform a set of parallel integer compute operations; a ternarization unit including a weight ternarization circuit and an activation quantization circuit; wherein the weight ternarization circuit is to convert a weight tensor from a floating-point representation to a ternary representation including a ternary weight and a scale factor; wherein the activation quantization circuit is to convert an activation tensor from a floating-point representation to an integer representation; and wherein the parallel processor compute unit includes one or more circuits to perform the set of parallel integer compute operations on the ternary representation of the weight tensor and the integer representation of the activation tensor.

Resource usage prediction for cluster provisioning

A system for provisioning resources includes a processor and a memory. The processor is configured to receive a time series of past usage data. The past usage data comprises process usage data and instance usage data. The processor is further configured to determine an upcoming usage data based at least in part on the time series of the past usage data, and provision a computing system according to the upcoming usage data.

Data query method, apparatus and device

A method including obtaining resource overheads according to feature information of a received query request; according to the resource overheads and a compute node resource, dynamically adjusting a compute node in a resource pool; and querying, by using the compute node, data corresponding to the query request. A compute node in a resource pool may be dynamically adjusted, so that the compute node in the resource pool may process all the received query requests, and therefore, the processing efficiency and a resource utilization rate of the compute node are more effectively improved, such that the compute node may more efficiently perform parallel processing on the multiple query requests, and the utilization rates of a CPU resource, a memory resource and a network bandwidth resource are increased, thus achieving better effect from the perspectives of overall computing resource and user query load and improving the usage experience of a user.

TECHNIQUES FOR MODIFYING CLUSTER COMPUTING ENVIRONMENTS

Systems, devices, and methods discussed herein are directed to intelligently adjusting the set of worker nodes within a computing cluster. By way of example, a computing device (or service) may monitor performance metrics of a set of worker nodes of a computing cluster. When a performance metric is detected that is below a performance threshold, the computing device may perform a first adjustment (e.g., an increase or decrease) to the number of nodes in the cluster. Training data may be obtained based at least in part on the first adjustment and utilized with supervised learning techniques to train a machine-learning model to predict future performance changes in the cluster. Subsequent performance metrics and/or cluster metadata may be provided to the machine-learning model to obtain output indicating a predicted performance change. An additional adjustment to the number of worker nodes may be performed based at least in part on the output.

System and method for supporting a usage calculation process in a cloud infrastructure environment

Systems and methods described herein support a usage calculation process in a cloud infrastructure environment. The usage calculation process can be used to determine whether a requested transaction that targets a compartment within a tree-structure of compartments violates any compartment quota or limit within parent compartments within the tree-structure.