G06F9/5027

DATA TRANSMISSION METHOD AND APPARATUS
20230038051 · 2023-02-09 ·

A data transmission method and apparatus are provided. The data transmission method is applied to a computer system including at least two coprocessors, for example, including a first coprocessor and a second coprocessor. A shared memory is deployed between the first coprocessor and the second coprocessor, and is configured to store data generated when subtasks are separately executed. Further, the shared memory further stores a storage address of data generated when a subtask is executed, and a mapping relationship between each subtask and a coprocessor that executes the subtask. Therefore, a storage address of data to be read by the coprocessor may be found based on the mapping relationship, and the data may further be directly read from the shared memory without being copied by using a system bus. This improves efficiency of data transmission between the coprocessors.

METHOD AND APPARATUS FOR DYNAMICALLY ADJUSTING PIPELINE DEPTH TO IMPROVE EXECUTION LATENCY

Apparatus and method for managing pipeline depth of a data processing device. For example, one embodiment of an apparatus comprises: an interface to receive a plurality of work requests from a plurality of clients; and a plurality of engines to perform the plurality of work requests; wherein the work requests are to be dispatched to the plurality of engines from a plurality of work queues, the work queues to store a work descriptor per work request, each work descriptor to include information needed to perform a corresponding work request, wherein the plurality of work queues include a first work queue to store work descriptors associated with first latency characteristics and a second work queue to store work descriptors associated with second latency characteristics; engine configuration circuitry to configure a first engine to have a first pipeline depth based on the first latency characteristics and to configure a second engine to have a second pipeline depth based on the second latency characteristics.

RESOURCE SCHEDULING METHOD AND RELATED APPARATUS
20230037783 · 2023-02-09 ·

The present disclosure relates to resource scheduling methods and apparatuses. In one example method, a scheduling node receives a task. The scheduling node obtains a target execution duration level to which the task belongs, where the target execution duration level represents a time length, and the target execution duration level indicates to use a target compute module of a target compute node in multiple compute nodes to execute the task. The scheduling node sends the task to the target compute node.

SYSTEMS, METHODS, AND APPARATUS TO IDENTIFY FUNCTIONS FOR COMPUTATIONAL DEVICES
20230039450 · 2023-02-09 ·

A method may include interacting with an interface for one or more computational devices, wherein the interacting is based on an identifier, and wherein the identifier comprises information that identifies a functionality of a computational device functions. The information may include a functionality identifier. The identifier may further include information that identifies a group of the computational device function. The group of the computational device function may be based on a source of the computational device function. The information that identifies the functionality of a computational device function may include a functionality identifier, and the information that identifies the group of the computational device function may include a group identifier. The functionality identifier may include a unique function identifier, and the group identifier may include an organizationally unique identifier.

LIMITING PROVISION AND DISPLAY OF REDUNDANT DIGITAL COMPONENTS ON A CLIENT DEVICE
20230038126 · 2023-02-09 ·

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for limiting provision and display of redundant digital components on a client. Methods can include storing, by a client device, a list of digital components specifying digital components available for provision to the client device. For a first digital component received within a first application, the client device detects a set of signals specifying a first user interaction with the first digital component and a second user interaction with content provided in response to the first user interaction. Based on whether an affirmative user action was performed, the client device can modify the list of digital components. When a request to access a content page within a second application is received, the client device can receive the second digital component, which can be selected from among digital components included on the modified list of digital components.

DYNAMIC RESOURCE PROVISIONING FOR USE CASES

A computer-implemented method, according to one embodiment, includes: receiving, at a computer, a request to facilitate a testing environment, where the request specifies a number and type of resources to be included in the testing environment. A database which lists available resources in systems and/or devices that are in communication with the computer is inspected and the available resources are compared to the number and type of resources specified in the request to be included in the testing environment. In response to determining that a valid combination of the available resources meets the number and type of resources specified in the request to be included in the testing environment, the database is updated to indicate that each of the resources in the valid combination are in use. Moreover, the request is satisfied by returning information about the resources in the valid combination.

OPTIMIZING VM NUMA CONFIGURATION AND WORKLOAD PLACEMENT IN A HETEROGENEOUS CLUSTER
20230038612 · 2023-02-09 ·

An example method of placing a virtual machine (VM) in a cluster of hosts is described. Each of the hosts having a hypervisor managed by a virtualization management server for the cluster, the hosts separated into a plurality of nonuniform memory access (NUMA) domains. The method including: comparing a virtual central processing unit (vCPU) and memory configuration of the VM with physical NUMA topologies of the hosts; selecting a set of the hosts spanning at least one of the NUMA domains, each host in the set of hosts having a physical NUMA topology that maximizes locality for vCPU and memory resources of the VM as specified in the vCPU and memory configuration; and providing the set of hosts to a distributed resource scheduler (DRS) executing in the virtualization management server, the DRS configured to place the VM in a host selected from the set of hosts.

CPU utilization for service level I/O scheduling

One or more aspects of the present disclosure relate to service level input/output scheduling to control central processing unit (CPU) utilization. Input/output (I/O) operations are processed with one or more of a first CPU pool and a second CPU pool of two or more CPU pools. The second CPU pool processes I/O operations that are determined to stall any of the CPU cores.

Data transformation for a machine learning model

Data transformation caching in an artificial intelligence infrastructure that includes one or more storage systems and one or more graphical processing unit (‘GPU’) servers, including: identifying, in dependence upon one or more machine learning models to be executed on the GPU servers, one or more transformations to apply to a dataset; generating, in dependence upon the one or more transformations, a transformed dataset; storing, within one or more of the storage systems, the transformed dataset; receiving a plurality of requests to transmit the transformed dataset to one or more of the GPU servers; and responsive to each request, transmitting, from the one or more storage systems to the one or more GPU servers without re-performing the one or more transformations on the dataset, the transformed dataset.

System and method for appraising resource configuration
11556383 · 2023-01-17 · ·

To more properly size resources in a destination to which IT resources will be migrated, a system for appraising a resource configuration estimates a source's load model representing a load of first resources in a first computer system which is the source of migration and estimates a destination's load model representing a load of second resources to be built by migrating the first resources to a second computer system based on the source's load model. The system compares performance requirements of the first resources against the destination's load model and finds the destination's load model that is conformable to the performance requirements. When determining design values of the second resources' configuration, the system corrects those design values based on the destination's load model estimated conformable to the performance requirements to decrease design margins of the resource configuration using a design correction value defined to meet a service level requested.