G06F12/0851

ACCELERATOR OR ACCELERATED FUNCTIONS AS A SERVICE USING NETWORKED PROCESSING UNITS

Various approaches for deploying and controlling distributed accelerated compute operations with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A system for orchestrating acceleration functions in a network compute mesh is configured to access a flowgraph, the flowgraph including data producer-consumer relationships between a plurality of tasks in a workload; identify available artifacts and resources to execute the artifacts to complete each of the plurality of tasks, wherein an artifact is an instance of a function to perform a task of the plurality of tasks; determine a configuration assigning artifacts and resources to each of the plurality of tasks in the flowgraph; and schedule, based on the configuration, the plurality of tasks to execute using the assigned artifacts and resources.

INFRASTRUCTURE-DELEGATED ORCHESTRATION BACKUP USING NETWORKED PROCESSING UNITS

Various approaches for monitoring and responding to orchestration or service failures with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A method performed by a computing device for deploying remedial actions in failure scenarios of an orchestrated edge computing environment may include: identifying an orchestration configuration of a controller entity (responsible for orchestration) and a worker entity (subject to the orchestration to provide at least one service); determining a failure scenario of the orchestration of the worker entity, such as at a networked processing unit implemented at a network interface located between the controller entity and the worker entity; and causing a remedial action to resolve the failure scenario and modify the orchestration configuration, such as replacing functionality of the controller entity or the worker entity with functionality at a replacement entity.

MEMORY INTERLEAVING COORDINATED BY NETWORKED PROCESSING UNITS

Various approaches for configuring interleaving in a memory pool used in an edge computing arrangement, including with the use of infrastructure processing units (IPUs) and similar networked processing units, are disclosed. An example system may discover and map disaggregated memory resources at respective compute locations connected to each another via at least one interconnect. The system may identify workload requirements for use of the compute locations by respective workloads, for workloads provided by client devices to the compute locations. The system may determine an interleaving arrangement for a memory pool that fulfills the workload requirements, to use the interleaving arrangement to distribute data for the respective workloads among the disaggregated memory resources. The system may configure the memory pool for use by the client devices of the network, as the memory pool causes the disaggregated memory resources to host data based on the interleaving arrangement.

MANAGEMENT OF WORKLOAD PROCESSING USING DISTRIBUTED NETWORKED PROCESSING UNITS

Various approaches for deploying and controlling distributed compute operations with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A system that includes a networked processing unit may perform workload processing with operations that: receive, from another networked processing unit, workload information for a workload, for a workload having respective tasks to be processed among distributed computing entities; perform an analysis of network conditions for a predicted execution of the workload, based on the workload information, to analyze network availability among the distributed computing entities; perform an analysis of compute conditions for the predicted execution of the workload, based on the workload information, to analyze processing availability among the distributed computing entities; and identify locations of the distributed computing entities to deploy the workload, based on the analysis of network conditions and the analysis of compute conditions.

OPTIMIZING CONCURRENT EXECUTION USING NETWORKED PROCESSING UNITS
20230136612 · 2023-05-04 ·

Various approaches for managing distributed compute operations for workload execution of concurrent tasks, including with the use of infrastructure processing units (IPUs) and similar networked processing units, are disclosed. An example method may include: identifying multiple tasks of a computing workload, for a workload that provides processing dependencies among the tasks, and that uses concurrent execution with one or more of the tasks; monitoring an execution time for each of the tasks, relative to an execution time threshold for each of the tasks; identifying the execution time of a particular task as exceeding an execution time threshold for the particular task; determining a remediation based on the particular task and the identified execution time, with the remediation including use of other compute resources in the distributed computing environment for the workload; and applying the remediation to increase speed of execution of the workload.

VIRTUAL POOLS AND RESOURCES USING DISTRIBUTED NETWORKED PROCESSING UNITS

Various approaches for deploying and using virtual pools of compute resources with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A host computing system may be configured to operate a virtual pool of resources, with operations including: identifying, at the host computing system, availability of a resource at the host computing system; transmitting, to a network infrastructure device, a notification that the resource at the host computing system is available for use in a virtual resource pool in the edge computing network; receiving a request for the resource in the virtual resource pool that is provided on behalf of a client computing system, based on the request being coordinated via the network infrastructure device and includes at least one quality of service (QoS) requirement; and servicing the request for the resource, based on the at least one QoS requirement.

IN-FLIGHT INCREMENTAL PROCESSING

Various approaches for deploying and controlling distributed compute operations with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. For example, a payload may be received at a networking infrastructure device. Here, the payload is part of a workload that is routed through the networking infrastructure device from a first network node to a second network node. The networking infrastructure device may obtain a workload graph for the workload with vertices of the workload graph specifying functions. The networking infrastructure device may apply a function to the payload in accordance with the workload graph to transform the payload into a processed payload. The processed payload may be transmitted towards the second network node.

SERVICE MESH SWITCHING

Various approaches for service mech switching, including the use of infrastructure processing units (IPUs) and similar networked processing units, are disclosed. For example, a packet that includes a service request for a service may be received at a networking infrastructure device. The service may include an application that spans multiple nodes in a network. An outbound interface of the networking infrastructure device may be selected through which to route the packet. The selection of the outbound interface may be based on a service component of the service request in the packet and network metrics that correspond to the service. The packet may then be transmitted using the outbound interface.

FEDERATED DISTRIBUTION OF COMPUTATION AND OPERATIONS USING NETWORKED PROCESSING UNITS

Various approaches for deploying and controlling distributed compute operations with the use of infrastructure processing units (IPUs) and similar network-addressable processing units are disclosed. A device for orchestrating functions in a network compute mesh is configured to receive, at a network-addressable processing unit of a network-addressable processing unit mesh from a requestor device, a computation request to execute a workflow with a set of objectives; query at least one other network-addressable processing units of the network-addressable processing unit mesh using the set of objectives, to determine aspects of available resources and data in the network-addressable processing unit mesh to apply to the workflow; transmit a list of recommended resources available to execute the workflow to the requestor device, the list of recommended resources being ranked based on at least one dimension of the resources; obtain a compute chain from the requestor device, the compute chain describing resource control transitions and data flow provided from the recommended resources and data in the network-addressable processing unit mesh; and schedule the execution of the workflow at one or more network-addressable processing units in the network-addressable processing unit mesh in accordance with the compute chain.

CONFIGURABLE CACHE FOR COHERENT SYSTEM
20230384931 · 2023-11-30 ·

A device includes a memory bank. The memory bank includes data portions of a first way group. The data portions of the first way group include a data portion of a first way of the first way group and a data portion of a second way of the first way group. The memory bank further includes data portions of a second way group. The device further includes a configuration register and a controller configured to individually allocate, based on one or more settings in the configuration register, the first way and the second way to one of an addressable memory space and a data cache.