G06F2213/0064

TECHNOLOGIES FOR MANAGING A FLEXIBLE HOST INTERFACE OF A NETWORK INTERFACE CONTROLLER
20230412365 · 2023-12-21 · ·

Technologies for processing network packets by a host interface of a network interface controller (NIC) of a compute device. The host interface is configured to retrieve, by a symmetric multi-purpose (SMP) array of the host interface, a message from a message queue of the host interface and process, by a processor core of a plurality of processor cores of the SMP array, the message to identify a long-latency operation to be performed on at least a portion of a network packet associated with the message. The host interface is further configured to generate another message which includes an indication of the identified long-latency operation and a next step to be performed upon completion. Additionally, the host interface is configured to transmit the other message to a corresponding hardware unit scheduler as a function of the subsequent long-latency operation to be performed. Other embodiments are described herein.

TECHNOLOGIES FOR ALLOCATING RESOURCES ACROSS DATA CENTERS
20230421358 · 2023-12-28 ·

Technologies for allocating resources across data centers include a compute device to obtain resource utilization data indicative of a utilization of resources for a managed node to execute a workload. The compute device is also to determine whether a set of resources presently available to the managed node in a data center in which the compute device is located satisfies the resource utilization data. Additionally, the compute device is to allocate, in response to a determination that the set of resources presently available to the managed node does not satisfy the resource utilization data, a supplemental set of resources to the managed node. The supplemental set of resources are located in an off-premises data center that is different from the data center in which the compute device is located. Other embodiments are also described and claimed.

SYSTEM AND METHOD FOR PRIORITY ORCHESTRATION AND SCHEDULING OF DIFFERENT BACKGROUND OPERATIONS

A method, computer program product, and computer system for receiving an indication about an amount of background IOs a background scheduler is capable of performing on a plurality of applications, wherein the indication may be based upon, at least in part, one of host IO latency and a rate of the host IO latency being at one of a plurality of levels. One or more applications of the plurality of applications on which to perform the background IOs may be determined. The background IO applications to be performed on the one or more applications of the plurality of applications may be scheduled based upon, at least in part, one of the host IO latency and the rate of the host IO latency being at one of the plurality of levels.

Computer Architecture Having Selectable Parallel and Serial Communication Channels Between Processors and Memory
20240004419 · 2024-01-04 ·

A computer architecture provides both a parallel memory bus and serial memory bus between a processor system and memory. Latency-tolerant memory access requests are steered to the serial memory bus which operates to increase the available memory bus bandwidth on the parallel memory. The invention also provides integrated circuit computer memory suitable for this application.

TECHNOLOGIES FOR LOAD BALANCING A NETWORK

Technologies for load balancing a storage network include a system. The system includes circuitry to adjust routing rules in a network interface controller to deliver a packet from one of multiple uplinks to one of any physical functions, circuitry to remap, in response to a failure of a switch, a port from one physical function to another physical function, and circuitry to communicate control data between a software defined network controller and one or more agents in one or more host endpoints with a hierarchical distributed hashing table.

Supporting timely and context triggered prefetching in microprocessors

Implementations of the disclosure implement timely and context triggered (TACT) prefetching that targets particular load IPs in a program contributing to a threshold amount of the long latency accesses. A processing device comprising an execution unit; and a prefetcher circuit communicably coupled to the execution unit is provided. The prefetcher circuit is to detect a memory request for a target instruction pointer (IP) in a program to be executed by the execution unit. A trigger IP is identified to initiate a prefetch operation of memory data for the target IP. Thereupon, an association is determined between memory addresses of the trigger IP and the target IP. The association comprising a series of offsets representing a path between the trigger IP and an instance of the target IP in memory. Based on the association, an offset from the memory address of the trigger IP to prefetch the memory data is produced.

System and method to improve input output command latency by dynamic size logical to physical caching

A method and apparatus are provided to divide a logical to physical table into multiple parts, one part in a first fast memory and a second part in a second non-volatile memory, wherein an algorithm may be used in the division.

Systems and methods for creating pipeline paths

A data pipeline including a plurality of graph pipeline paths is obtained. Each of the graph pipeline paths may include pipeline nodes, at least one of the pipeline node may have at least two inputs. Each of the inputs may comprise outputs of two or more other pipeline nodes. A linear pipeline path interface is provided for creating and presenting a linear pipeline path of the data pipeline. The linear pipeline path may include pipeline nodes, each being limited to a single input and a single output. The linear pipeline path interface may be limited to interacting with linear pipeline paths. A graph pipeline interface is provided for presenting and modifying the data pipeline. The linear pipeline path interface may be switched to the graph pipeline interface in response to a first user input.

Technologies for load balancing a network

Technologies for load balancing a storage network include a system. The system includes circuitry to adjust routing rules in a network interface controller to deliver a packet from one of multiple uplinks to one of any physical functions, circuitry to remap, in response to a failure of a switch, a port from one physical function to another physical function, and circuitry to communicate control data between a software defined network controller and one or more agents in one or more host endpoints with a hierarchical distributed hashing table.

System, apparatus and method for handshaking protocol for low power state transitions

In one embodiment, an apparatus includes: at least one processing circuit; at least one array associated with the at least one processing circuit; a power controller to manage power consumption of the apparatus; and a fabric bridge coupled to the power controller. The fabric bridge and power controller may be configured to implement a handshaking protocol to enable the fabric bridge to receive data from the at least one array via a sideband communication path and send the data to a system memory coupled to the apparatus via a primary communication path, prior to entry of the apparatus into a first low power state. Other embodiments are described and claimed.