H04L12/879

Payload cache

In one embodiment, a computer system includes a payload sub-system including interfaces to connect with respective devices, transfer data with the respective devices, and receive write transactions from the respective devices, a classifier to classify the received write transactions into payload data and control data, and a payload cache to store the classified payload data, and a processing unit (PU) sub-system including a local PU cache to store the classified control data, wherein the payload cache and the local PU cache are different physical caches in respective different physical locations in the computer system, and processing core circuitry configured to execute software program instructions to perform control and packet processing responsively to the control data stored in the local PU cache.

Router and routing method

The application discloses a router coupled to a first device and a second device. The router includes a first packet input interface, a second packet input interface, a first register, a second register, a control circuit and a switch module. The switch module includes a control port, a first packet output interface and a second packet output interface. The application further discloses a routing method. The router and routing method saves idle time and improves quality of service.

USER-MODE PROTOCOL STACK-BASED NETWORK ISOLATION METHOD AND DEVICE
20210392091 · 2021-12-16 ·

A user-mode protocol stack-based network isolation method includes: at a bottom-layer network card interface of a user-mode protocol stack, for each network card, adding an isolation space pointer for binding to a network isolation space; when a service application is initialized, configuring a plurality of user-mode protocol stack network isolation spaces with independent protocol stack private tables based on the user-mode protocol stack; for each network card, designating a corresponding user-mode protocol stack network isolation space through the isolation space pointer of each network card; and for service data received from each network card, performing data processing on the service data through protocol stack private tables in a user-mode protocol stack network isolation space corresponding to the network card.

DATA TRANSMISSION AND NETWORK INTERFACE CONTROLLER

Implementations of this disclosure provide data transmission operations and network interface controllers. An example method performed by a first RDMA network interface controller includes obtaining m data packets from a host memory of a first host; sending the m data packets to a second RDMA network interface controller of a second host; backing up the m data packets to a network interface controller memory integrated into the first RDMA network interface controller; determining that the second RDMA network interface controller does not receive n data packets of the m data packets; and in response, obtaining the n data packets from the m data packets that have been backed up to the network interface controller memory integrated into the first RDMA network interface controller, and retransmitting the n data packets to the second RDMA network interface controller.

STREAMING PLATFORM READER

A streaming platform reader includes: a plurality of reader threads configured to retrieve messages from a plurality of partitions of a streaming platform, wherein each message in the plurality of partitions is associated with a unique identifier; a plurality of queues coupled to the plurality of reader threads configured to store messages or an end of partition signal from the reader threads, wherein each queue includes a first position that stores the earliest message stored by a queue; a writer thread controlled by gate control logic that: compares the identifiers of all of the messages in the first positions of the queues of the plurality of queues, and forwards, to a memory, the message associated with the earliest identifier; and wherein the gate control logic blocks the writer thread unless each of the queues contains a message or an end of partition signal.

Determining whether to suppress an interrupt indicating completion of processing of a frame based on priority of the frame and current value information
11188399 · 2021-11-30 · ·

According to an embodiment, a notification control device includes a memory and one or more hardware processors configured to function as a determination unit and a notification unit. The determination unit is configured to determine, using notification control information set according to a priority of a frame, whether to notify of completion notification indicating completion of processing of the frame. The notification unit is configured to notify the completion notification when it is determined to notify of the completion notification.

Direct packet placement

Communication apparatus includes a host interface and a network interface, which receives from a packet communication network at least one packet stream including a sequence of data packets, which include headers containing respective sequence numbers and data payloads containing slices of the data segment having a predefined, fixed size per slice. Packet processing circuitry is configured to receive the data packets from the network interface, and to map the data payloads of the data packets in the at least one packet stream, using a linear mapping of the sequence numbers, to respective addresses in the buffer.

Packet processing with reduced latency

Generally, this disclosure provides devices, methods, and computer readable media for packet processing with reduced latency. The device may include a data queue to store data descriptors associated with data packets, the data packets to be transferred between a network and a driver circuit. The device may also include an interrupt generation circuit to generate an interrupt to the driver circuit. The interrupt may be generated in response to a combination of an expiration of a delay timer and a non-empty condition of the data queue. The device may further include an interrupt delay register to enable the driver circuit to reset the delay timer, the reset postponing the interrupt generation.

STORAGE EDGE CONTROLLER WITH A METADATA COMPUTATIONAL ENGINE

Embodiments described herein provide improved methods and systems for generating metadata for media objects at a computational engine (such as an artificial intelligence engine) within the storage edge controller, and for storing and using such metadata, in data processing systems.

NETWORK STORAGE FAILOVER SYSTEMS AND ASSOCIATED METHODS

Failover methods and systems for a networked storage environment are provided. A metadata data structure is generated, before starting a replay of entries at a log stored in a non-volatile memory of a second storage node, during a failover operation initiated in response to a failure at a first storage node. The second storage node operates as a partner node of the first storage node, and the metadata structure stores a metadata attribute of each log entry. Furthermore, the metadata attribute of each log entry is persistently stored. The persistently stored metadata attribute is used to respond to a read request received during the replay by the second storage node, while a write request metadata attribute of a write request is used for executing the write request received by the second storage node during the replay.