H04L49/9063

Spatial dispersion buffer
11522817 · 2022-12-06 · ·

An improved buffer for networking and other computing devices comprises multiple memory instances, each having a distinct set of entries. Transport data units (“TDUs”) are divided into storage data units (“SDUs”), and each SDU is stored within a separate entry of a separate memory instance in a logical bank. One or more grids of the memory instances are organized into overlapping logical banks. The logical banks are arranged into views. Different destinations or other entities are assigned different views of the buffer. A memory instance may be shared between logical banks in different views. When overlapping logical banks are accessed concurrently, data in a memory instance that they share may be recovered using a parity SDU in another memory instance. The shared buffering enables more efficient buffer usage in a network device with a traffic manager shared amongst egress bocks. Example read and write algorithms for such buffers are disclosed.

STORAGE APPARATUS AND ADDRESS SETTING METHOD
20220337532 · 2022-10-20 ·

A storage apparatus includes: a plurality of storage controllers including controller interfaces including a plurality of interface ports for connection to the plurality of switches having switch ports, a plurality of virtual networks configured by one of the switch ports is configured in the switch, and the storage controller sends a first packet for specifying the switch port to which the interface port is to be connected from the interface port to the plurality of virtual networks, and determines an address of the interface port used for data transfer between the storage controllers based on a switch number of the switch and a switch port number of the switch port in a case where a second packet including information for specifying the switch number of the switch and the switch port number of the switch port to which the interface port is to be connected is received.

LIVE SOCKET REDIRECTION
20230164235 · 2023-05-25 ·

Networking methods and systems include determining a first state of a connection on a first network based on connection buffers at a host. A first system call relating to the connection is identified. A next state of the connection that would result from the first system call is determined. The first system call is executed responsive to a determination that the next state does not move the connection farther from a safe transition state.

System and method for modem stabilization when waiting for AP-driven link recovery
11606316 · 2023-03-14 · ·

Various embodiments of methods and systems for a modem-directed application processor boot flow in a portable computing device (“PCD”) are disclosed. An exemplary method includes an application processor that transitions into an idle state, such as a WFI state, for durations of time during a boot sequence that coincide with processing by a DMA engine and/or crypto engine. That is, the application processor may “sleep” while the DMA engine and/or crypto engine process workloads in response to instructions they received from the application processor.

ON CHIP ROUTER
20230208776 · 2023-06-29 ·

There is disclosed a router for routing data on a computing chip comprising a plurality of processing elements, the router comprising: a packet processing pipeline; a dropped packet buffer; and one or more circuits configured to: determine that a data packet in the packet processing pipeline is to be dropped; move the data packet that is to be dropped from the packet processing pipeline to the dropped packet buffer; and re-insert the dropped data packet from the dropped packet buffer into the packet processing pipeline for re-processing.

Adaptive interrupt coalescing for energy efficient mobile platforms

Methods and apparatus relating to adaptive interrupt coalescing for energy efficient mobile platforms are discussed herein. In one embodiment, one or more interrupts are buffered based on communication throughput. At least one of the one or more interrupts are released in response to expiration of an interrupt coalescing time period. Other embodiments are also claimed and described.

PROCESSOR AND IMPLEMENTATION METHOD, ELECTRONIC DEVICE, AND STORAGE MEDIUM

A processor includes a system controller configured to send predetermined packet information to a data packing and unpacking module; the data packing and unpacking module configured to acquire corresponding packet data from a storage array module according to the packet information, pack the packet data with the packet information, send a first packet obtained from packing to an operation module for operation processing, acquire a second packet returned by the operation module, unpack the second packet to obtain operation result data, and store the operation result data in the storage array module; the storage array module configured to store data; and the operation module configured to perform operation processing on the acquired first packet, generate the second packet according to the operation result data, and return the second packet to the data packing and unpacking module.

PARALLEL PROCESSING APPARATUS AND METHOD FOR CONTROLLING COMMUNICATION
20170293589 · 2017-10-12 · ·

A packet transmitting unit transmits, to a node via RDMA communication, a packet with a first identifier that represents a predetermined process and a second identifier that represents a destination communication interface and is a logical identifier, as a destination, being added thereto. A plurality of communication interfaces exist. A packet receiving unit receives a packet transmitted from the node via RDMA communication, selects a communication interface that is a destination of a received packet and is used in the predetermined process, based on the first identifier and the second identifier added to the received packet, and transfers the received packet to a selected communication interface.

Automated multi-fabric link aggregation system

An automated multi-fabric link aggregation system includes leaf switch devices that have leaf switch device downlink ports, that are included in a first network fabric, and that are aggregated to provide a first aggregation fabric. Each leaf switch device generates discovery communications including a first network fabric identifier for the first network fabric, and a first aggregation fabric identifier for the first aggregation fabric. The leaf switch devices then transmit the discovery communications via the leaf switch device downlink ports. I/O modules that have I/O module uplink port are included in a second network fabric and are aggregated to provide a second aggregation fabric. The I/O modules receive the discovery communications via each of the I/O module uplink ports, determine that each received discovery communication includes the first network fabric identifier and the first aggregation fabric identifier and, in response, automatically configure the I/O module uplink ports in a LAG.

Terabit-scale network packet processing via flow-level parallelization

In one example, the present disclosure describes a device, computer-readable medium, and method for organizing terabit-scale packet volumes into flows for downstream processing stages. For instance, in one example, a method includes extracting a first flow key from a first data packet, inputting the first flow key into a hash function to obtain a first output value, selecting a first partition in a memory to which to store the first data packet, wherein the first partition is selected based on the first output value, and storing the first data packet to the first partition.