H04L49/9042

REAL-TIME ON-CHIP DATA TRANSFER SYSTEM

A system and method for real-time data transfer on a system-on-chip (SoC) allows MIPI-CSI (camera serial interface) data received on a first interface to be output on another MIPI-CSI interface without using system memory or delaying the loopback path. The system includes a CSI receiver, a loopback buffer, and a CSI transmitter. The loopback buffer is used for the data transfer between the CSI receiver and the CSI transmitter. The CSI transmitter receives a payload included in a data packet from the CSI receiver by way of the loopback buffer. The CSI receiver communicates a packet header of the data packet to the CSI transmitter. The CSI transmitter reads the payload from the loopback buffer based on the packet header and at least one of a buffer threshold capacity and payload size.

SYSTEMS AND METHODS FOR ACCELERATING OBJECT STORES WITH DISTRIBUTED CACHING

An illustrative embodiment disclosed herein is an object store with distributed caching including a distributed cache cluster including a first cache on a first node device and a second cache on a second node device. The object store with distributed caching further includes a gateway server communicatively coupled to the distributed cache cluster. The gateway server receives a request to store an object from a client device, determines whether the object satisfies an object policy, determines whether the request indicates that the object is to be split up into a plurality of shards, and stores a first shard of the plurality of shards in the first cache and a second shard of the plurality of shards in the second cache.

Packet reconstruction at deparser

Some embodiments provide a method for a deparser of a processing pipeline. The method receives, from a set of match-action stages of the pipeline, packet header field (PHF) values for a first set of PHFs of a packet processed by the match-action stages. The method also receives, directly from a parser of the pipeline, (i) packet data for the packet prior to any modification by the match-action stages and (ii) descriptive data that specifies locations within the packet data for a second set of PHFs of the packet that are not included in the first set of PHFs. The method constructs a packet from (i) the PHF values received for the first set of PHFs and (ii) the packet data received for the second set of PHFs. The descriptive data is used to extract packet header field values for the second set of PHFs from the packet data.

Splitting data into an information vector and a control vector and processing, at a stage of a control pipeline, the control vector and a data block of the information vector extracted from a corresponding stage of a data pipeline
10680977 · 2020-06-09 · ·

Disclosed herein are techniques for improving the performance of a pipeline in an integrated circuit. An integrated circuit includes a pipeline including a plurality of stages, and a plurality of storage circuits coupled to corresponding stages in the plurality of stages of the pipeline. A first stage of the plurality of stages is configured to split a set of data into a first vector and a second vector. The plurality of stages is configured to transport the first vector through the pipeline, and sequentially perform operations on the first vector of the set of data. The plurality of storage circuits is configured to transport the second vector among the plurality of storage circuits. Each storage circuit of the plurality of storage circuits is configured to provide a data block in the second vector to the corresponding stage in the plurality of stages of the pipeline for data processing.

Processing packets according to hierarchy of flow entry storages

Some embodiments provide a method for processing a packet received by a managed forwarding element. The method performs a series of packet classification operations based on header values of the received packet. The packet classifications operations determine a next destination of the received packet. When the series of packet classification operations specifies to send the packet to a network service that performs payload transformations on the packet, the method (1) assigns a service operation identifier to the packet that identifies the service operations for the network service to perform on the packet, (2) sends the packet to the network service with the service operation identifier, and (3) stores a cache entry for processing subsequent packets without the series of packet classification operations. The cache entry includes the assigned service operation identifier. The network service uses the assigned service operation identifier to process packets without performing its own classification operations.

PACKET STORAGE BASED ON PACKET PROPERTIES
20240022528 · 2024-01-18 ·

In an example, a system includes a network port that receives a packet; a first memory; a second memory; and a packet analyzer coupled to the network port. The packet analyzer operates to divide the packet into multiple fragments, analyze each of the multiple fragments to determine whether the corresponding fragment has a first priority level or a second, lower, priority level, determine whether to store each of the multiple fragments in the first memory or the second memory based on the priority level determined for that fragment, store each fragment determined to have the first priority level in the first memory, and store each fragment determined to have the second priority level in the second memory. The network port, packet analyzer and the first memory, which may be a cache memory, may be embodied on a chip, and the second memory may be external to the chip.

ULTRA-SCALABLE, DISAGGREGATED INTERNET PROTOCOL (IP) AND ETHERNET SWITCHING SYSTEM FOR A WIDE AREA NETWORK
20200092228 · 2020-03-19 ·

Systems and Methods for IP and Ethernet switching in an ultra-scalable disaggregated wide area common carrier (WACC) disaggregated networking switching system. The WACC network switching system may include an Ethernet fabric having a set of M Ethernet switches each including a set of N switch ports, and a set of N input/output (IO) devices each including a set of W IO ports, a set of M Ethernet ports, an IO side packet processor (IOSP), and a fabric side packet processor (FSP). Each Ethernet switch may establish switch queues. Each IO device may establish a set of M hierarchical virtual output queues each including a set of N ingress-IOSP queues and ingress-virtual output queues, a set of W egress-IOSP queues, a set of M ingress-FSP queues, and a set of N hierarchical virtual input queues each including a set of N egress-FSP queues and egress-virtual input queues.

Network Forensic System and Method
20240031270 · 2024-01-25 ·

A high performance packet stream storage method. Original packet data from data traffic transmitted over a network is collected. Collected original packet data is written in a memory. Metadata from the collected original package data is extracted and metadata is written in the memory. The original packet data and the metadata is stored in a storage unit.

Forwarding Element Data Plane with Computing Parameter Distributor

Some embodiments provide a network forwarding element with a data-plane forwarding circuit that has a parameter collecting circuit to store and distribute parameter values computed by several machines in a network. In some embodiments, the machines perform distributed computing operations, and the parameter values that compute are parameter values associated with the distributed computing operations. The parameter collecting circuit of the data-plane forwarding circuit (data plane) in some embodiments (1) stores a set of parameter values computed and sent by a first set of machines, and (2) distributes the collected parameter values to a second set of machines once it has collected the set of parameter values from all the machines in the first set. The first and second sets of machines are the same set of machines in some embodiments, while they are different sets of machines (e.g., one set has at least one machine that is not in the other set) in other embodiments. In some embodiments, the parameter collecting circuit performs computations on the parameter values that it collects and distributes the result of the computations once it has processed all the parameter values distributed by the first set of machines. The computations are aggregating operations (e.g., adding, averaging, etc.) that combine corresponding subset of parameter values distributed by the first set of machines.

Forwarding Element Data Plane with Computing Parameter Distributor

Some embodiments provide a network forwarding element with a data-plane forwarding circuit that has a parameter collecting circuit to store and distribute parameter values computed by several machines in a network. In some embodiments, the machines perform distributed computing operations, and the parameter values that compute are parameter values associated with the distributed computing operations. The parameter collecting circuit of the data-plane forwarding circuit (data plane) in some embodiments (1) stores a set of parameter values computed and sent by a first set of machines, and (2) distributes the collected parameter values to a second set of machines once it has collected the set of parameter values from all the machines in the first set. The first and second sets of machines are the same set of machines in some embodiments, while they are different sets of machines (e.g., one set has at least one machine that is not in the other set) in other embodiments. In some embodiments, the parameter collecting circuit performs computations on the parameter values that it collects and distributes the result of the computations once it has processed all the parameter values distributed by the first set of machines. The computations are aggregating operations (e.g., adding, averaging, etc.) that combine corresponding subset of parameter values distributed by the first set of machines.