Patent classifications
H04L2012/5681
DATA PROCESSING METHOD AND DEVICE
Embodiments of the present disclosure provide a data processing method and a device. The method includes: determining, by a controller, a data packet filtering rule and a data forwarding path; sending, by the controller, the data packet filtering rule and configuration information carrying the data forwarding path to a first node, where the filtering rule is used to match configuration information corresponding to a data packet of the first node, the configuration information includes routing information corresponding to each node in the data forwarding path, routing information includes a queue identifier which is used to identify a transmission queue to which the data packet belongs; forwarding, by the first node and a remaining node, the data packet.
Bypass FIFO for multiple virtual channels
A group of low-level FIFOs may be logically bound together to form a super-FIFO. The super-FIFO may treat each low-level FIFO as a storage location. The super-FIFO may enable a push to (or a pop from) every low-level FIFO, simultaneously. The super-FIFO may enable a virtual channel (VC) to use the super-FIFO, bypassing a VC FIFO for the VC, removing several cycles of latency otherwise needed for enqueuing and dequeuing messages in the VC FIFO. In addition, the super-FIFO may enable bypassing of an arbiter, further reducing latency by avoiding a penalty of the arbiter.
SYNCHRONIZATION OF AUDIO STREAMS AND SAMPLING RATE FOR WIRELESS COMMUNICATION
Disclosed herein, among other things, are methods and apparatus for providing a time-stamp based controller for synchronization of sink or source sampling rate with external packet rate. A method for wireless communications includes receiving a transmission of a packet using a wireless transceiver of an electronic device, and using a processor of the electronic device to read a first value of a system timer and store the first value as an arrival time-stamp. The packet is decoded and processed by the processor, and sent to an output. When the processed packet is sent, a second value of the system timer is read, adjusted and stored as a departure time-stamp. The arrival time-stamp and the departure time-stamp are used to calculate an adjustment stimulus for a sample rate actuator of the electronic device. The sample rate actuator is configured to maintain synchronization of sampling rate with an external packet rate.
TECHNOLOGIES FOR JITTER-ADAPTIVE LOW-LATENCY, LOW POWER DATA STREAMING BETWEEN DEVICE COMPONENTS
Technologies for low-latency data streaming include a computing device having a processor that includes a producer and a consumer. The producer generates a data item, and in a local buffer producer mode adds the data item to a local buffer, and in a remote buffer producer mode adds the data item to a remote buffer. When the local buffer is full, the producer switches to the remote buffer producer mode, and when the remote buffer is below a predetermined low threshold, the producer switches to the local buffer producer mode. The consumer reads the data item from the local buffer while operating in a local buffer consumer mode and reads the data item from the remote buffer while operating in a remote buffer consumer mode. When the local buffer is above a predetermined high threshold, the consumer may switch to a catch-up operating mode. Other embodiments are described and claimed.
COMBINED INPUT AND OUTPUT QUEUE FOR PACKET FORWARDING IN NETWORK DEVICES
An apparatus for switching network traffic includes an ingress packet forwarding engine and an egress packet forwarding engine. The ingress packet forwarding engine is configured to determine, in response to receiving a network packet, an egress packet forwarding engine for outputting the network packet and enqueue the network packet in a virtual output queue. The egress packet forwarding engine is configured to output, in response to a first scheduling event and to the ingress packet forwarding engine, information indicating the network packet in the virtual output queue and that the network packet is to be enqueued at an output queue for an output port of the egress packet forwarding engine. The ingress packet forwarding engine is further configured to dequeue, in response to receiving the information, the network packet from the virtual output queue and enqueue the network packet to the output queue.
MULTIPATH TRAFFIC MANAGEMENT
One embodiment provides an apparatus. The apparatus includes client traffic management (CTM) logic. The CTM logic is to trigger implementation of a selected network traffic flow related to the client device, the triggering based, at least in part, on a network traffic flow related to the client device. The network traffic flow is associated with a connection and includes at least one subflow. Each subflow is carried by a respective path associated with the connection. The triggering includes at least one of constraining and/or adjusting an allowable throughput at a service provider for one or more of the at least one subflow. The selected traffic policy is to be implemented in a transport layer.
FLOW QUEUEING METHOD AND SYSTEM
A method includes receiving a packet. The method further includes determining whether the packet is part of a responsive connection. The method further includes determining whether a responsive buffer is full in response to a determination that the packet is part of the responsive connection. The method further includes applying a responsive probability to the packet in response to a determination that the responsive buffer is full. The method further includes determining whether to drop the packet based on the responsive probability. The method further includes accepting the packet for processing in response to a determination that the responsive buffer is not full or in response to a determination not to drop the packet.
Combined input and output queue for packet forwarding in network devices
An apparatus for switching network traffic includes an ingress packet forwarding engine and an egress packet forwarding engine. The ingress packet forwarding engine is configured to determine, in response to receiving a network packet, an egress packet forwarding engine for outputting the network packet and enqueue the network packet in a virtual output queue. The egress packet forwarding engine is configured to output, in response to a first scheduling event and to the ingress packet forwarding engine, information indicating the network packet in the virtual output queue and that the network packet is to be enqueued at an output queue for an output port of the egress packet forwarding engine. The ingress packet forwarding engine is further configured to dequeue, in response to receiving the information, the network packet from the virtual output queue and enqueue the network packet to the output queue.
Multipath traffic management
One embodiment provides an apparatus. The apparatus includes client traffic management (CTM) logic. The CTM logic is to trigger implementation of a selected network traffic flow related to the client device, the triggering based, at least in part, on a network traffic flow related to the client device. The network traffic flow is associated with a connection and includes at least one subflow. Each subflow is carried by a respective path associated with the connection. The triggering includes at least one of constraining and/or adjusting an allowable throughput at a service provider for one or more of the at least one subflow. The selected traffic policy is to be implemented in a transport layer.
Data processing unit for compute nodes and storage nodes
A new processing architecture is described in which a data processing unit (DPU) is utilized within a device. Unlike conventional compute models that are centered around a central processing unit (CPU), example implementations described herein leverage a DPU that is specially designed and optimized for a data-centric computing model in which the data processing tasks are centered around, and the primary responsibility of, the DPU. For example, various data processing tasks, such as networking, security, and storage, as well as related work acceleration, distribution and scheduling, and other such tasks are the domain of the DPU. The DPU may be viewed as a highly programmable, high-performance input/output (I/O) and data-processing hub designed to aggregate and process network and storage I/O to and from multiple other components and/or devices. This frees resources of the CPU, if present, for computing-intensive tasks.