Patent classifications
H04L49/9036
MESSAGE ORDERING BUFFER
The disclosed embodiments, collectively referred to as the “Message Ordering Buffer” or “MOB”, relate to an improved messaging platform, or processing system, which may also be referred to as a message processing architecture or platform, which routes messages from a publisher to a subscriber ensuring related messages, e.g., ordered messages, are conveyed to a single recipient, e.g., processing thread, without unnecessarily committing resources of the architecture to that recipient or otherwise preventing message transmission to other recipients. The disclosed embodiments further include additional features which improve efficient and facilitate deployment in different application environments. The disclosed embodiments may be deployed as a message oriented middleware component directly installed, or accessed as a service, and accessed by publishers and subscribers, as described herein, so as to electronically exchange messages therebetween.
METHOD AND APPARATUS FOR WIRELESS COMMUNICATION OF LOW LATENCY DATA BETWEEN MULTILINK DEVICES
In a multilink system, block acknowledgment is performed at TID level. It is commonly accepted that acknowledgment policy refrains low latency delivery. However, it is often that traffic streams combine low-latency data and data requiring a high level of reliability. A need exists to facilitate the transmission of low-latency data units of a TID when a multilink operation, such as in 802.11be, is implemented with an acknowledgment policy. One specific link from amongst the multiple links may be dedicated to low latency data for which the originator and the recipient no longer take into account the acknowledgment. It turns that the originator can remove the low latency data from its transmit buffer immediately after transmission, while the recipient can deliver these data to the upper layer without waiting for the good reception of all preceding data, according to a sequence numbering. Low-latency services are consequently improved.
System And Method For Using Dynamic Thresholds With Route Isolation For Heterogeneous Traffic In Shared Memory Packet Buffers
A system and method for managing shared memory packet buffers is disclosed. In some embodiments, the system is configured to receive and classify a packet as one of: network-network, network-host, host-network, or host-host; select a minimum guarantee space for the packet according to the classification thereof; if the selected minimum guarantee space is available, store the packet therein; otherwise, if a dedicated shared space is available, store the packet therein; otherwise, if a global shared space is available, store the packet therein; and otherwise, drop the packet.
COMBINED INPUT AND OUTPUT QUEUE FOR PACKET FORWARDING IN NETWORK DEVICES
An apparatus for switching network traffic includes an ingress packet forwarding engine and an egress packet forwarding engine. The ingress packet forwarding engine is configured to determine, in response to receiving a network packet, an egress packet forwarding engine for outputting the network packet and enqueue the network packet in a virtual output queue. The egress packet forwarding engine is configured to output, in response to a first scheduling event and to the ingress packet forwarding engine, information indicating the network packet in the virtual output queue and that the network packet is to be enqueued at an output queue for an output port of the egress packet forwarding engine. The ingress packet forwarding engine is further configured to dequeue, in response to receiving the information, the network packet from the virtual output queue and enqueue the network packet to the output queue.
Packet filtering using binary search trees
A packet filtering system uses linked zero-based binary search trees to filter received packets. The binary search trees may be generated from filter conditions defining filter parameters for filtering packets.
Message ordering buffer
The disclosed embodiments, collectively referred to as the “Message Ordering Buffer” or “MOB”, relate to an improved messaging platform, or processing system, which may also be referred to as a message processing architecture or platform, which routes messages from a publisher to a subscriber ensuring related messages, e.g., ordered messages, are conveyed to a single recipient, e.g., processing thread, without unnecessarily committing resources of the architecture to that recipient or otherwise preventing message transmission to other recipients. The disclosed embodiments further include additional features which improve efficient and facilitate deployment in different application environments. The disclosed embodiments may be deployed as a message oriented middleware component directly installed, or accessed as a service, and accessed by publishers and subscribers, as described herein, so as to electronically exchange messages therebetween.
Congestion drop decisions in packet queues
A packet switch includes an ingress port; queue admission control circuitry connected to the ingress port; one or more egress queues configured to manage packet buffers; and an egress port connected to the packet buffers, wherein the packet buffers are managed such that already queued lower priority packets are discarded from the packet buffers when it is required to drop higher priority packets that should otherwise be accepted in the packet buffers. The queue admission control circuitry can be configured to determine if a packet should be dropped or not, and the queue admission control circuitry communicates to buffer reallocation circuitry that is configured to discard one or more lower priority packets to support enqueuing the higher priority packet.
System and method for facilitating efficient management of non-idempotent operations in a network interface controller (NIC)
A network interface controller (NIC) capable of efficient management of non-idempotent operations is provided. The NIC can be equipped with a network interface, storage management logic block, and an operation management logic block. During operation, the network interface can receive a request for an operation from a remote device. The storage management logic block can store, in a local data structure, outcome of operations executed by the NIC. The operation management logic block can determine whether the NIC has previously executed the operation. If the NIC has previously executed the operation, the operation management logic block can obtain an outcome of the operation from the data structure and generate a response comprising the obtained outcome for responding to the request.
Cyclic Queuing and Forwarding (CQF) Segmentation
A method for communicating time sensitive data streams in a network. The method includes synchronizing queuing and transmission of data in a set of output buffers for buffering streams associated with a certain class of service. The method includes processing data packets of a stream using a cyclic flow meter, wherein the cyclic flow meter limits transfer of the data packets of the stream to the set of output buffers of the network node according to a predetermined amount per cycle based on an output frequency. The method includes transmitting, according to the output frequency, the data packets of the stream from a non-empty output buffer of the set of output buffers, and transferring the data packets of the stream from the cyclic flow meter to an empty output buffer of the set of output buffers.
Combined input and output queue for packet forwarding in network devices
An apparatus for switching network traffic includes an ingress packet forwarding engine and an egress packet forwarding engine. The ingress packet forwarding engine is configured to determine, in response to receiving a network packet, an egress packet forwarding engine for outputting the network packet and enqueue the network packet in a virtual output queue. The egress packet forwarding engine is configured to output, in response to a first scheduling event and to the ingress packet forwarding engine, information indicating the network packet in the virtual output queue and that the network packet is to be enqueued at an output queue for an output port of the egress packet forwarding engine. The ingress packet forwarding engine is further configured to dequeue, in response to receiving the information, the network packet from the virtual output queue and enqueue the network packet to the output queue.