Patent classifications
H04L47/6235
PROCESSING QUEUE ASSIGNMENT BASED ON A FLAG IN A GENEVE HEADER
Described herein are systems, methods, and software to manage the identification of control packets in an encapsulation header. In one implementation, a computing system may receive a Geneve packet at a network interface and determine that the Geneve packet includes an Operations and Management (OAM) flag. Once the OAM flag is identified, the computing system can select a processing queue from a plurality of processing queues for a main processing system of the computing system based on the OAM flag and assign the Geneve packet to the processing queue.
DATA PROCESSING METHOD, DATA PROCESSING APPARATUS, ELECTRONIC DEVICE, STORAGE MEDIUM, AND PROGRAM PRODUCT
Provided in the present disclosure are a data processing method and apparatus, and an electronic device, the method includes: determining a plurality of candidate data pieces, where the candidate data pieces are provided from corresponding data sources; and determining a target data piece based on priorities of the data sources corresponding to the plurality of candidate data pieces in a current cycle, wherein a same data source has different priorities in different processing cycles, and priority sequence numbers of a same data source in different processing cycles satisfy a nonlinear relationship.
IN-ORDER PROCESSING OF NETWORK PACKETS
The described technology relates to a real-time processing of network packets. An example system relates to reordering messages received at a server over a communication network from distributed clients, in order to, among other things, eliminate or at least substantially reduce the effects of jitter (delay variance) experienced in the network. The reordering of messages may enable the example data processing application to improve the consistency of processing packets in the time order of when the respective packets entered a geographically distributed network.
METHOD AND SYSTEM FOR FACILITATING LOSSY DROPPING AND ECN MARKING
Methods and systems are provided for performing lossy dropping and ECN marking in a flow-based network. The system can maintain state information of individual packet flows, which can be set up or released dynamically based on injected data. Each flow can be provided with a flow-specific input queue upon arriving at a switch. Packets of a respective flow are acknowledged after reaching the egress point of the network, and the acknowledgement packets are sent back to the ingress point of the flow along the same data path. As a result, each switch can obtain state information of each flow and perform per-flow packet dropping and ECN marking.
Data center network having optical permutors
A network system for a data center is described in which a switch fabric may provide full mesh interconnectivity such that any servers may communicate packet data to any other of the servers using any of a number of parallel data paths. Moreover, according to the techniques described herein, edge-positioned access nodes, optical permutation devices and core switches of the switch fabric may be configured and arranged in a way such that the parallel data paths provide single L2/L3 hop, full mesh interconnections between any pairwise combination of the access nodes, even in massive data centers having tens of thousands of servers. The plurality of optical permutation devices permute communications across the optical ports based on wavelength so as to provide, in some cases, full-mesh optical connectivity between edge-facing ports and core-facing ports.
DEVICE AND METHOD FOR QUEUES RELEASE AND OPTIMIZATION BASED ON RUN-TIME ADAPTIVE AND DYNAMIC INTERNAL PRIORITY VALUE STRATEGY
The present disclosure relates to controlling queue release in a network. In particular, the disclosure proposes a controller configured to obtain a state of each of a plurality of queues of a network node and determine, based on the states of the queues, whether the utilization of one or more queues exceeds one or more thresholds. If one or more thresholds are exceeded, the controller is configured to generate one or more new priority entries for one or more queues of the plurality of queues and provide the one or more new priority entries to the one or more queues of the network node. Further, the disclosure proposes a network node being configured to provide a state of each of a plurality of queues to a controller, and obtain one or more new priority entries for one or more queues of the plurality of queues from the controller.
DEVICE AND METHOD FOR QUEUES RELEASE AND OPTIMIZATION BASED ON RUN-TIME ADAPTIVE AND DYNAMIC GATE CONTROL LIST STRATEGY
A controller is configured to: obtain a state of each of a plurality of queues of a network node; determine, based on the states of the queues, whether the utilization of one or more queues exceeds one or more thresholds; generate one or more new entries for a gate control list of the network node that controls the plurality of queues, if one or more thresholds are exceeded; and provide the one or more new entries to the network node. Further, a network node is configured to provide a state of each of a plurality of queues to a controller, and obtain one or more new entries for a gate control list of the network node that controls the plurality of queues, from the controller.
SYSTEM AND METHOD FOR FACILITATING DATA-DRIVEN INTELLIGENT NETWORK WITH INGRESS PORT INJECTION LIMITS
Data-driven intelligent networking systems and methods are provided. The system can accommodate dynamic traffic while applying injection limits to different traffic classes at an ingress edge port. The system can maintain state information of individual packet flows, which can be set up or released dynamically based on injected data. Each flow can be provided with a flow-specific input queue upon arriving at a switch. Packets of a respective flow can be acknowledged after reaching the egress point of the network, and the acknowledgement packets can be sent back to the ingress point of the flow along the same data path. Furthermore, an edge switch can dynamically allocate the ingress port bandwidth among the traffic classes that are active at a given moment.
RECEPTION INTERFACE CIRCUITS SUPPORTING MULTIPLE COMMUNICATION STANDARDS AND MEMORY SYSTEMS INCLUDING THE SAME
A reception interface circuit includes a termination circuit, a buffer and an interface controller. The termination circuit is configured to change a termination mode in response to a termination control signal. The buffer is configured to change a reception characteristic in response to a buffer control signal. The interface controller is configured to generate the termination control signal and the buffer control signal such that the reception characteristic of the buffer is changed in association with the change in the termination mode. The reception interface circuit may support various communication standards by changing the reception characteristic of the buffer in association with the termination mode. Using the reception interface circuit, communication efficiency of transceiver systems such as a memory system and/or compatibility between a transmitter device and a receiver device may be improved.
SYSTEM AND METHOD FOR FACILITATING EFFICIENT PACKET FORWARDING IN A NETWORK INTERFACE CONTROLLER (NIC)
A network interface controller (NIC) capable of efficient packet forwarding is provided. The NIC can be equipped with a host interface, a packet generation logic block, and a forwarding logic block. During operation, the packet generation logic block can obtain, via the host interface, a message from the host device and for a remote device. The packet generation logic block may generate a plurality of packets for the remote device from the message. The forwarding logic block can then send a first subset of packets of the plurality of packets based on ordered delivery. If a first condition is met, the forwarding logic block can send a second subset of packets of the plurality of packets based on unordered delivery. Furthermore, if a second condition is met, the forwarding logic block can send a third subset of packets of the plurality of packets based on ordered delivery.