H04L47/627

System and methods to filter out noisy application signatures to improve precision of first packet classification
11582158 · 2023-02-14 ·

The systems and methods discussed herein provide for classifying CDN connections to the originating application on the first packet. In some implementations, the system identifies application connections established within a predetermined time period prior to the CDN connection and increments a value associated with these connections. The system classifies the CDN connection as corresponding to the application connection with the highest associated value, allowing routing of network traffic to take advantage of QoS benefits and reduce the need for deep packet inspection.

Communication control apparatus and communication control method

A communication control apparatus includes an observed data rate acquiring unit configured to acquire observed data rates indicating input data rates of queues, the observed data rates being observed in a layer 2 switch, a threshold value storage unit configured to store predetermined threshold values of the queues, a shaping rate determination unit configured to determine a shaping rate of each queue based on both observed data based on an observed data rate of the observed data rates acquired by the observed data rate acquiring unit and a threshold value of the predetermined threshold values stored in the threshold value storage unit, and a shaping rate setting unit configured to set, in the layer 2 switch, the shaping rate of each queue determined by the shaping rate determination unit.

Spatial dispersion buffer
11522817 · 2022-12-06 · ·

An improved buffer for networking and other computing devices comprises multiple memory instances, each having a distinct set of entries. Transport data units (“TDUs”) are divided into storage data units (“SDUs”), and each SDU is stored within a separate entry of a separate memory instance in a logical bank. One or more grids of the memory instances are organized into overlapping logical banks. The logical banks are arranged into views. Different destinations or other entities are assigned different views of the buffer. A memory instance may be shared between logical banks in different views. When overlapping logical banks are accessed concurrently, data in a memory instance that they share may be recovered using a parity SDU in another memory instance. The shared buffering enables more efficient buffer usage in a network device with a traffic manager shared amongst egress bocks. Example read and write algorithms for such buffers are disclosed.

Throttling queue for a request scheduling and processing system

Various methods and systems for implementing request scheduling and processing in a multi-tenant distributed computing environment are provided. Requests to utilize system resources in the distributed computing environment are stored in account queues corresponding to tenant accounts. If storing a request in an account queue would exceed a throttling threshold such as a limit on the number of requests stored per account, the request is dropped to a throttling queue. A scheduler prioritizes processing requests stored in the processing queue before processing requests stored in the account queues. The account queues can be drained using dominant resource scheduling. In some embodiments, a request is not picked up from an account queue if processing the request would exceed a predefined hard limit on system resource utilization for the corresponding tenant account. In some embodiments, the hard limit is defined as a percentage of threads the system has to process requests.

THROTTLING QUEUE FOR A REQUEST SCHEDULING AND PROCESSING SYSTEM

Various methods and systems for implementing request scheduling and processing in a multi-tenant distributed computing environment are provided. Requests to utilize system resources in the distributed computing environment are stored in account queues corresponding to tenant accounts. If storing a request in an account queue would exceed a throttling threshold such as a limit on the number of requests stored per account, the request is dropped to a throttling queue. A scheduler prioritizes processing requests stored in the processing queue before processing requests stored in the account queues. The account queues can be drained using dominant resource scheduling. In some embodiments, a request is not picked up from an account queue if processing the request would exceed a predefined hard limit on system resource utilization for the corresponding tenant account. In some embodiments, the hard limit is defined as a percentage of threads the system has to process requests.

Communication Method and Related Device
20230098662 · 2023-03-30 ·

A communication method includes: sending, by a policy control network element, first query information to a data analytics network element, where the first query information is used to obtain service quality information; and receiving, by the policy control network element, first response information sent by the data analytics network element, where the first response information includes the quality information that is of the service and that is requested using the first query information.

Regulating enqueueing and dequeuing border gateway protocol (BGP) update messages

A network device, associated with peer network devices, may receive policy information for a protocol; and compute a first update message based on information regarding a route associated with the policy information. The network device may determine that an upper utilization threshold for one or more of peer queues, associated with the peer network devices, is not satisfied; and write the first update message to the peer queues based on determining that the upper utilization threshold is not satisfied. The network device may compute a second update message based on the information regarding the route; determine that the upper utilization threshold for one or more of the peer queues is satisfied; and pause writing the second update message to the peer queues based on the upper utilization threshold being satisfied. The network device may permit the peer network devices to obtain data from corresponding ones of the peer queues.

DATA COPY ACCELERATION FOR SERVICE MESHES
20230077147 · 2023-03-09 ·

Examples described herein relate to a system for accelerating data operations of a service mesh using a data mover accelerator.

Communication method and related device

A communication method includes: sending, by a policy control network element, first query information to a data analytics network element, where the first query information is used to obtain service quality information; and receiving, by the policy control network element, first response information sent by the data analytics network element, where the first response information includes the quality information that is of the service and that is requested using the first query information.

Automatic flow management

Packet-switching operations in a network device are managed based on the detection of excessive-rate traffic flows. A network device receives a data unit, determines the traffic flow to which the data unit belongs, and updates flow tracking information for that flow. The network device utilizes the tracking information to determine when a rate at which the network device is receiving data belonging to the flow exceeds an excessive-rate threshold and is thus an excessive-rate flow. The network device may enable one or more excessive-rate policies on an excessive-rate traffic flow. Such a policy may include any number of features that affect how the device handles data units belonging to the flow, such as excessive-rate notification, differentiated discard, differentiated congestion notification, and reprioritization. Memory and other resource optimizations for such flow tracking and management are also described.