H04L49/503

DISTRIBUTED VIRTUAL GATEWAY APPLIANCE

Methods, computer program products, and systems are presented. The methods include, for instance: providing a distributed virtual gateway for Network Virtualization over Layer 3 (NVO3) network. A gateway stack having three or more nodes is implemented as a distributed virtual gateway, providing Layer 2 or Layer 3 gateway services in a fail-safe manner. Nodes of the gateway stack are configured to autonomously process and forward inbound NVO3 data packets with known destination addresses without engaging a master of the gateway stack.

Buffer management method and apparatus

A memory management method includes: determining that available storage space of a first memory in a network device is less than a first threshold, where the first threshold is greater than 0 and the first memory stores a first packet queue; and deleting at least one packet at the tail of the first packet queue from the first memory based on the available storage space of the first memory being less than the first threshold. When the available storage space of the first memory is less than the first threshold, a packet queue, namely, the first packet queue, is selected and a packet at the tail of the packet queue is deleted from the first memory.

GRADED THROTTLING FOR NETWORK-ON-CHIP TRAFFIC
20230036142 · 2023-02-02 ·

Graded throttling for network-on-chip traffic, including: calculating, by an agent of a network-on-chip, a number of outstanding transactions issued by the agent; determining that the number of outstanding transactions meets a threshold; and implementing, by the agent, in response to the number of outstanding transactions meeting the threshold, a traffic throttling policy.

Overload protection for data sinks in a distributed computing system

Described in this document, among other things, is an overload protection system that can protect data sinks from overload by controlling the volume of data sent to those data sinks in a fine-grained manner. The protection system preferably sits in between edge servers, or other producers of data, and data sinks that will receive some or all of the data. Preferably, each data sink owner defines a policy to control how and when overload protection will be applied. Each policy can include definitions of how to monitor the stream of data for overload and specify one or more conditions upon which throttling actions are necessary. In embodiments, a policy can contain a multi-part specification to identify the class(es) of traffic to monitor to see if the conditions have been triggered.

System and method for efficient network isolation and load balancing in a multi-tenant cluster environment

A system and method for supporting load balancing in a multi-tenant cluster environment, in accordance with an embodiment. One or more tenants can be supported and each associated with a partition, which are each in turn associated with one or more end nodes. The method can provide a plurality of switches, the plurality of switches comprising a plurality of leaf switches and at least one switch at another level, wherein each of the plurality of switches comprise at least one port. The method can assign each node a weight parameter, and based upon this parameter, the method can route the plurality of end nodes within the multi-tenant cluster environment, wherein the routing attempts to preserve partition isolation.

OVERLOAD PROTECTION FOR DATA SINKS IN A DISTRIBUTED COMPUTING SYSTEM

Described in this document, among other things, is an overload protection system that can protect data sinks from overload by controlling the volume of data sent to those data sinks in a fine-grained manner. The protection system preferably sits in between edge servers, or other producers of data, and data sinks that will receive some or all of the data. Preferably, each data sink owner defines a policy to control how and when overload protection will be applied. Each policy can include definitions of how to monitor the stream of data for overload and specify one or more conditions upon which throttling actions are necessary. In embodiments, a policy can contain a multi-part specification to identify the class(es) of traffic to monitor to see if the conditions have been triggered.

Method for operating an ethernet communication device, and ethernet communication device
11252107 · 2022-02-15 · ·

A method for operating an Ethernet communication device having multiple external physical interfaces for a motor vehicle includes the following steps: detecting a special state on at least one of the interfaces, generating a control signal to adapt the interface in the special state if the special state is detected; and blocking a forwarding of a message packet arriving on the interface in the special state to a media access control unit of the Ethernet communication device on the basis of the control signal.

BUFFER MANAGEMENT METHOD AND APPARATUS
20210392092 · 2021-12-16 ·

A memory management method includes: determining that available storage space of a first memory in a network device is less than a first threshold, where the first threshold is greater than 0 and the first memory stores a first packet queue; and deleting at least one packet at the tail of the first packet queue from the first memory based on the available storage space of the first memory being less than the first threshold. When the available storage space of the first memory is less than the first threshold, a packet queue, namely, the first packet queue, is selected and a packet at the tail of the packet queue is deleted from the first memory.

CONVERGED AVIONICS DATA NETWORK
20220131598 · 2022-04-28 ·

An apparatus and method for operating an avionics data network includes a network switch core configured for a time-sensitive networking (TSN) schema, a set of ARINC 664 (A664) and a set of TSN networking end nodes communicatively connected with the network switch core. The network switch core is configured to receive, from the first set of networking end nodes, a set of data frames, determine the respective schema of the set of data frames, police the set of data frames based on the determined respective schema using a set of predetermined rules, forward the set of data frames to a predetermined queue on an egress port of the network switch core based on the determined respective schema, and transmit set of data frames to an end node of the second set of networking end nodes having a corresponding schema.

Maintaining bandwidth utilization in the presence of packet drops

Examples describe a manner of scheduling packet segment fetches at a rate that is based on one or more of: a packet drop indication, packet drop rate, incast level, operation of queues in SAF or VCT mode, or fabric congestion level. Headers of packets can be fetched faster than payload or body portions of packets and processed prior to queueing of all body portions. In the event a header is identified as droppable, fetching of the associated body portions can be halted and any body portion that is queued can be discarded. Fetch overspeed can be applied for packet headers or body portions associated with packet headers that are approved for egress.