Patent classifications
H04L49/1515
Efficient highly connected deployment units
Efficient and highly-scalable network solutions are provided that each utilize deployment units based on Clos networks, but in an environment such as a data center of Internet Protocol-based network. Each of the deployment units can include multiple stages of devices, where connections between devices are only made between stages and the deployment units are highly connected. In some embodiments, the level of connectivity between two stages can be reduced, providing available connections to add edge switches and additional host connections while keeping the same number of between-tier connections. In some embodiments, where deployment units (or other network groups) can be used at different levels to connect other deployment units, the edges of the deployment units can be fused to reduce the number of devices per host connection.
Methods and apparatus for flow control associated with a switch fabric
In some embodiments, an apparatus includes a switch fabric having at least a first switch stage and a second switch stage, an edge device operatively coupled to the switch fabric and a management module. The edge device is configured to send a first portion of a data stream to the switch fabric such that the first portion of the data stream is received at a queue of the second switch stage of the switch fabric via the first switch stage of the switch fabric. The management module is configured to send a flow control signal configured to trigger the edge device to suspend transmission of a second portion of the data stream when a congestion level of the queue of the second switch stage of the switch fabric satisfies a condition in response to the first portion of the data stream being received at the queue.
LOOP CONFLICT AVOIDANCE IN A NETWORK COMPUTING ENVIRONMENT
Systems, methods, and devices for improved routing operations in a network computing environment. A system includes a network topology comprising a plurality of spine nodes and a plurality of leaf nodes, wherein a link between a first spine node and a first leaf node is inactive. The first spine node includes one or more processors configurable to execute instructions stored in non-transitory computer readable storage media. The instructions include receiving a packet to be transmitted to the first leaf node. The instructions include identifying an alternative spine node at a same level in the network topology. The instructions include attaching a tunnel label to the packet, wherein the tunnel label indicates the packet should be transmitted to the alternative spine node.
MULTIPLE NETWORK INTERFACING
This disclosure describes techniques for enabling interoperability between asymmetric and symmetric Integrated Routing and Bridging (IRB) modes. An interfacing component may be configured to receive a first route advertisement from a first edge node in a Layer-2 (L2) fabric. The first route advertisement may correspond to an asymmetric format route, for instance. The interfacing component may be further configured to receive a second route advertisement from a second edge node in a L2/Layer-3 (L3) fabric. The second edge node may be configured for symmetric integrated routing and bridging (IRB). The interfacing component may be configured to re-originate the first route and the second route such that the interfacing component is included as a hop in the resultant routes between the L2 fabric and the L2/L3 fabric.
Redundant multicast tree in a fat tree network topology
A method includes identifying within a network topology, by an apparatus, a plurality of network devices; and establishing by the apparatus, a multiple tree topology comprising a first multicast tree and a second multicast tree, the first and second multicast trees operable as redundant trees for multicast traffic in the network topology, the establishing including: allocating a first of the network devices as a corresponding root of the first multicast tree, allocating a first group of intermediate devices from the network devices as first forwarding devices in the first multicast tree, allocating a second group of intermediate devices as belonging to first leaf devices in the first multicast tree, and allocating terminal devices of the network devices as belonging to the first leaf devices, and allocating a second of the network devices as the corresponding root of the second multicast tree, allocating the second group of intermediate devices as second forwarding devices in the second multicast tree, allocating the first group of intermediate devices as belonging to second leaf devices in the second multicast tree, and allocating the terminal devices as belonging to the second leaf devices.
NETWORK CONGESTION CONTROL METHOD, NODE, SYSTEM, AND STORAGE MEDIUM
A network congestion control method, a node and a system are disclosed, where the method is applied to a spine-leaf network system. The method includes: a spine node receives network information sent by the at least one leaf node, where the network information includes network topology information of the leaf node and a network performance indicator of the leaf node; networks the at least one leaf node and the spine node based on the network topology information of the at least one leaf node, to obtain a combined network topology; and if the combined network topology is a global network topology of the spine-leaf network system, performs network congestion control on the at least one leaf node based on the network performance indicator of the at least one leaf node.
COMMUNICATION NETWORK HOPPING ARCHITECTURE
Communication network systems are disclosed. In one or more implementations, the communication network system includes a plurality of network devices. Each of the plurality of network devices incorporates one or more multi-port switches, where each multi-port switch includes a connection to the network device incorporating the multi-port switch and a connection to at least one other port of another multi-port switch incorporated by another respective one of the plurality of network devices.
IDENTIFYING COMPONENTS FOR REMOVAL IN A NETWORK CONFIGURATION
Systems, methods, and computer-readable media analyzing memory usage in a network node. A network assurance appliance may be configured to determine a hit count for a concrete level rule implemented on a node and identify one or more components of a logical model, wherein each of the one or more components are associated with the concrete level rule. The network assurance appliance may attribute the hit count for the concrete level rule to each of the components of the logical model, determine a number of hardware level entries associated with the each of the one or more components, and generate a report comprising the one or more components of the logical model, the hit count attributed to each of the one or more components of the logical model, and the number of hardware level entries associated with the one or more components of the logical model.
Methods and apparatus to manage a physical network to reduce network dependencies in a multi-fabric virtual network
A disclosed example of managing a network includes a packet receiver to receive a packet at a second switch from a first switch via an inter-switch link between the first and second switches; a packet analyzer to determine whether the packet includes a load-based teaming (LBT) egress control value; a packet transmitter to handle the packet according to an LBT policy when the packet analyzer determines the packet does not include the LBT egress control value; and a packet modifier to terminate the packet at the second switch when the packet analyzer determines the packet does include the LBT egress control value.
Identifying components for removal in a network configuration
Systems, methods, and computer-readable media analyzing memory usage in a network node. A network assurance appliance may be configured to determine a hit count for a concrete level rule implemented on a node and identify one or more components of a logical model, wherein each of the one or more components are associated with the concrete level rule. The network assurance appliance may attribute the hit count for the concrete level rule to each of the components of the logical model, determine a number of hardware level entries associated with the each of the one or more components, and generate a report comprising the one or more components of the logical model, the hit count attributed to each of the one or more components of the logical model, and the number of hardware level entries associated with the one or more components of the logical model.