H04L49/15

Provisioning logical entities in a multidatacenter environment

A system provisions global logical entities that facilitate the operation of logical networks that span two or more datacenters. These global logical entities include global logical switches that provide L2 switching as well as global routers that provide L3 routing among network nodes in multiple datacenters. The global logical entities operate along side local logical entities that are for operating logical networks that are local within a datacenter.

Provisioning logical entities in a multidatacenter environment

A system provisions global logical entities that facilitate the operation of logical networks that span two or more datacenters. These global logical entities include global logical switches that provide L2 switching as well as global routers that provide L3 routing among network nodes in multiple datacenters. The global logical entities operate along side local logical entities that are for operating logical networks that are local within a datacenter.

CUSTOMER ACTIVATION ON EDGE COMPUTING ENVIRONMENT

A system and method for providing on-demand edge compute. The system may include an orchestrator that provides a UI and that controls an abstraction layer for implementing a workflow for providing on-demand edge compute. The abstraction layer may include a server configuration orchestration (SCO) system (e.g., a Metal-as-a-Service (MaaS) system) and API that may provide an interface between the orchestrator and the SCO. The API may enable the orchestrator to communicate with the SCO for receiving requests that enable the SCO to integrate with existing compute resources to perform various compute provisioning tasks (e.g., to build and provision a server instance). The various tasks, when executed, may provide on-demand edge compute service to users. The SCO API may further enable the ECS orchestrator to receive information from the SCO (e.g., compute resource information, status messages).

CUSTOMER ACTIVATION ON EDGE COMPUTING ENVIRONMENT

A system and method for providing on-demand edge compute. The system may include an orchestrator that provides a UI and that controls an abstraction layer for implementing a workflow for providing on-demand edge compute. The abstraction layer may include a server configuration orchestration (SCO) system (e.g., a Metal-as-a-Service (MaaS) system) and API that may provide an interface between the orchestrator and the SCO. The API may enable the orchestrator to communicate with the SCO for receiving requests that enable the SCO to integrate with existing compute resources to perform various compute provisioning tasks (e.g., to build and provision a server instance). The various tasks, when executed, may provide on-demand edge compute service to users. The SCO API may further enable the ECS orchestrator to receive information from the SCO (e.g., compute resource information, status messages).

TECHNOLOGIES FOR SWITCHING NETWORK TRAFFIC IN A DATA CENTER

Technologies for switching network traffic include a network switch. The network switch includes one or more processors and communication circuitry coupled to the one or more processors. The communication circuity is capable of switching network traffic of multiple link layer protocols. Additionally, the network switch includes one or more memory devices storing instructions that, when executed, cause the network switch to receive, with the communication circuitry through an optical connection, network traffic to be forwarded, and determine a link layer protocol of the received network traffic. The instructions additionally cause the network switch to forward the network traffic as a function of the determined link layer protocol. Other embodiments are also described and claimed.

Automated network change system

Methods, systems, and apparatus, for automatically changing a network system. A method includes receiving a set of first intents that describe a state of a first switch fabric; receiving a set of second intents that describe a state of a second switch fabric; computing a set of network operations to perform on the first switch fabric to achieve the second switch fabric, the set of operations also defining an order in which the operations are to be executed, and the set of operations determined based on the set of first intents, the set of second intents, and migration logic that defines a ruleset for selecting the operations based on the set of first intents and the second intents; and executing the set of network operations according to the order, to apply changes to elements within the first switch fabric to achieve the state of the second switch fabric.

Automated network change system

Methods, systems, and apparatus, for automatically changing a network system. A method includes receiving a set of first intents that describe a state of a first switch fabric; receiving a set of second intents that describe a state of a second switch fabric; computing a set of network operations to perform on the first switch fabric to achieve the second switch fabric, the set of operations also defining an order in which the operations are to be executed, and the set of operations determined based on the set of first intents, the set of second intents, and migration logic that defines a ruleset for selecting the operations based on the set of first intents and the second intents; and executing the set of network operations according to the order, to apply changes to elements within the first switch fabric to achieve the state of the second switch fabric.

CONNECTING PROCESSORS USING TWISTED TORUS CONFIGURATIONS
20230094933 · 2023-03-30 ·

Methods, systems, and apparatus, including computer programs encoded on computer-storage media, for connecting processors using twisted torus configurations. In some implementations, a cluster of processing nodes is coupled using a reconfigurable interconnect fabric. The system determines a number of processing nodes to allocate as a network within the cluster and a topology for the network. The system selects an interconnection scheme for the network, where the interconnection scheme is selected from a group that includes at least a torus interconnection scheme and a twisted torus interconnection scheme. The system allocates the determined number of processing nodes of the cluster in the determined topology, sets the reconfigurable interconnect fabric to provide the selected interconnection scheme for the processing nodes in the network, and provides access to the network for performing a computing task.

CONNECTING PROCESSORS USING TWISTED TORUS CONFIGURATIONS
20230094933 · 2023-03-30 ·

Methods, systems, and apparatus, including computer programs encoded on computer-storage media, for connecting processors using twisted torus configurations. In some implementations, a cluster of processing nodes is coupled using a reconfigurable interconnect fabric. The system determines a number of processing nodes to allocate as a network within the cluster and a topology for the network. The system selects an interconnection scheme for the network, where the interconnection scheme is selected from a group that includes at least a torus interconnection scheme and a twisted torus interconnection scheme. The system allocates the determined number of processing nodes of the cluster in the determined topology, sets the reconfigurable interconnect fabric to provide the selected interconnection scheme for the processing nodes in the network, and provides access to the network for performing a computing task.

METHODS AND APPARATUS FOR COMMUNICATING BETWEEN NODE DEVICES

Aspects of the present disclosure relate to an interconnect comprising interfaces to communicate with respective requester and receiver node devices, and home nodes. Each home node is configured to: receive requests from one or more requester nodes, each request comprising a target address corresponding to a target receiver nodes; and transmit each said request to the corresponding target receiver node. Mapping circuitry is configured to: associate each of said plurality of home nodes with a given home node cluster; perform a first hashing of the target address of a given request, to determine a target cluster; perform a second hashing of the target address, to determine a target home node within said target cluster; and direct the given message, to the target home node.