G06F15/161

Cloud-based scale-up system composition

Technologies for composing a managed node with multiple processors on multiple compute sleds to cooperatively execute a workload include a memory, one or more processors connected to the memory, and an accelerator. The accelerator further includes a coherence logic unit that is configured to receive a node configuration request to execute a workload. The node configuration request identifies the compute sled and a second compute sled to be included in a managed node. The coherence logic unit is further configured to modify a portion of local working data associated with the workload on the compute sled in the memory with the one or more processors of the compute sled, determine coherence data indicative of the modification made by the one or more processors of the compute sled to the local working data in the memory, and send the coherence data to the second compute sled of the managed node.

Data center network device sensing

Provided is a process of managing rack-mounted computing devices in a data center with a distributed peer-to-peer management system, the process including: determining roles of data-center management computing devices in a distributed peer-to-peer data-center management system; receiving, via an out-of-band network, a data-center management command at a given data-center management computing device; distributing, based on at least some of the roles, via the out-of-band network, the data-center management command.

MULTI-CLUSTER BOOT-STRAPPING

Disclosed herein are system, method, and computer program product embodiments for multi-cluster boot-strapping. In some embodiments, a server residing on a primary computing cluster receives a first request to establish a temporary connection between the primary computing cluster and a secondary computing cluster. The server establishes the temporary connection between the primary computing cluster and the secondary computing cluster using the first set of credentials. Furthermore, the server receives a second request to establish a persistent connection between the primary computing cluster and the secondary computing cluster. The server establishes the persistent connection by transmitting a third request comprising the configuration settings to the secondary computing cluster thereby causing the secondary computing cluster to generate a second set of credentials corresponding to the primary computing cluster. The server receives and stores the second set of credentials.

Portable blockchain mining systems and methods of use
11659682 · 2023-05-23 · ·

Portable blockchain mining systems and methods of use are discussed here. Systems include a portable building; a plurality of blockchain mining processors mounted within, or a plurality of blockchain mining processor mounts located within, an interior of the portable building; an air inlet defined in the portable building; and an air outlet defined in the portable building. Air outlets may be above the air inlet and oriented to direct exhaust air in an upward direction out of the portable building. A cooling fan may be connected to convey air through the air inlet, across the plurality of blockchain mining processors and out the air outlet. The cooling fan may simultaneously cool a genset and processors 72. Compact, stackable mining modules are discussed.

Techniques to configure physical compute resources for workloads via circuit switching

Embodiments are generally directed apparatuses, methods, techniques and so forth to select two or more processing units of the plurality of processing units to process a workload, and configure a circuit switch to link the two or more processing units to process the workload, the two or more processing units each linked to each other via paths of communication and the circuit switch.

METHOD OF MANAGING A NETWORK OF CALCULATION NODES
20170353358 · 2017-12-07 ·

A method of managing a network of calculation nodes interconnected by a plurality of interconnection devices, includes organizing the calculation nodes into groups of calculation nodes, for each group of calculation nodes, connecting the interconnection devices interconnecting the nodes of the group to a group management node, the management node being dedicated to the group of calculation nodes on each management node execution of an administration function by the implementation of independent management modules, each management module of a management node being able to communicate with the other management modules of the same management node.

CLOUD-BASED SCALE-UP SYSTEM COMPOSITION

Technologies for composing a managed node with multiple processors on multiple compute sleds to cooperatively execute a workload include a memory, one or more processors connected to the memory, and an accelerator. The accelerator further includes a coherence logic unit that is configured to receive a node configuration request to execute a workload. The node configuration request identifies the compute sled and a second compute sled to be included in a managed node. The coherence logic unit is further configured to modify a portion of local working data associated with the workload on the compute sled in the memory with the one or more processors of the compute sled, determine coherence data indicative of the modification made by the one or more processors of the compute sled to the local working data in the memory, and send the coherence data to the second compute sled of the managed node.

Simultaneous multi-processor apparatus applicable to achieving exascale performance for algorithms and program systems
11669418 · 2023-06-06 · ·

Apparatus adapted for exascale computers are disclosed. The apparatus includes, but is not limited to at least one of: a system, data processor chip (DPC), Landing module (LM), chips including LM, anticipator chips, simultaneous multi-processor (SMP) cores, SMP channel (SMPC) cores, channels, bundles of channels, printed circuit boards (PCB) including bundles, floating point adders, accumulation managers, QUAD Link Anticipating Memory (QUADLAM), communication networks extended by coupling links of QUADLAM, log2 calculators, exp2 calculators, logALU, Non-Linear Accelerator (NLA), and stairways. Methods of algorithm and program development, verification and debugging are also disclosed. Collectively, embodiments of these elements disclose a class of supercomputers that obsolete Amdahl's Law, providing cabinets of petaflop performance and systems that may meet or exceed an exaflop of performance for Block LU Decomposition (Linpack).

System for provisioning racks autonomously in data centers

A provisioning system autonomously and asynchronously brings up data center racks. In an embodiment, the provisioning system determines presence of a first and second device connected to a network. The provisioning system generates a first and second thread for validation of the first and second devices, respectively. Responsive to determining by the first thread that the first device is not validated, the provisioning system notifies a detection system that the validation of the first device has not passed. Responsive to determining by the second thread that the second device is validated, the provisioning system provisions the second device for integration with one or more provisioned devices on the network.

VIRTUALIZED RACK MANAGEMENT MODULES
20170295053 · 2017-10-12 ·

Systems, methods, and computer-readable media for managing nodes through virtual rack management modules. A system can have a first rack that includes a first top-of-rack (ToR) switch and a first group of nodes. The first ToR switch can be connected to the first group of nodes. The system can also have a second rack that includes a second ToR switch and a second group of nodes. The second ToR switch can be connected to the second group of nodes, and the second ToR switch can be connected to the first ToR switch. Furthermore, the system can include a rack management node that executes a hypervisor. The hypervisor can run a first virtual rack management module (vRMM) and a second vRMM. The first vRMM and the second vRMM can manage the first group of nodes and the second group of nodes, respectively.