Patent classifications
H05K7/1442
Technologies For Switching Network Traffic In A Data Center
Technologies for switching network traffic include a network switch. The network switch includes one or more processors and communication circuitry coupled to the one or more processors. The communication circuitry is capable of switching network traffic of multiple link layer protocols. Additionally, the network switch includes one or more memory devices storing instructions that, when executed, cause the network switch to receive, with the communication circuitry through an optical connection, network traffic to be forwarded, and determine a link layer protocol of the received network traffic. The instructions additionally cause the network switch to forward the network traffic as a function of the determined link layer protocol. Other embodiments are also described and claimed.
Technologies For Switching Network Traffic In A Data Center
Technologies for switching network traffic include a network switch. The network switch includes one or more processors and communication circuitry coupled to the one or more processors. The communication circuity is capable of switching network traffic of multiple link layer protocols. Additionally, the network switch includes one or more memory devices storing instructions that, when executed, cause the network switch to receive, with the communication circuitry through an optical connection, network traffic to be forwarded, and determine a link layer protocol of the received network traffic. The instructions additionally cause the network switch to forward the network traffic as a function of the determined link layer protocol. Other embodiments are also described and claimed.
Technologies for assigning workloads based on resource utilization phases
Technologies for assigning workloads based on resource utilization phases include an orchestrator server to assign a set of workloads to the managed nodes. The orchestrator server is also to receive telemetry data from the managed nodes and identify, as a function of the telemetry data, historical resource utilization phases of the workloads. Further, the orchestrator server is to determine, as a function of the historical resource utilization phases and as the workloads are performed, predicted resource utilization phases for the workloads, and apply, as a function of the predicted resources utilization phases, adjustments to the assignments of the workloads among the managed nodes as the workloads are performed.
Technologies for dynamic allocation of tiers of disaggregated memory resources
Technologies for dynamically allocating tiers of disaggregated memory resources include a compute device. The compute device is to obtain target performance data, determine, as a function of target performance data, memory tier allocation data indicative of an allocation of disaggregated memory sleds to tiers of performance, in which one memory sled of one tier is to act as a cache for another memory sled of a subsequent tier, send the memory tier allocation data and the target performance data to the corresponding memory sleds through a network, receive performance notification data from one of the memory sleds in the tiers, and determine, in response to receipt of the performance notification data, an adjustment to the memory tier allocation data.
Technologies for allocating ephemeral data storage among managed nodes
Technologies for allocating ephemeral data storage among managed nodes include an orchestrator server to receive ephemeral data storage availability information from the managed nodes, receive a request from a first managed node of the managed nodes to allocate an amount of ephemeral data storage as the first managed node executes one or more workloads, determine, as a function of the ephemeral data storage availability information, an availability of the requested amount of ephemeral data storage, and allocate, in response to a determination that the requested amount of ephemeral data storage is available from one or more other managed nodes, the requested amount of ephemeral data storage to the first managed node as the first managed node executes the one or more workloads. Other embodiments are also described and claimed.
Techniques to process packets in a dual-mode switching environment
Various embodiments are generally directed to an apparatus, method and other techniques to receive a packet via an optical fabric, the packet comprising a switch mode indicator, determine a switch mode for the packet based on the switch mode indicator, and process the packet in accordance with a first protocol or a second protocol based on the switch mode.
Memory sharing for physical accelerator resources in a data center
Examples may include sleds for a rack in a data center including physical accelerator resources and memory for the accelerator resources. The memory can be shared between the accelerator resources. One or more memory controllers can be provided to couple the accelerator resources to the memory to provide memory access to all the accelerator resources. Each accelerator resource can include a memory controller to access a portion of the memory while the accelerator resources can be coupled via an out-of-band channel to provide memory access to the other portions of the memory.
Configurable computing resource physical location determination
Examples may include techniques to determine locations of a physical resource in a data center. A data center can include a number of racks having sled spaced. The sled spaces accommodate sleds having one or more physical resources disposed on each sled. The racks and sleds can include a beacon and beacon sensor, respectively, operable to determine a location of the sleds within the data center. Beacons and beacon sensors can exchange signals, a pod controller can receive an information element including indications of the exchanged signals and determine a location of the physical resource within the data center.
APPARATUS FOR MOUNTING PROCESSORS FOR CLUSTER COMPUTING
A bracket for mounting a processor and a support structure for receiving bracket-supported processors for cluster computing are provided. In some embodiments, a bracket may be configured to receive a processor and fasten the processor to the bracket. The bracket may be configured to mount the processor to a support structure. The support structure may be configured to receive an array of brackets. The support structure may be configured to be stacked in combination with additional support structures.
Technologies for cooling rack mounted sleds
Technologies for rack cooling includes monitoring a temperature of a sled mounted in a rack and controlling a cooling system of the rack based on the temperature of the sled. The cooling system includes a cooling fan array, which may be controlled to cool the sled. Additionally, if needed, one or more adjacent cooling fan arrays that are located adjacent to the controlled cooling fan array may be adjusted to provide additional cooling to the sled.