H03M7/4081

Technologies for optical communication in rack clusters

Technologies for optical communication in a rack cluster in a data center are disclosed. In the illustrative embodiment, a network switch is connected to each of 1,024 sleds by an optical cable that enables communication at a rate of 200 gigabits per second. The optical cable has low loss, allowing for long cable lengths, which in turn allows for connecting to a large number of sleds. The optical cable also has a very high intrinsic bandwidth limit, allowing for the bandwidth to be upgraded without upgrading the optical infrastructure.

Technologies For Switching Network Traffic In A Data Center
20190342642 · 2019-11-07 ·

Technologies for switching network traffic include a network switch. The network switch includes one or more processors and communication circuitry coupled to the one or more processors. The communication circuitry is capable of switching network traffic of multiple link layer protocols. Additionally, the network switch includes one or more memory devices storing instructions that, when executed, cause the network switch to receive, with the communication circuitry through an optical connection, network traffic to be forwarded, and determine a link layer protocol of the received network traffic. The instructions additionally cause the network switch to forward the network traffic as a function of the determined link layer protocol. Other embodiments are also described and claimed.

Technologies For Switching Network Traffic In A Data Center
20190342643 · 2019-11-07 ·

Technologies for switching network traffic include a network switch. The network switch includes one or more processors and communication circuitry coupled to the one or more processors. The communication circuity is capable of switching network traffic of multiple link layer protocols. Additionally, the network switch includes one or more memory devices storing instructions that, when executed, cause the network switch to receive, with the communication circuitry through an optical connection, network traffic to be forwarded, and determine a link layer protocol of the received network traffic. The instructions additionally cause the network switch to forward the network traffic as a function of the determined link layer protocol. Other embodiments are also described and claimed.

Technologies for assigning workloads based on resource utilization phases

Technologies for assigning workloads based on resource utilization phases include an orchestrator server to assign a set of workloads to the managed nodes. The orchestrator server is also to receive telemetry data from the managed nodes and identify, as a function of the telemetry data, historical resource utilization phases of the workloads. Further, the orchestrator server is to determine, as a function of the historical resource utilization phases and as the workloads are performed, predicted resource utilization phases for the workloads, and apply, as a function of the predicted resources utilization phases, adjustments to the assignments of the workloads among the managed nodes as the workloads are performed.

Technologies for dynamic allocation of tiers of disaggregated memory resources

Technologies for dynamically allocating tiers of disaggregated memory resources include a compute device. The compute device is to obtain target performance data, determine, as a function of target performance data, memory tier allocation data indicative of an allocation of disaggregated memory sleds to tiers of performance, in which one memory sled of one tier is to act as a cache for another memory sled of a subsequent tier, send the memory tier allocation data and the target performance data to the corresponding memory sleds through a network, receive performance notification data from one of the memory sleds in the tiers, and determine, in response to receipt of the performance notification data, an adjustment to the memory tier allocation data.

DATA PROCESSING METHOD AND APPARATUS
20190312589 · 2019-10-10 · ·

A method of compression is disclosed in which an input sequence of bits is divided into a plurality of portions. Each portion is sub-divided into a plurality of sub-divisions. Frequency analysis is performed to determine the number of occurrences of each sub-division permutation and new values are assigned, based on the frequency analysis, to each of the sub-division permutations. For each portion a label representing the permutation of bits in that portion is assigned. The label comprises a representation of a combined value resulting from combining the new values associated with the sub-division permutations of that portion. A processed sequence of bits is generated by replacing, within the input sequence of bits, bit portions with the respective label representing the permutation of bits in that portion.

Technologies for allocating ephemeral data storage among managed nodes
10411729 · 2019-09-10 · ·

Technologies for allocating ephemeral data storage among managed nodes include an orchestrator server to receive ephemeral data storage availability information from the managed nodes, receive a request from a first managed node of the managed nodes to allocate an amount of ephemeral data storage as the first managed node executes one or more workloads, determine, as a function of the ephemeral data storage availability information, an availability of the requested amount of ephemeral data storage, and allocate, in response to a determination that the requested amount of ephemeral data storage is available from one or more other managed nodes, the requested amount of ephemeral data storage to the first managed node as the first managed node executes the one or more workloads. Other embodiments are also described and claimed.

Techniques to process packets in a dual-mode switching environment

Various embodiments are generally directed to an apparatus, method and other techniques to receive a packet via an optical fabric, the packet comprising a switch mode indicator, determine a switch mode for the packet based on the switch mode indicator, and process the packet in accordance with a first protocol or a second protocol based on the switch mode.

Memory sharing for physical accelerator resources in a data center
10390114 · 2019-08-20 · ·

Examples may include sleds for a rack in a data center including physical accelerator resources and memory for the accelerator resources. The memory can be shared between the accelerator resources. One or more memory controllers can be provided to couple the accelerator resources to the memory to provide memory access to all the accelerator resources. Each accelerator resource can include a memory controller to access a portion of the memory while the accelerator resources can be coupled via an out-of-band channel to provide memory access to the other portions of the memory.

Memory device and memory system

A memory device includes: a plurality of memory cells; soft read logic configured to generate soft data by reading data from the plurality of memory cells in response to a soft read command from a controller, the soft data including at least a major symbol and at least a minor symbol; a compressor configured to generate compressed data by: encoding, into a code alphabet having a second length, a major source alphabet including repetitions of the major symbol by a first length among a plurality of source alphabets included in the soft data, and encoding, into a code alphabet having a longer length than the second length, a minor source alphabet including repetitions of the major symbol by a shorter length than the first length and ending with one minor symbol; and an interface configured to provide the compressed data to the controller.