Patent classifications
H04L12/2881
TECHNOLOGIES FOR OFFLOADING ACCELERATION TASK SCHEDULING OPERATIONS TO ACCELERATOR SLEDS
Technologies for offloading acceleration task scheduling operations to accelerator sleds include a compute device to receive a request from a compute sled to accelerate the execution of a job, which includes a set of tasks. The compute device is also to analyze the request to generate metadata indicative of the tasks within the job, a type of acceleration associated with each task, and a data dependency between the tasks. Additionally the compute device is to send an availability request, including the metadata, to one or more micro-orchestrators of one or more accelerator sleds communicatively coupled to the compute device. The compute device is further to receive availability data from the one or more micro-orchestrators, indicative of which of the tasks the micro-orchestrator has accepted for acceleration on the associated accelerator sled. Additionally, the compute device is to assign the tasks to the one or more micro-orchestrators as a function of the availability data.
Technologies for providing accelerated functions as a service in a disaggregated architecture
Technologies for providing accelerated functions as a service in a disaggregated architecture include a compute device that is to receive a request for an accelerated task. The task is associated with a kernel usable by an accelerator sled communicatively coupled to the compute device to execute the task. The compute device is further to determine, in response to the request and with a database indicative of kernels and associated accelerator sleds, an accelerator sled that includes an accelerator device configured with the kernel associated with the request. Additionally, the compute device is to assign the task to the determined accelerator sled for execution. Other embodiments are also described and claimed.
Independent datastore in a network routing environment
Systems, methods, and devices for offloading network data to a datastore. A system includes a publisher device in a network computing environment. The system includes a subscriber device in the network computing environment. The system includes a datastore independent of the publisher device and the subscriber device, the datastore comprising one or more processors in a processing platform configurable to execute instructions stored in non-transitory computer readable storage media. The instructions includes receiving data from the publisher device. The instructions include storing the data across one or more of a plurality of shared storage devices. The instructions include providing the data to the subscriber device.
Host routed overlay with deterministic host learning and localized integrated routing and bridging
Systems, methods, and devices for improved routing operations in a network computing environment. A system includes a virtual customer edge router and a host routed overlay comprising a plurality of host virtual machines. The system includes a routed uplink from the virtual customer edge router to one or more of the plurality of leaf nodes. The system is such that the virtual customer edge router is configured to provide localized integrated routing and bridging (IRB) service for the plurality of host virtual machines of the host routed overlay.
Independent Datastore In A Network Routing Environment
Systems, methods, and devices for offloading network data to a datastore. A system includes a publisher device in a network computing environment. The system includes a subscriber device in the network computing environment. The system includes a datastore independent of the publisher device and the subscriber device, the datastore comprising one or more processors in a processing platform configurable to execute instructions stored in non-transitory computer readable storage media. The instructions includes receiving data from the publisher device. The instructions include storing the data across one or more of a plurality of shared storage devices. The instructions include providing the data to the subscriber device.
Routing optimizations in a network computing environment
Systems, methods, and devices for improved routing operations in a network computing environment. A system includes a network topology comprising a spine node and a plurality of leaf nodes. The system is such that at least one of the plurality of leaf nodes is associated with one or more networking prefixes. The spine node stores a prefix table. The prefix table includes a listing of networking prefixes in the network topology. The prefix table includes an indication of at least one equal-cost multipath routing (ECMP) group associated with each of the networking prefixes in the network topology. The prefix table includes an indication of at least one leaf node of the plurality of leaf nodes associated with each of the networking prefixes in the network topology.
TECHNOLOGIES FOR DIVIDING WORK ACROSS ACCELERATOR DEVICES
Technologies for dividing work across one or more accelerator devices include a compute device. The compute device is to determine a configuration of each of multiple accelerator devices of the compute device, receive a job to be accelerated from a requester device remote from the compute device, and divide the job into multiple tasks for a parallelization of the multiple tasks among the one or more accelerator devices, as a function of a job analysis of the job and the configuration of each accelerator device. The compute engine is further to schedule the tasks to the one or more accelerator devices based on the job analysis and execute the tasks on the one or more accelerator devices for the parallelization of the multiple tasks to obtain an output of the job.
Loop conflict avoidance in a spine-and-leaf network topology
Systems, methods, and devices for routing operations in a network computing environment. A system includes a network topology comprising a plurality of spine nodes and a plurality of leaf nodes, wherein a link between a first spine node and a first leaf node is inactive. The first spine node includes one or more processors configurable to execute instructions stored in non-transitory computer readable storage media. The instructions include receiving a packet to be transmitted to the first leaf node. The instructions include identifying an alternative spine node at a same level in the network topology. The instructions include attaching a tunnel label to the packet, wherein the tunnel label indicates the packet should be transmitted to the alternative spine node.
ILLUMINATION CONTROL DEVICE AND RELATED ILLUMINATION CONTROL SYSTEM
A control device for controlling at least one LED module includes: a power supply module arranged to receive power over Ethernet to generate a first supply power; and an illumination controlling module coupled to the power supply module for receiving a communication signal from the Ethernet to generate a serial bus signal to control an illumination of the at least one LED module; wherein the illumination controlling module is powered by the first supply power.
Technologies for dividing work across accelerator devices
Technologies for dividing work across one or more accelerator devices include a compute device. The compute device is to determine a configuration of each of multiple accelerator devices of the compute device, receive a job to be accelerated from a requester device remote from the compute device, and divide the job into multiple tasks for a parallelization of the multiple tasks among the one or more accelerator devices, as a function of a job analysis of the job and the configuration of each accelerator device. The compute engine is further to schedule the tasks to the one or more accelerator devices based on the job analysis and execute the tasks on the one or more accelerator devices for the parallelization of the multiple tasks to obtain an output of the job.