Patent classifications
H04L12/2881
TECHNOLOGIES FOR PROVIDING ACCELERATED FUNCTIONS AS A SERVICE IN A DISAGGREGATED ARCHITECTURE
Technologies for providing accelerated functions as a service in a disaggregated architecture include a compute device that is to receive a request for an accelerated task. The task is associated with a kernel usable by an accelerator sled communicatively coupled to the compute device to execute the task. The compute device is further to determine, in response to the request and with a database indicative of kernels and associated accelerator sleds, an accelerator sled that includes an accelerator device configured with the kernel associated with the request. Additionally, the compute device is to assign the task to the determined accelerator sled for execution. Other embodiments are also described and claimed.
Technologies for coordinating disaggregated accelerator device resources
A compute device to manage workflow to disaggregated computing resources is provided. The compute device comprises a compute engine receive a workload processing request, the workload processing request defined by at least one request parameter, determine at least one accelerator device capable of processing a workload in accordance with the at least one request parameter, transmit a workload to the at least one accelerator device, receive a work product produced by the at least one accelerator device from the workload, and provide the work product to an application.
Analog and digital communication system for interfacing plain old telephone service devices with a network
An analog and digital communication system establishes a connection between a plain old telephone service (POTS) device and a destination device. The system includes a customer appliance that includes multiple ports, an internal power source device, a subscriber identification module (SIM) card, an electronic control unit, an analog to digital convertor, and a processor device. At least one of the ports are operably connected to the POTS device. A controller communicates remotely with the customer appliance via the network to provide routing of the telephone call between the POTS device and the destination device. The controller is configured to at least one of transmit the telephone call or receive the telephone call via a closest public switched telephone network (PSTN) handoff or via an internal customer appliance network establishing a connection between the customer appliance and a second customer appliance.
Technologies for offloading acceleration task scheduling operations to accelerator sleds
Technologies for offloading acceleration task scheduling operations to accelerator sleds include a compute device to receive a request from a compute sled to accelerate the execution of a job, which includes a set of tasks. The compute device is also to analyze the request to generate metadata indicative of the tasks within the job, a type of acceleration associated with each task, and a data dependency between the tasks. Additionally the compute device is to send an availability request, including the metadata, to one or more micro-orchestrators of one or more accelerator sleds communicatively coupled to the compute device. The compute device is further to receive availability data from the one or more micro-orchestrators, indicative of which of the tasks the micro-orchestrator has accepted for acceleration on the associated accelerator sled. Additionally, the compute device is to assign the tasks to the one or more micro-orchestrators as a function of the availability data.
METHOD FOR DETECTING AND IDENTIFYING ITEMS OF EQUIPMENT COMMUNICATING ACCORDING TO A MODBUS PROTOCOL AND COMMUNICATION CONTROLLER FOR THE IMPLEMENTATION OF SUCH A METHOD
A method for detecting and identifying items of equipment communicating according to a Modbus communication protocol, connected by means of a first communication network of Ethernet type and by means of at least one second communication network, the first and the second communication network being linked by a communication gateway. A communication controller implementing such a method also is disclosed.
Asynchronous object manager in a network routing environment
Systems, methods, and devices for offloading network data to a datastore. A system includes routing chip hardware and an asynchronous object manager in communication with the routing chip hardware. The asynchronous object manager is configurable to execute instructions stored in non-transitory computer readable storage media. The instructions include asynchronously receiving a plurality of objects from one or more producers. The instructions include identifying one or more dependencies between two or more of the plurality of objects. The instructions include reordering the plurality of objects according to the one or more dependencies. The instructions include determining whether the one or more dependencies is resolve. The instructions include, in response to determining the one or more dependencies is resolved, calling back an application and providing one or more of the plurality of objects to the application.
TECHNOLOGIES FOR LIFECYCLE MANAGEMENT WITH REMOTE FIRMWARE
Technologies for lifecycle management include multiple computing devices in communication with a lifecycle management server. On boot, a computing device loads a lightweight firmware boot environment. The lightweight firmware boot environment connects to the lifecycle management server and downloads one or more firmware images for controllers of the computing device. The controllers may include baseboard management controllers, network interface controllers, solid-state drive controllers, or other controllers. The lifecycle management server may select firmware images and/or versions of firmware images based on the controllers or the computing device. The computing device installs each firmware image to a controller memory device coupled to a controller, and in use, each controller accesses the firmware image in the controller memory device. The controller memory device may be a DRAM device or a high-performance byte-addressable non-volatile memory. Other embodiments are described and claimed.
TECHNOLOGIES FOR PROVIDING SHARED MEMORY FOR ACCELERATOR SLEDS
Technologies for providing shared memory for accelerator sleds includes an accelerator sled to receive, with a memory controller, a memory access request from an accelerator device to access a region of memory. The request is to identify the region of memory with a logical address. Additionally, the accelerator sled is to determine from a map of logical addresses and associated physical address, the physical address associated with the region of memory. In addition, the accelerator sled is to route the memory access request to a memory device associated with the determined physical address.
Single node and multiple node datastore architecture in a network routing environment
Systems, methods, and devices for offloading network data to a datastore. A system includes a plurality of instances of a datastore node in a single networking device, each of the plurality of instances of the datastore node comprising: a datastore; a publisher independent of the datastore for publishing data to the datastore; a subscriber independent of the datastore for receiving information from the datastore; a replicator agent configured to connect to the datastore as a publisher or a subscriber; and a persistent storage agent configured to connect to the datastore as a publisher or a subscriber.
Technologies for dynamically managing the reliability of disaggregated resources in a managed node
Technologies for dynamically managing the reliability of disaggregated resources in a managed node include a resource manager server. The resource manager server includes communication circuit to receive resource data from a set of disaggregated resources that indicates reliability of each disaggregated resource of the set of disaggregated resources and a node request to compose a managed node. The resource manager server further includes a compute engine to determine node parameters from the node request indicative of a target reliability of one or more disaggregated resources of the set of disaggregated resources to be included in the managed node, compose a managed node from the set of disaggregated resources that satisfies the node parameters by configuring the compute sled to utilize the disaggregated resources of the managed node for the execution of a workload, and monitor the disaggregated resources of the managed node for a failure.