Patent classifications
H04L41/0889
Integrity verification for managing network configurations
In one embodiment, a method includes determining a first configuration information for configuring a user plane device of the first network slice associated with the control plane device, generating a first configuration identifier based on the first configuration information, sending, to the user plane device of the first network slice, the first configuration identifier and the first configuration information, receiving, from the user plane device of the first network slice, a second configuration identifier, wherein the second configuration identifier is generated by the user plane device of the first network slice based on the first configuration information received from the control plane device, and verifying, by comparing the first configuration identifier and the second configuration identifier, that the user plane device of the first network slice is configured using the first configuration information.
PERIPHERAL DEVICE CONNECTIONS RESPONSIVE TO IDENTIFICATION DATA MATCHING
In an example in accordance with the present disclosure, a compute device is described. The compute device includes a host controller to 1) receive identification data of peripheral devices coupled to a computing dock and 2) compare the identification data with values stored in a database. A stored value indicates a combination of peripheral devices previously coupled to the computing dock. The compute device also includes a device connection controller to, responsive to the identification data matching a stored value, retrieve operational data for the peripheral devices from the database. The device connection controller is also to connect the compute device with the peripheral devices based on operational data retrieved from the database.
SERVER APPARATUS, WIRELESS BASE STATION, MONITORING APPARATUS, COMMUNICATION CONTROLLING METHOD, NON-TRANSITORY COMPUTER-READABLE MEDIUM, AND COMMUNICATION SYSTEM
Provided is a server apparatus capable of reducing a workload in a setting process. A server apparatus (10) includes: a registering unit (11) configured to register first address information set in a monitoring apparatus (30) that monitors a wireless base station (20) and VLAN setting information serving as setting information of a Virtual Local Area Network (VLAN) including the monitoring apparatus (30); a receiving unit (12) configured to receive, from the wireless base station (20), an assignment request for second address information; and a transmitting unit (13) configured to transmit the first address information, the second address information, and the VLAN setting information to the wireless base station (20).
Method and system for virtual server dormancy
A method for providing a dormant state for content management servers is provided. Client devices are allowed to conduct transactions with servers when the servers are active. However, in a dormant state, the servers are not allowed to accept new transactions. Thus, by utilizing the dormant state, software upgrades can be made to one server at a time. Alternatively, all servers can be taken down for major upgrades, with the servers still operated in a read-only mode based on a file image from a point in time just prior to the shutdown. When the upgrade is completed, the servers can be returned to the active state.
ACCELERATOR OR ACCELERATED FUNCTIONS AS A SERVICE USING NETWORKED PROCESSING UNITS
Various approaches for deploying and controlling distributed accelerated compute operations with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A system for orchestrating acceleration functions in a network compute mesh is configured to access a flowgraph, the flowgraph including data producer-consumer relationships between a plurality of tasks in a workload; identify available artifacts and resources to execute the artifacts to complete each of the plurality of tasks, wherein an artifact is an instance of a function to perform a task of the plurality of tasks; determine a configuration assigning artifacts and resources to each of the plurality of tasks in the flowgraph; and schedule, based on the configuration, the plurality of tasks to execute using the assigned artifacts and resources.
INFRASTRUCTURE-DELEGATED ORCHESTRATION BACKUP USING NETWORKED PROCESSING UNITS
Various approaches for monitoring and responding to orchestration or service failures with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A method performed by a computing device for deploying remedial actions in failure scenarios of an orchestrated edge computing environment may include: identifying an orchestration configuration of a controller entity (responsible for orchestration) and a worker entity (subject to the orchestration to provide at least one service); determining a failure scenario of the orchestration of the worker entity, such as at a networked processing unit implemented at a network interface located between the controller entity and the worker entity; and causing a remedial action to resolve the failure scenario and modify the orchestration configuration, such as replacing functionality of the controller entity or the worker entity with functionality at a replacement entity.
MEMORY INTERLEAVING COORDINATED BY NETWORKED PROCESSING UNITS
Various approaches for configuring interleaving in a memory pool used in an edge computing arrangement, including with the use of infrastructure processing units (IPUs) and similar networked processing units, are disclosed. An example system may discover and map disaggregated memory resources at respective compute locations connected to each another via at least one interconnect. The system may identify workload requirements for use of the compute locations by respective workloads, for workloads provided by client devices to the compute locations. The system may determine an interleaving arrangement for a memory pool that fulfills the workload requirements, to use the interleaving arrangement to distribute data for the respective workloads among the disaggregated memory resources. The system may configure the memory pool for use by the client devices of the network, as the memory pool causes the disaggregated memory resources to host data based on the interleaving arrangement.
MANAGEMENT OF WORKLOAD PROCESSING USING DISTRIBUTED NETWORKED PROCESSING UNITS
Various approaches for deploying and controlling distributed compute operations with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A system that includes a networked processing unit may perform workload processing with operations that: receive, from another networked processing unit, workload information for a workload, for a workload having respective tasks to be processed among distributed computing entities; perform an analysis of network conditions for a predicted execution of the workload, based on the workload information, to analyze network availability among the distributed computing entities; perform an analysis of compute conditions for the predicted execution of the workload, based on the workload information, to analyze processing availability among the distributed computing entities; and identify locations of the distributed computing entities to deploy the workload, based on the analysis of network conditions and the analysis of compute conditions.
OPTIMIZING CONCURRENT EXECUTION USING NETWORKED PROCESSING UNITS
Various approaches for managing distributed compute operations for workload execution of concurrent tasks, including with the use of infrastructure processing units (IPUs) and similar networked processing units, are disclosed. An example method may include: identifying multiple tasks of a computing workload, for a workload that provides processing dependencies among the tasks, and that uses concurrent execution with one or more of the tasks; monitoring an execution time for each of the tasks, relative to an execution time threshold for each of the tasks; identifying the execution time of a particular task as exceeding an execution time threshold for the particular task; determining a remediation based on the particular task and the identified execution time, with the remediation including use of other compute resources in the distributed computing environment for the workload; and applying the remediation to increase speed of execution of the workload.
VIRTUAL POOLS AND RESOURCES USING DISTRIBUTED NETWORKED PROCESSING UNITS
Various approaches for deploying and using virtual pools of compute resources with the use of infrastructure processing units (IPUs) and similar networked processing units are disclosed. A host computing system may be configured to operate a virtual pool of resources, with operations including: identifying, at the host computing system, availability of a resource at the host computing system; transmitting, to a network infrastructure device, a notification that the resource at the host computing system is available for use in a virtual resource pool in the edge computing network; receiving a request for the resource in the virtual resource pool that is provided on behalf of a client computing system, based on the request being coordinated via the network infrastructure device and includes at least one quality of service (QoS) requirement; and servicing the request for the resource, based on the at least one QoS requirement.