Patent classifications
G06F9/5094
Neural network operational method and apparatus, and related device
The present disclosure describes methods, devices, and storage mediums for adjusting computing resource. The method includes obtaining an expected pooling time of a target pooling layer and a to-be-processed data volume of the target pooling layer; obtaining a current clock frequency corresponding to at least one computing resource unit used for pooling; determining a target clock frequency according to the expected pooling time of the target pooling layer and the to-be-processed data volume of the target pooling layer; and in response to that the convolution layer associated with the target pooling layer completes convolution and the current clock frequency is different from the target clock frequency, switching the current clock frequency of the at least one computing resource unit to the target clock frequency, and performing pooling in the target pooling layer based on the at least one computing resource unit having the target clock frequency.
Autonomous release management in distributed computing systems
Implementations described herein relate to methods, systems, and computer-readable media to provide an alert based on a release of a software application implemented in a distributed computing system. In some implementations, the method includes receiving, at a processor, an indication of the release of the software application, obtaining a first set of metric values for each metric of a list of metrics for a first time period preceding a time of release of the release, obtaining a second set of metric values for each metric of the list of metrics for a second time period following the time of release, comparing the first set of metric values to the second set of metric values to determine a deviation score, generating an alert based on the deviation score, and transmitting the alert via one of a user interface and a communication channel.
Processing data stream modification to reduce power effects during parallel processing
Certain aspects of the present disclosure provide a method for performing parallel data processing, including: receiving data for parallel processing from a data processing requestor; generating a plurality of data sub-blocks; determining a plurality of data portions in each data sub-block of the plurality of data sub-blocks; changing an order of the plurality of data portions in at least one data sub-block of the plurality of data sub-blocks; providing the plurality of data sub-blocks, including the at least one data sub-block comprising the changed order of the plurality of data portions, to a plurality of processing units for parallel processing; and receiving processed data associated with the plurality of data sub-blocks from the plurality of processing units.
Hybrid cloud orchestration system
A system, method, and computer-readable medium are disclosed for performing a data center monitoring and management operation. The data center monitoring and management operation includes: identifying a plurality of asset resources; selecting a workload for allocation of asset resources; determining which asset resources of the plurality of asset resources may be needed for allocation, determination of which asset resources of the plurality of asset resources may be needed for allocation taking into account on-premises asset resources and cloud-based asset resources the inventory of the available asset resources; and, performing a data center hybrid cloud asset allocation operation, the data center asset allocation operation allocating resources the workload based upon the determining.
Allocation method, system and device for power consumption of complete machine box, and readable storage medium
Provided are an allocation method, system and device for power consumption of a complete machine box, and a readable storage medium. The method, applied to a CMC, includes: determining reserved total power consumption, based on rated power consumption of an integral chassis and preset power consumption of each node; allocating the preset power consumption to each node correspondingly; detecting, at a detection frequency, actual power consumption of each node; and re-allocating, for each node, the reserved total power consumption and the preset power consumption of the node, based on a rate of change in the actual power consumption of the node and/or a power consumption utilization ratio that is a ratio of the actual power consumption of the node to the preset power consumption of the node, thereby maximizing the effective utilization ratio of the integral chassis, reducing the unused power consumption, and reducing operating cost.
Pattern-recognition enabled autonomous configuration optimization for data centers
A model-based approach to determining an optimal configuration for a data center may use an environmental chamber to characterize the performance of various data center configurations at different combinations of temperature and altitude. Telemetry data may be recorded from different configurations as they execute a stress workload at each temperature/altitude combination, and the telemetry data may be used to train a corresponding library of models. When a new data center is being configured, the temperature/altitude of the new data center may be used to select a pre-trained model from a similar temperature/altitude. Performance of the current configuration can be compared to the performance of the model, and if the model performs better, a new configuration based on the model may be used as an optimal configuration for the data center.
Method of task transition between heterogenous processors
A method, system, and apparatus determines that one or more tasks should be relocated from a first processor to a second processor by comparing performance metrics to associated thresholds or by using other indications. To relocate the one or more tasks from the first processor to the second processor, the first processor is stalled and state information from the first processor is copied to the second processor. The second processor uses the state information and then services incoming tasks instead of the first processor.
Task offloading and routing in mobile edge cloud networks
A method implemented by a network element (NE) in a mobile edge cloud (MEC) network, comprising receiving, by the NE, an offloading request message from a client, the offloading request message comprising task-related data describing a task associated with an application executable at the client, determining, by the NE, whether to offload the task to an edge cloud server of a plurality of edge cloud servers distributed within the MEC network based on the task-related data and server data associated with each of the plurality of edge cloud servers, transmitting, by the NE, a response message to the client based on whether the task is offloaded to the edge cloud server.
Serialization floors and deadline driven control for performance optimization of asymmetric multiprocessor systems
Closed loop performance controllers of asymmetric multiprocessor systems may be configured and operated to improve performance and power efficiency of such systems by adjusting control effort parameters that determine the dynamic voltage and frequency state of the processors and coprocessors of the system in response to the workload. One example of such an arrangement includes applying hysteresis to the control effort parameter and/or seeding the control effort parameter so that the processor or coprocessor receives a returning workload in a higher performance state. Another example of such an arrangement includes deadline driven control, in which the control effort parameter for one or more processing agents may be increased in response to deadlines not being met for a workload and/or decreased in response to deadlines being met too far in advance. The performance increase/decrease may be determined by comparison of various performance metrics for each of the processing agents.
Managing workloads of a deep neural network processor
A computing system includes processor cores for executing applications that utilize functionality provided by a deep neural network (“DNN”) processor. One of the cores operates as a resource and power management (“RPM”) processor core. When the RPM processor receives a request to execute a DNN workload, it divides the DNN workload into workload fragments. The RPM processor then determines whether a workload fragment is to be statically allocated or dynamically allocated to a DNN processor. Once the RPM processor has selected a DNN processor, the RPM enqueues the workload fragment on a queue maintained by the selected DNN processor. The DNN processor dequeues workload fragments from its queue for execution. Once execution of a workload fragment has completed, the DNN processor generates an interrupt indicating that execution of the workload fragment has completed. The RPM processor can then notify the processor core that originally requested execution of the workload fragment.