Patent classifications
G06F9/3555
AUTOSCALING IN AN ELASTIC CLOUD SERVICE
Techniques described herein can optimize usage of computing resources in a data system. Dynamic throttling can be performed locally on a computing resource in the foreground and autoscaling can be performed in a centralized fashion in the background. Dynamic throttling can lower the load without overshooting while minimizing oscillation and reducing the throttle quickly. Autoscaling may involve scaling in or out the number of computing resources in a cluster as well as scaling up or down the type of computing resources to handle different types of situations.
DATA PROCESSING APPARATUS AND RELATED PRODUCTS
The present disclosure provides a data processing apparatus and related products. The products include a control module including an instruction caching unit, an instruction processing unit, and a storage queue unit. The instruction caching unit is configured to store computation instructions associated with an artificial neural network operation; the instruction processing unit is configured to parse the computation instructions to obtain a plurality of operation instructions; and the storage queue unit is configured to store an instruction queue, where the instruction queue includes a plurality of operation instructions or computation instructions to be executed in the sequence of the queue. By adopting the above-mentioned method, the present disclosure can improve the operation efficiency of related products when performing operations of a neural network model.
System and method for pipelined time-domain computing using time-domain flip-flops and its application in time-series analysis
Systems and/or methods can include a ring based inverter chain that constructs multi-bit flip-flops that store time. The time flip-flops serve as storage units and enable pipeline operations. Single cells used in time series analysis, such as dynamic time warping are rendered by the time-domain circuits. The circuits include time flip-flops, Min, and ABS circuits. A and the matrix can be constructed through the single cells.
Autoscaling and throttling in an elastic cloud service
Techniques described herein can optimize usage of computing resources in a data system. Dynamic throttling can be performed locally on a computing resource in the foreground and autoscaling can be performed in a centralized fashion in the background. Dynamic throttling can lower the load without overshooting while minimizing oscillation and reducing the throttle quickly. Autoscaling may involve scaling in or out the number of computing resources in a cluster as well as scaling up or down the type of computing resources to handle different types of situations.
Orchestrated scaling of cloud architecture components
A computer-implemented method is disclosed. The method can comprise: monitoring utilization of a cloud architecture component that is being used by a component utilizer; determining, via a machine learning model, a pattern of usage of the cloud architecture component based on the monitoring; determining, based on the pattern of usage, a first time period when the cloud architecture component is excessively used by the component utilizer and a second time period when the cloud resource is scantily used by the component utilizer; and orchestrating, based on the first and second time periods, a scaling of the cloud architecture immediately before a subsequent iteration of the pattern of usage by the component utilizer.
GENERATION OF A RECOMMENDATION FOR AUTOMATIC TRANSFORMATION OF TIMES SERIES DATA AT INGESTION
In a computer-implemented method for generating a recommendation for automatic transformation of times series data at ingestion, historical query data of a time series data monitoring system is analyzed, where the historical query data includes a plurality of queries and data associated with execution of the plurality of queries. Based on the analyzing, it is determined whether an execution cost of a query of the plurality of queries can be reduced by performing automatic transformation of at least a portion of times series data accessed responsive to the query at ingestion into the time series data monitoring system. In response to determining that the execution cost of the query can be reduced by performing automatic transformation at ingestion, a recommendation to perform the automatic transformation of the at least a portion of the times series data at ingestion is generated.
SERVICE LOAD INDEPENDENT RESOURCE USAGE DETECTION AND SCALING FOR CONTAINER-BASED SYSTEM
A computer implemented method and related system determine a current load result of a software container executing on a compute node in a container system. In response to determining that the current load result exceeds a predetermined scale-up threshold for the software container, the method adds a first plurality of replicas of the software container to the compute node, where a quantity of the first plurality of replicas is related to the current load result. In response to determining that the current load result is less than a predetermined scale-down threshold for the software container, the method deletes a second plurality of replicas of the software container from the compute node, where a quantity of the second plurality of replicas is related to the current load result.
THERMAL STATE INFERENCE BASED FREQUENCY SCALING
The systems and methods monitor thermal states associated with a device. The systems and methods set thermal thresholds associated with the device. The systems and methods infer the thermal thresholds from information gathered by a client application running on the device. The systems and methods implement a stored policy associated with a violation of one of the thermal thresholds by one of the monitored thermal states.
VECTOR CONVERT HEXADECIMAL FLOATING POINT TO SCALED DECIMAL INSTRUCTION
An instruction to perform converting and scaling operations is provided. Execution of the instruction includes converting an input value in one format to provide a converted result in another format. The converted result is scaled to provide a scaled result. A result obtained from the scaled result is placed in a selected location. Further, an instruction to perform scaling and converting operations is provided. Execution of the instruction includes scaling an input value in one format to provide a scaled result and converting the scaled result from the one format to provide a converted result in another format. A result obtained from the converted result is placed in a selected location.
HIERARCHICAL WORKLOAD ALLOCATION IN A STORAGE SYSTEM
A method for hierarchical workload allocation in a storage system, the method may include determining to reallocate a compute workload of a current compute core of the storage system; wherein the current compute core is responsible for executing a workload allocation unit that comprises one or more first type shards; and reallocating the compute workload by (a) maintaining the responsibility of the current compute core for executing the workload allocation unit, and (b) reallocating at least one first type shard of the one or more first type shards to a new workload allocation unit that is allocated to a new compute core of new compute cores.