G06F2209/5022

CLOUD-BASED SYSTEMS FOR OPTIMIZED MULTI-DOMAIN PROCESSING OF INPUT PROBLEMS USING MACHINE LEARNING SOLVER TYPE SELECTION

Various embodiments of the present disclosure provide methods, apparatuses, systems, computing devices, computing entities, and/or the like for determining optimized solutions to input problems in a containerized, cloud-based (e.g., serverless) manner. In one embodiment, an example method is provided. The method comprises: receiving a problem type of an input problem originating from a client computing entity; mapping the problem type to one or more selected solver types; generating one or more container instances of one or more compute containers, each compute container corresponding to a selected solver type; generating a problem output using the one or more container instances; and providing the problem output comprising a solution to the input problem to the client computing entity. In various embodiments, optimized solutions for input problems are determined using a cloud-based multi-domain solver system configured to dynamically allocate computing and processing resources between different solution-determining tasks.

Processing rest API requests based on resource usage satisfying predetermined limits

A request manager analyzes API calls from a client to a host application for state and performance information. If current utilization of host application processing or memory footprint resources exceed predetermined levels, then the incoming API call is not forwarded to the application. If current utilization of the host application processing and memory resources do not exceed the predetermined levels, then the request manager quantifies the processing or memory resources required to report the requested information and determines whether projected utilization of the host application processing or memory resources inclusive of the resources required to report the requested information exceed predetermined levels. If the predetermined levels are not exceeded, then the request manager forwards the API call to the application for processing.

Electronic device for securing usable dynamic memory and operating method thereof
11579927 · 2023-02-14 · ·

An electronic device including an application processor and a communication processor. The communication processor including a resource memory, the communication processor configured to monitor an occupancy rate of the resource memory, determine whether the electronic device is in an idle state, forcibly release a network connection, clear the resource memory, and reconnect the network connection.

SYSTEM FOR MONITORING AND OPTIMIZING COMPUTING RESOURCE USAGE OF CLOUD BASED COMPUTING APPLICATION
20230043579 · 2023-02-09 ·

A system of monitoring and optimizing computing resources usage for computing application may include predicting a first performance metric for job load capacity of a computing application for optimal job concurrency and optimal resource utilization. The system may include generating an alerting threshold based on the first performance metric. The system may further include, in response to a difference between the alerting threshold and a job load of the computing application within an interval exceeding a threshold, predicting a second performance metric for job load capacity of the computing application for optimal job concurrency and optimal resource utilization. The system may further include, in response to a difference between the first performance metric and the second performance metric exceeding a difference threshold, updating the alerting threshold with a job load capacity with the optimal resource utilization rate corresponding to the second performance metric.

SYSTEMS AND METHODS FOR UNIVERSAL AUTO-SCALING
20230040512 · 2023-02-09 ·

Systems and methods for universal auto-scaling are disclosed. In one embodiment, a method may include: (1) monitoring, by an auto-scale computer program executed by a computer processor, a utilization level at each of a plurality of data layers in a data pod, wherein each data layer comprises at least one node; (2) comparing, by the auto-scale computer program, each of the utilization levels to a threshold; (3) identifying, by the auto-scale computer program, that one of the thresholds is met or exceeded; and (4) deploying, by the auto-scale computer program, an additional node to the data layer with the met or exceeded utilization level.

Predicting and halting runaway queries

Operations include halting a runaway query in response to determining that a performance metric of the query exceeds a performance threshold. The runaway query halting system receives a query execution plan associated with a query and divides the received execution plan into one or more components. For each component, the system determines a predicted resource usage associated with executing the component. The system further determines a predicted resource usage associated with the query execution plan based on the predicted resource usage associated with each component. The system executes the query associated with the received query execution plan and compares the predicted resource usage associated with the query to a resource usage threshold. In response to determining that the predicted resource usage of the query execution plan exceeds the resource usage threshold, the system halts execution of the query associated with the query execution plan.

INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND COMPUTER-READABLE RECORDING MEDIUM STORING INFORMATION PROCESSING PROGRAM
20230010895 · 2023-01-12 · ·

An information processing apparatus includes: a memory; and a processor coupled to the memory and configured to: divide a job in units of computing nodes for a plurality of computing nodes; determine execution of scale-out or scale-in on the basis of a load in a case where each of the computing nodes is caused to execute a job obtained by the division; execute, in a case where determining execution of the scale-out, the scale-out according to the division of the job in units of computing nodes; and execute, in a case where determining execution of the scale-in, the scale-in according to the division of the job in units of computing nodes.

DIFFERENTIATED WORKLOAD TELEMETRY
20230009332 · 2023-01-12 ·

In an approach for generating differentiated workload telemetry data, a processor corresponds one or more services with a workload related telemetry generating an event emitter. A processor performs a correlation analysis of corresponding relationship and connection among connected resources and current traffic into and out of the one or more services. A processor labels domain context for each telemetry event. A processor communicates each telemetry event to a global event handler. A processor performs a cross-correlation in real-time of telemetry data with the global event handler. A processor updates a real-time differentiated workload report.

Distribution of quantities of an increased workload portion into buckets representing operations

In some examples, a computing system receives an indication of an increased workload portion to be added to a workload of a storage system, the workload comprising buckets of operations of different characteristics. The computing system computes, based on quantities of operations of the different characteristics in the workload, factor values that indicate distribution of operations of the increased workload portion to the buckets of operations of the different characteristics, and distributes, according to the factor values, the operations of the increased workload portion into the buckets of operations of the different characteristics.

Systems and methods for autoscaling instance groups of computing platforms

Systems and methods scale an instance group of a computing platform by determining whether to scale up or down the instance group by using historical data from prior jobs wherein the historical data includes one or more of: a data set size used in a prior related job and a code version for a prior related job. The systems and methods also scale the instance group up or down based on the determination. In some examples, systems and methods scale an instance group of a computing platform by determining a job dependency tree for a plurality of related jobs, determining runtime data for each of the jobs in the dependency tree and scaling up or down the instance group based on the determined runtime data.