Patent classifications
G06F9/5005
Vehicular arbitration system
A vehicular arbitration system includes: a main manager configured to receive one or more requests from a plurality of first application execution units and to determine a request for operating a predetermined on-vehicle device based on the received one or more requests and a predetermined rule; and a plurality of sub-managers respectively configured to arbitrate the request determined by the main manager and a request input from at least one second application execution unit that is different from the plurality of first application execution units and to control the on-vehicle device based on an arbitration result.
Systems and methods for autoscaling instance groups of computing platforms
Systems and methods scale an instance group of a computing platform by determining whether to scale up or down the instance group by using historical data from prior jobs wherein the historical data includes one or more of: a data set size used in a prior related job and a code version for a prior related job. The systems and methods also scale the instance group up or down based on the determination. In some examples, systems and methods scale an instance group of a computing platform by determining a job dependency tree for a plurality of related jobs, determining runtime data for each of the jobs in the dependency tree and scaling up or down the instance group based on the determined runtime data.
Data discovery solution for data curation
Disclosed are methods and systems for a data discovery solution which harnesses the power of crowdsourcing to improve automated data curation. This is done in two complimentary ways: (a) large scale collective curation through anonymized multi-tenancy, and (b) and through internet scale matching and validation gaming platform using mobile application game. The result is the most extensive library of semantic-technical mappings of the enterprise data, which are immediately at hand to provide a fast, easy and a good understanding of the enterprise data. The data discovery solution forms a gateway for governing and unlocking value from big data.
STORAGE ARRAY RESOURCE ALLOCATION BASED ON FEATURE SENSITIVITIES
Aspects of the present disclosure relate to tuning resource allocations based on a storage array feature's impact on the array's global performance. In embodiments, one or more input/output (IO) features used by a storage array to process one or more IO workloads are determined. Additionally, each IO feature's impact for processing the IO workload within a threshold performance requirement can be determined. Further, at least one IO feature's resource allocation can be tuned based on its IO workload processing impact.
DATA STREAMING PIPELINE FOR COMPUTE MAPPING SYSTEMS AND APPLICATIONS
Approaches presented herein provide systems and methods for a data platform to identify, evaluate, and map data for processing via one or more compute instances. The data platform may receive an instruction and retrieve data from a variety of remote data locations. The data may be processed, such as to label or file the data, and then streamed to a compute instance for further evaluation. The data platform may be used to provide a centralized system for managing system data that combine both legacy and modern storage solutions into a single integrated platform.
Resource monitor for monitoring long-standing computing resources
Disclosed herein are system, apparatus, article of manufacture, method, and/or computer program product embodiments for monitoring long-standing computing resources. An apparatus may operate by receiving a cloud monitoring notification, where the cloud monitoring notification may indicate an occurrence of a monitored condition. The apparatus may then operate by scanning a cluster computing system for resource having a client assigned resource identifier and a computing resource attribute based on a resource identifier scan parameter and a resource attribute scan parameter. The apparatus may further operate by generating a resource notification request based on the scanning of the cluster computing system and transmitting the resource notification request to a communications system to notify a user that the resource has a computing resource attribute that match the resource attribute scan parameter.
Systems and methods for reducing forced application termination
The disclosed computer-implemented method may include detecting an application running in a background state on a client device. The method may also include collecting state data about a current state of the client device. Additionally, the method may include determining, by applying a machine learning model to the collected state data, that a likelihood of forcible termination of the application within a predetermined timeframe exceeds a threshold. Furthermore, the method may include reducing a computing resource footprint of the application on the client device to reduce the likelihood of forcible termination of the application. Various other methods, systems, and computer-readable media are also disclosed.
Trade platform with reinforcement learning network and matching engine
A system for reinforcement learning in a dynamic resource environment includes at least one memory and at least one processor configured to provide an electronic resource environment comprising: a matching engine and the resource generating agent configured for: obtaining from a historical data processing task database a plurality of historical data processing tasks, each historical data processing task including respective task resource requirement data; for a historical data processing task of the plurality of historical data processing tasks, generating layers of data processing tasks wherein a first layer data processing task has an incremental variant in its resource requirement data relative to resource requirement data for a second layer data processing task; and providing the layers of data processing tasks for matching by the machine engine.
OPERATING SYSTEM-LEVEL ASSISTIVE FEATURES FOR CONTEXTUAL PRIVACY
Systems and methods are described that include operations such as detecting a plurality of computing devices configured as a distributed ambient computing system, receiving a request to execute a computing task, obtaining, from the distributed ambient computing system, data representing a device context for at least two of the plurality of devices, generating a combined context corresponding to the distributed ambient computing system, the combined context representing a combination of the device context for the at least two devices, generating and providing at least one decision request based on the computing task and the combined context, receiving a response to the at least one decision request, and triggering execution of the computing task based on the response and the combined context.
EDGE FUNCTION BURSTING
One example method includes determining that local resources at an edge site are inadequate to support performance of a function needed by software running on the edge site, invoking a client agent, in response to invoking the client agent, receiving an execution manifest, determining, by the client agent, where to execute the function, wherein the determining comprises identifying a target execution environment for the function and the determining is based in part on information contained in the execution manifest, and transmitting, by the client agent, the execution manifest to a server agent of the target execution environment, and the execution manifest facilitates execution of the function in the target execution environment.