Patent classifications
G06F2209/5013
Acquisition and maintenance of compute capacity
A system for providing low-latency computational capacity from a virtual compute fleet is provided. The system may be configured to maintain a plurality of virtual machine instances on one or more physical computing devices, wherein the plurality of virtual machine instances comprises a first pool comprising a first sub-pool of virtual machine instances and a second sub-pool of virtual machine instances, and a second pool comprising virtual machine instances used for executing one or more program codes thereon. The first sub-pool and/or the second sub-pool may be associated with one or more users of the system. The system may be further configured to process code execution requests and execute program codes on the virtual machine instances of the first or second sub-pool.
Systems and methods for distributed resource management
Methods, nontransitory computer readable media, and systems are disclosed for servicing a job queue. Each job has node resource requirements. Composite job memory and processor requirements is determined from these requirements. Nodes that satisfy these requirements are identified by obtaining, for each class of a plurality of node classes: an availability score, a number of processers, and a memory capability. A request for nodes of a class is made when a demand score for the class satisfies the class availability score. An acknowledgement and updated availability score is received upon request acceptance. A declination is received upon request rejection. The submitting and receiving is performing multiple times, if needed, until each class has been considered for a request or sufficient acknowledgements are received to satisfy the composite requirements of the jobs. Each node in the cluster draws jobs from the queue subject to the collective requirements of the drawn jobs.
Resource allocation using distributed segment processing credits
Systems and methods for allocating resources are disclosed. Resources as processing time, writes or reads are allocated. Credits are issued to the clients in a manner that ensure the system is operating in a safe allocation state. The credits can be used not only to allocate resources but also to throttle clients where necessary. Credits can be granted fully, partially, and in a number greater than requested. Zero or negative credits can also be issued to throttle clients. Segment credits are associated with identifying unique fingerprints or segments and may be allocated by determining how many credits a CPU/cores can support. This maximum number may be divided amongst clients connected with the server.
System and Method for Confluence Resource Allocation
A computer-implemented method including: receiving a request to schedule a confluence involving a plurality of entities during a future time interval, wherein the confluence is associated with one or more confluence locations; obtaining a plurality of expected location indicators, wherein the plurality of expected location indicators indicate the expected location of the plurality of entities during the future time interval; and allocating resources for the confluence during the future time interval based on the plurality of expected location indicators.
METHOD, CONTROLLER, AND SYSTEM FOR SERVICE FLOW CONTROL IN OBJECT-BASED STORAGE SYSTEM
A method, a controller, and a system for service flow control in an object-based storage system are disclosed. The method is: receiving, by a controller, a first object IO request; acquiring a processing quantity threshold and a to-be-processed quantity; if the to-be-processed quantity is less than the processing quantity threshold, sending the first object IO request to a storage device client, and updating the to-be-processed quantity; receiving a first response message replied by the storage device client for the first object IO request, where the first response message carries a processing result of the first object IO request; and adjusting the processing quantity threshold according to a received processing result of an object IO request when a preset condition is met. The storage device is not overloaded with object IO requests and can use all resources to effectively, thereby improving performance and a success rate of the object-based storage system.
High-load business process scalability
The present disclosure involves systems, software, and computer implemented methods for providing high-load business process scalability in cloud-based infrastructures. One process includes operations for receiving a message at a first computer node executing a first business process instance. A second business process instance associated with the message is identified. The message is sent to a messaging queue for retrieval by the second business process instance if the second business process instance is not located at the first computer node.
Scheduling execution requests to allow partial results
The subject disclosure is directed towards scheduling requests using quality values that are defined for partial responses to the requests. For each request in a queue, an associated processing time is determined using a system load and/or the quality values. The associated processing time is less than or equal to a service demand, which represents an amount of time to produce a complete response.
METHOD FOR EXECUTING AND MANAGING DISTRIBUTED PROCESSING, AND CONTROL APPARATUS
A non-transitory computer-readable recording medium stores a control program that causes a computer to execute a process, the process includes collecting a processing result of a subjob distributed to a plurality of nodes, each of the plurality of nodes processing a to-be-processed job distributed among the nodes estimating an overall processing result, based on the collected processing results of the subjobs, the overall processing result being a result of overall processing corresponding to the subjobs; and determining whether or not to continue processing remaining subjobs of the subjobs corresponding to the overall processing depending on the estimated overall processing result.
Tenant selection in quota enforcing request admission mechanisms for shared applications
Methods, systems, and computer-readable storage media for selecting requests from a plurality of tenant queues in a multi-tenant system including actions of determining a plurality of weights, each weight being associated with a respective tenant in a set of tenants, each weight being dynamically adjusted based on a quota associated with the respective tenant for each iteration of a plurality of iterations within a period, selecting a tenant from the set of tenants based on the plurality of weights to provide a selected tenant, and transmitting a request from a request queue associated with the selected tenant for processing by a shared resource.
MANAGED INTEGRATION OF CONSTITUENT SERVICES OF MULTI-SERVICE APPLICATIONS
At a managed control plane service, constituent services and operational requirements of an application are identified. In response to an end-user request directed to the application, contents of an inter-service request are generated at a resource selected by the managed control plane service for a first constituent service, and a response to the message is generated at another resource selected for a second constituent service. Tasks to be performed for the operational requirements are initiated by the managed control plane service.