Patent classifications
G06F2209/5013
Development and analysis of quantum computing programs
Techniques regarding the development and/or analysis of one or more quantum computing programs are provided. For example, one or more embodiments described herein can comprise a system, which can comprise a memory that can store computer executable components. The system can also comprise a processor, operably coupled to the memory, and that can execute the computer executable components stored in the memory. The computer executable components can comprise a circuit component, operatively coupled to the processor, that can create a quantum computing program over a period of time. The computer executable components can also comprise a visualization component, operatively coupled to the processor, that can generates a quantum state visualization that depicts a characterization of the quantum computing program over the period of time.
Systems and methods of distributed processing
In some embodiments a distributed computing system is provided that includes a plurality of different feature modules and a matching engine. The different feature modules each provide different processing for handling parent requests and submitting, to the matching engine, commands for child data transaction requests that are associated with the parent request.
Re-initiation of microservices utilizing context information provided via service calls
An apparatus comprises a processing device configured to identify, at a first microservice, a service call that is to be transmitted to a second microservice, and to modify the service call to include context information, the context information characterizing a current state of execution of one or more tasks by one of the first microservice and the second microservice. The processing device is further configured to provide, from the first microservice to the second microservice, the modified service call including the context information. The context information enables re-initiation of said one of the first microservice and the second microservice to continue execution of the one or more tasks from the current state.
Resource management unit for capturing operating system configuration states and swapping memory content
This disclosure describes methods, devices, systems, and procedures in a computing system for capturing a configuration state of an operating system executing on a central processing unit (CPU), and offloading memory management tasks, based on the configuration state, to a resource management unit such as a system-on-a-chip (SoC). The resource management unit identifies a status of a resource requiring memory swapping based on the captured configuration state of the operating system. The resource management unit then swaps the memory to alleviate the CPU from processing the swap thereby improving overall computing system performance.
Systems and methods for distributed resource management
Methods, computer readable media, and systems service a queue, comprising a plurality of jobs, by identifying nodes satisfying a hardware requirement for at least a subset of jobs in the queue. Each job indicates when it was submitted to the queue and one or more node resource requirements. A current availability score for each node class in a plurality of node classes is determined and nodes of a first node class in the plurality of node classes are reserved when a demand score for the class satisfies the current availability score for the first node class by a first threshold amount. Reserved nodes are permitted to draw jobs from the queue in accordance with satisfaction by such nodes of the node resource requirements of the jobs but are terminated, without completing the jobs, when the current availability score for their node class exceeds a second threshold amount.
Methods, devices, and a computer program product for processing an access request and updating a storage system
Techniques for processing an access request and updating a storage system are provided. For instance, a method comprises: receiving an access request for an object associated with a storage system, the storage system including a plurality of physical nodes, each of the plurality of physical nodes including at least one set of virtual units, each set of virtual units including at least one virtual unit; determining, from a plurality of sets of virtual units included in the plurality of physical nodes of the storage system, a target set of virtual units associated with the object; and determining, from the target set of virtual units, a target virtual unit corresponding to the object. With the technical solution of the present disclosure, not only a set of virtual units on a physical node may be easily split and merged, but also huge computing resources that need to be allocated may be saved, so better user experience may be brought about at a lower cost.
Dynamic insertion of variablized secrets in a pipeline integration system
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for dynamically injecting secrets in a pipeline integration system. One of the methods includes receiving, by a pipeline controller, a job identifying a pipeline for executing a plurality of tasks. A preparatory task is invoked for the pipeline that obtains one or more original task definitions for the identified pipeline and generates one or more modified task definitions for the pipeline, wherein each modified task definition references one or more variablized secret names specified as the value of the reserved variable of the one or more configuration files. One or more subsequent execution tasks are invoked using each respective modified task definition. One or more retrieved secrets are injected into an execution environment for each particular execution task.
Methods and apparatus for state objects in cluster computing
Embodiments of a mobile state object for storing and transporting job metadata on a cluster computing system may use a database as an envelope for the metadata. A state object may include a database that stores the job metadata and wrapper methods. A small database engine may be employed. Since the entire database exists within a single file, complex, extensible applications may be created on the same base state object, and the state object can be sent across the network with the state intact, along with history of the object. An SQLite technology database engine, or alternatively other single file relational database engine technologies, may be used as the database engine. To support the database engine, compute nodes on the cluster may be configured with a runtime library for the database engine via which applications or other entities may access the state file database.
SYSTEMS AND METHODS FOR DYNAMIC ALLOCATION OF COMPUTING RESOURCES FOR MICROSERVICE ARCHITECTURE TYPE APPLICATIONS
Systems and methods for auto-scaling computing resources for applications that rely on microservices, such as media streaming applications. In example embodiments, the systems and methods use a hybrid approach that takes into account proactive and reactive metrics to predict future processing needs. The proactive metrics use data obtained from the microservices pipeline. In example embodiments, computing units are deployed or removed from a cluster of computing units in anticipation of a future computational need and in consideration of quality of service requirements.
Rescheduling of virtual machine migrations with less impact on an existing migration schedule
An apparatus, for an existing schedule indicating time slots during which virtual-machines are to be migrated between physical-machines, acquires change information indicating a change of first time slots during which first virtual-machines are to be migrated, and generates constraint information including a group of constraints regarding the change information and second time slots during which second virtual-machines other than the first virtual-machines are to be migrated. The apparatus generates semi-constraint information items each generated by excluding, from the constraint information, an exclusion target constraint that is selected in turn from among the group of constraints, and generates, for each semi-constraint information item, a rescheduling result by rescheduling migrations of the virtual-machines. The apparatus outputs a first rescheduling result that is one of the generated rescheduling results which reduces a difference between a scheduled migration time indicated by the first rescheduling result and scheduled migration times of the second virtual-machines.