Patent classifications
G06F11/1482
ACTIVE-STANDBY PODS IN A CONTAINER ORCHESTRATION ENVIRONMENT
Container orchestration platforms, such as Kubernetes, automatically manage the execution of applications in containers on host computing systems. An application may include component applications that execute in different containers and a container orchestration platform may organize the containers for those component applications into a pod for the application. To spread the load for the application, the container orchestration platform may enable creation of more than one pod for a single application. Once created, any of the multiple pods may be selected to handle a request to the application. As such, all of the pods are considered active by the container orchestration platform.
Query retry using quiesce notification
The subject technology requests information regarding an instance identifier of a compute service manager instance to a particular job. The subject technology retrieves information related to a set of instances of compute service managers in a set of virtual warehouses. The subject technology filters the information to determine a set of candidates from the set of instances of compute service managers. The subject technology sorts the set of candidates based at least in part on a workload. The subject technology selects a candidate compute service manager to issue a query restart by randomly selecting an execution node.
Transparent checkpointing and process migration in a distributed system
A distributed system for creating a checkpoint for a plurality of processes running on the distributed system. The distributed system includes a plurality of compute nodes with an operating system executing on each compute node. A checkpoint library resides at the user level on each of the compute nodes, and the checkpoint library is transparent to the operating system residing on the same compute node and to the other compute nodes. Each checkpoint library uses a windowed messaging logging protocol for checkpointing of the distributed system. Processes participating in a distributed computation on the distributed system may be migrated from one compute node to another compute node in the distributed system by re-mapping of hardware addresses using the checkpoint library.
Distributed components in computing clusters
The subject disclosure is directed towards components in different server clusters, e.g., comprising software components such as components of a distributed computing system. Components are available for use by distributed computing system applications, yet managed by the distributed computing system runtime such that only a single instance can be activated and exist within communicating (non-partitioned) clusters. Also described is recovery from a situation in which no longer partitioned clusters each have created the same component.
QUIESCE NOTIFICATIONS FOR QUERY RETRIES
The subject technology retrieves information related to a set of instances of compute service managers, each instance of a particular compute service manager being associated with a set of virtual warehouses. The subject technology filters the information to determine a set of candidates from the set of instances of compute service managers. The subject technology sorts the set of candidates based at least in part on each workload of each of the set of candidates. The subject technology selects a candidate compute service manager from the set of instances of compute service managers to issue a query restart by randomly selecting an execution node, the execution node being included in a particular virtual warehouse associated with the candidate compute service manager, the selecting facilitating improving utilization of cluster resources and improving query execution on the selected candidate compute service manager.
Data Center Restoration and Migration
A system can maintain a first data center that comprises a virtualized overlay network and virtualized volume identifiers. The system can determine to perform a restore of data of the first data center to a second data center, the data comprising first instances of virtualized workloads. The system can transfer the data to the second data center. The system can configure the second data center with the virtualized overlay network and the virtualized volume identifiers. The system can operate the virtualized workloads on the second data center, the second instances of the virtualized workloads invoking the second instance of the virtualized overlay network and the second instance of the virtualized volume identifiers.
Data Center Restoration
A system can determine to restore a datacenter that comprises a group of virtualized workloads. The system can determine respective associations between respective virtualized workloads and respective datastores. The system can determine to restore a first virtualized workload of the group of virtualized workloads first. The system can restore a first portion of infrastructure that corresponds to the first virtualized workload first among a group of infrastructure. The system can, after restoring the first portion of infrastructure, restore a first portion of data that corresponds to the first virtualized workload first among a group of data. The system can, after restoring the first portion of data, restore a first portion of a virtualization layer that corresponds to the first virtualized workload first among a group of virtualization layers. The system can, after restoring the first portion of the virtualization layer, restore the first virtualized workload.
Standby Data Center as a Service
A system can maintain a first data center that comprises a virtualized overlay network and virtualized volume identifiers, and store data comprising virtualized workloads. The system can determine a service level agreement associated with providing a second data center as a backup to the first data center. The system can, based on the service level agreement, divide, into a first portion of tasks and a second portion of tasks deploying the data to a secondary storage of the second data center, deploying the data to a primary storage of the second data center, and configuring the second data center with the virtualized overlay network and the virtualized volume identifiers. The system can perform the first portion of tasks before determining to restore the first data center to the second data center. The system can perform the second portion of tasks in response to determining to restore the first data center.
Failure shield
An example graphics system can include a first portion including a graphics driver and graphics hardware and a second portion communicatively coupled to the first portion. The second portion can include a display system communicatively coupled to a GUI application and a shim layer to shield the second portion from failure responsive to failure of the first portion.
Storage system and control software deployment method
A storage system includes a plurality of storage nodes. The storage node includes: one or more storage devices which respectively provide a storage area; and one or more control software which read/write requested data from/into the corresponding storage device according to a request from a higher-level device, wherein each of the control software retains predetermined configuration information required for reading/writing requested data from/into the corresponding storage device according to a request from the higher-level device, wherein a plurality of the control software are managed as a redundancy group, and the configuration information retained in each of the control software belonging to the same redundancy group is synchronously updated, and wherein the plurality of control software configuring the redundancy group are each deployed in respectively different storage nodes so as to distribute a load of each of the storage nodes.