Patent classifications
G06F11/2069
HITLESS CONTAINER UPGRADE WITHOUT AN ORCHESTRATOR
Systems, methods, and computer-readable media are disclosed for performing a hitless upgrade of executable code in the absence of an orchestrator or other upgrade manager. A mechanism is disclosed that utilizes containers to update software functionality, features, or the like without interrupting a service provided by a container and without relying on an orchestrator or other upgrade manager to coordinate the upgrade process. State information indicative of a current state of module(s) within a container is maintained in an external data store such as a state database. A hand-off from a current container to a new container that updates module code/functionality of the current container can be initiated upon determining that a state metric calculated by the old container at a future timestamp matches a state metric independently calculated by the new container at the same timestamp.
Providing coordination between mapper circuitry and redundant array of independent disks (RAID) circuitry to manage RAID storage space
Techniques are directed to managing Redundant Array of Independent Disks (RAID) storage space. One technique involves providing, by RAID circuitry, a storage space request to mapper circuitry; receiving, by the RAID circuitry, a storage space reply from the mapper circuitry in response to the storage space request; and performing, by the RAID circuitry, a follow-up operation based on the storage space reply from the mapper circuitry. Another technique involves, receiving, by mapper circuitry, a storage space request from RAID circuitry; performing, by the mapper circuitry, a storage space management operation in response to the storage space request; and providing, by the mapper circuitry, a storage space reply to the RAID circuitry, the storage space reply identifying a result of the storage space management operation.
Systems and methods for enabling a highly available managed failover service
a computing system that receives and stores configuration information for the application in a data store. The configuration information comprises (1) identifiers for a plurality of cells of the application that include at least a primary cell and a secondary cell, (2) a defined state for each of the plurality of cells, (3) one or more dependencies for the application, and (4) a failover workflow defining actions to take in a failover event. The computing system receives an indication, from a customer, of a change in state of the primary cell or a request to initiate the failover event. The computing system updates, in the data store, the states for corresponding cells of the plurality of cells based on the failover workflow and updates, in the data store, the one or more dependencies for the application based on the failover workflow.
Servicing of network software components of nodes of a cluster storage system
Described herein are method and apparatus for servicing software components of nodes of a cluster storage system. During data-access sessions with clients, client IDs and file handles for accessing files are produced and stored to clients and stored (as session data) to each node. A serviced node is taken offline, whereby network connections to clients are disconnected. Each disconnected client is configured to retain its client ID and file handles and attempt reconnections. Session data of the serviced node is made available to a partner node (by transferring session data to the partner node). After clients have reconnected to the partner node, the clients may use the retained client IDs and file handles to continue a data-access session with the partner node since the partner node has access to the session data of the serviced node and thus will recognize and accept the retained client ID and file handles.
Node Mode Adjustment Method for when Storage Cluster BBU Fails and Related Component
A node mode adjustment method for when a storage cluster BBU fails and related component. The storage cluster comprises multiple IO GROUPs; each IO GROUP comprises two nodes, and each node is connected to a corresponding BBU of the node. The method comprises: monitoring the status of each BBU in the storage cluster; after detecting that any one BBU has failed, when it is determined that the BBU connected to the peer node of the node to which the failed BBU is connected has not failed, and it is determined that the storage cluster is not downgraded, maintaining in a write-back mode the node connected to the BBU detected to have failed and the peer node of said node. Applying the solution of the present application will help to ensure that the storage cluster has the advantages of high and high IO performance.
Vendor-neutral models of vendors' application resources
Vendor-neutral models of vendors' application resources are described. A host outputs capabilities of data protection operations which are specified by a vendor of an application that is installed on the host. The host inputs a vendor-neutral version of a data protection operation, based on any of the capabilities, for a resource of the application. The host uses a vendor-neutral model of the resource of the application to perform the vendor-neutral version of the data protection operation on the application resource.
FAST RECOVERY WITH ENHANCED RAID PROTECTION
The disclosure includes a computer-implemented method for providing fast data access after a drive failure, a computer program product, and a RAID controller. One embodiment may comprise identifying a RAID array, the RAID array comprising a plurality of storage volumes, identifying an unused block of a provisioned volume in the RAID array, and copying a redundant copy of high value host writes to the unused block. The copying may comprise, for primary strips in the RAID array, creating one or more secondary strips mirroring the primary strips such that each of pair of primary-secondary strips reside on different storage volumes from each other.
Streaming database recovery using cluster live mounts
A database can be instantly recovered by a cluster mapped to the database. Nodes of the cluster are mapped over channels to directories of the database. Scripts are generated from one or more templates that specify the order and values to be executed to perform a database job, such as database recovery. To recover the database, a template is executed that generates and populates scripts, which are processed on the host of the database to recover the database in a nearly instant manner without transferring data files.
Fast recovery with enhanced raid protection
The disclosure includes a computer-implemented method for providing fast data access after a drive failure, a computer program product, and a RAID controller. One embodiment may comprise identifying a RAID array, the RAID array comprising a plurality of storage volumes, identifying an unused block of a provisioned volume in the RAID array, and copying a redundant copy of high value host writes to the unused block. The copying may comprise, for primary strips in the RAID array, creating one or more secondary strips mirroring the primary strips such that each of pair of primary-secondary strips reside on different storage volumes from each other.
Method and apparatus for designing dual-mirror shared conf partition file
A method, an apparatus, a device and a computer readable storage medium for designing a dual-mirror shared conf partition file are provided. The method includes pre-configuring disk space occupation capacities for a first flash mirror file, a second flash mirror file, and a shared conf partition of the BMC, and generating a FW file of the BMC which does not include a shared conf partition file. The method further includes mounting partitions configured in one of the first and second flash mirror files firstly and then mounting the shared Conf partition at last, in response to an instruction for starting the one of the first and second flash mirror files. The shared conf partition stores a configuration file of the BMC. The shared conf partition and the configuration file are automatically generated when the BMC runs normally for the first time.