Patent classifications
G06F11/1662
SYNCHRONIZATION OF A DATABASE BY RESTORING COPIES OF CHANGED DATABASE OBJECTS
Systems and methods for performing backup and other secondary copy operations for large databases (e.g., “big data”), such as the Greenplum database, are described. In some cases, the systems and methods may maintain a second instance of a source database (e.g., Greenplum) using live synchronization (e.g., “Live Sync”), which performs incremental replication between a virtual machine containing a large database (e.g., a virtual machine containing a Greenplum database) and a synced copy of the virtual machine.
Low cost, heterogeneous method of transforming replicated data for consumption in the cloud
Disclosed are methods and the like that provide for transforming replicated data for consumption in the cloud, for example. Such methods can include attaching a target gateway node at a secondary site to a storage device at the secondary site, searching for an identifier stored in the storage device, and storing replicated data in the replication volume. The identifier is associated with an offset stored in the storage device, and the offset identifies a starting location of a replication volume in the storage device. The replicated data is received by the target gateway node from a source gateway node at a primary site. A starting location is received with the replicated data. The target gateway node stores the replicated data at a first location in the storage volume, and the first location is determined based, at least in part, on the starting location and the first storage location.
Regulating storage device rebuild rate in a storage system
An apparatus comprises at least one processing device comprising a processor coupled to a memory; the at least one processing device being configured to: obtain a set of rebuild rate parameters for a given storage device from a storage array comprising a plurality of storage devices; and dynamically regulate a rebuild rate associated with a rebuild process for the given storage device based on the set of rebuild rate parameters obtained from the storage array for the given storage device. For example, the set of rebuild rate parameters include a rebuild capacity parameter and a rebuild time parameter.
METHODS AND APPARATUS FOR PROVIDING HYPERVISOR LEVEL DATA SERVICES FOR SERVER VIRTUALIZATION
A data center for data backup and replication, including a pool of multiple storage units for storing a journal of I/O write commands issued at respective times, wherein the journal spans a history window of a pre-specified time length, and a journal manager for dynamically allocating more storage units for storing the journal as the journal size increases, and for dynamically releasing storage units as the journal size decreases.
SYNCHRONIZATION STORAGE SOLUTION AFTER AN OFFLINE EVENT
Disclosed are systems and methods of synchronization between a source and a target. The synchronization relationship can be quickly and easily be created for disaster recovery, real-time backup and failover, thereby ensuring that data on the source is fully-protected at an off-site location or on another server or VM, for example, at another data center, a different building or elsewhere in the cloud. Common snapshots available on both the source and target can act as common recovery points. The common recovery points can be used to locate the most recent snapshot in common, between the source and target, to enable a delta sync of all subsequently written data at the source to the target after an offline event.
Redundant cloud memory storage for a memory subsystem
A method for managing memory element failures in a memory subsystem is described. The method includes detecting, by the memory subsystem, a failed memory element in the memory subsystem and transmitting a redundant memory request based on detection of the failed memory element. The redundant memory request seeks to utilize memory storage in an external storage system in place of the failed memory element in the memory subsystem. Thereafter, the memory subsystem receives, from the external storage system, a redundant memory request confirmation, which indicates that the redundant memory request has been fulfilled and includes an address of a location in the external storage system. In response to receipt of the redundant memory request confirmation, the memory subsystem updates memory management information to map a logical address, which was previously mapped to the failed memory element, to the location in the external storage system.
Updating stateful system in server cluster
A computer-implemented method for seamlessly performing a maintenance operation on a stateful system includes mapping a network address of the stateful system to a primary server that uses a primary database to respond to incoming data requests. In response to receiving a maintenance request, the primary database is replicated to a secondary database of a secondary server. The secondary server is updated according to the maintenance request. The method further includes caching, in a replay buffer of the primary server, incoming data requests during the replicating. After the replicating, the data requests from the replay buffer are executed by the secondary server. Write operations to the primary server are disabled during the replicating, and the network address of the stateful system is mapped to the secondary server. Subsequently, the primary server is updated and reinstated by mapping the network address, and enabling the write operations.
ENABLING DATA INTEGRITY CHECKING AND FASTER APPLICATION RECOVERY IN SYNCHRONOUS REPLICATED DATASETS
One or more techniques and/or computing devices are provided for utilizing snapshots for data integrity validation and/or faster application recovery. For example, a first storage controller, hosting first storage, has a synchronous replication relationship with a second storage controller hosting second storage. A snapshot replication policy rule is defined to specify that a replication label is to be used for snapshot create requests, targeting the first storage, that are to be replicated to the second storage. A snapshot creation policy is created to issue snapshot create requests comprising the replication label. Thus a snapshot of the first storage and a replication snapshot of the second storage are created based upon a snapshot create request comprising the replication label. The snapshot and the replication snapshot may be compared for data integrity validation (e.g., determine whether the snapshots comprise the same data) and/or quickly recovering an application after a disaster.
IDENTIFYING FAULT DOMAINS FOR DELTA COMPONENTS OF A DISTRIBUTED DATA OBJECT
The disclosure herein describes placing delta components of a base component in target fault domains. One or more delta components are generated. When a first fault domain that lacks a sibling component of the base component is identified, the first fault domain is selected as a single delta target fault domain and a single delta component is placed on the single delta target fault domain. When a second fault domain that includes a first sibling component of the base component is identified and a third fault domain that includes a second sibling component of the base component is identified, the second fault domain and the third fault domain are selected as a first double delta target fault domain and a second double delta target fault domain, and a first double delta component and a second double delta component are placed on the first and second double delta target fault domains.
Availability of a storage system
Embodiments of the present disclosure provide a method for a storage system, a storage system and a computer program product. The method comprises determining a first drive in a drive array is temporarily unavailable. The method further comprises setting the first drive in a frozen state. The method further comprises: in response to receiving a write request for the first drive during the frozen state, pending the write request or recording the write request in a second drive in the drive array. The method further comprises: in response to receiving a read request for the first drive during the frozen state, reconstructing data to which the read request is directed through data stored in a third drive in the drive array.