G06F11/1662

Staging data within a unified storage element

Staging data on a storage element integrating fast durable storage and bulk durable storage, including: receiving, at a storage element integrating fast durable storage and bulk durable storage, a data storage operation from a host computer; storing data corresponding to the data storage operation within fast durable storage in accordance with a first data resiliency technique; and responsive to detecting a condition for transferring data between fast durable storage and bulk durable storage, transferring the data from fast durable storage to bulk durable storage in accordance with a second data resiliency technique.

Initiating Recovery Actions When A Dataset Ceases To Be Synchronously Replicated Across A Set Of Storage Systems
20220358019 · 2022-11-10 ·

Initiating recovery actions when a dataset ceases to be synchronously replicated across a set of storage systems, including: receiving, by at least one storage system among a plurality of storage systems implementing a symmetric input/output model for a synchronously replicated dataset, a request to modify the dataset; identifying one or more operations associated with the request to modify the dataset that have not been applied to at least one storage system of the plurality of storage systems; and responsive to a system fault among the plurality of storage systems synchronously replicating the dataset, applying a recovery action based on recovery information that identifies one or more operations that have not been applied to the plurality of storage systems.

METHOD FOR FULL DATA RECONTRUCTION IN A RAID SYSTEM HAVING A PROTECTION POOL OF STORAGE UNITS
20220357881 · 2022-11-10 ·

A method of performing a full data reconstruction in a redundant array of independent disks (RAID) system with a protection pool of storage units includes determining that a physical disk of a storage cluster has been removed from service. The physical disk includes a set of physical extents and at least one physical extent of the set of physical extents is associated with an array of physical extents distributed across physical disks of the storage cluster. The method further includes transmitting a message to of the physical disks, to allocate replacement physical extents and assign the replacement physical extents to the array of physical extents and initiating reconstruction of data from the set of physical extents of the physical disk to the replacement physical extents.

Data synchronization in a distributed data storage system
11494404 · 2022-11-08 · ·

The present disclosure relates to a method in a distributed and non-hierarchical node comprising a set of data items for determining a synchronization state between said node and one or more distributed and non-hierarchical nodes communicatively coupled to form a cluster, wherein the set of data items are locally replicated at each node. The method comprises generating a snapshot comprising information relating to the set of data items at a snapshot time T.sub.s, said information identifying changes to the data items where each change comprises a time stamp, where only changes to the data items with a time stamp<T.sub.0, where T.sub.s>T.sub.0, are included in the snapshot, although changes to the data items up to the time T.sub.s, have been received and stored in the nodes, receiving corresponding generated snapshots from all other nodes in the cluster, and determining a synchronization state based on a comparison between the generated and received snapshots. The synchronization state is determined to be not synchronized if the generated snapshot and the received corresponding snapshots do not all match.

RESYNCHRONIZATION TO A SYNCHRONOUS REPLICATION RELATIONSHIP

Techniques are provided for resynchronizing a synchronous replication relationship. Asynchronous incremental transfers are performed to replicate data of a storage object to a replicated storage object. Incoming write requests, targeting the storage object, are logged into a dirty region log during a last asynchronous incremental transfer. Metadata operations, executed on the storage object, are logged into a metadata log during the last asynchronous incremental transfer. Sequence numbers are assigned to the metadata operations based upon an order of execution. The metadata operations are replicated to the replicated storage object for execution according to the sequence numbers, and the dirty regions are replicated to the replicated storage object in response to the metadata operations having been replicated to the replicated storage object. The storage object and replicated storage object are transitioned to a synchronous replication state where incoming operations are synchronously replicated to the replicated storage object.

HETEROGENEOUS INDEXING AND LOAD BALANCING OF BACKUP AND INDEXING RESOURCES

Indexing preferences generally associate each data source with a type of indexing technology and/or with an index/catalog and/or with a computing device that hosts the index/catalog for tracking backup data generated from the source data. Indexing preferences govern which index/catalog receives transaction logs for a given storage operation. Thus, indexing destinations are defined granularly and flexibly in reference to the source data. Load balancing without user intervention assures that the various index/catalogs are fairly distributed in the illustrative backup systems by autonomously initiating migration jobs. Criteria for initiating migration jobs are based on past usage and going-forward trends. An illustrative migration job re-associates data sources with a different destination media agent and/or index/catalog, including transferring some or all relevant transaction logs and/or indexing information from the old host to the new host.

Techniques for LIF placement in SAN storage cluster synchronous disaster recovery

Improved techniques for disaster recover within storage area networks are disclosed. Embodiments include replicating a LIF of a primary cluster on a secondary cluster. LIF configuration information is extracted from the primary cluster. A peer node from a secondary cluster is located. One or more ports are located on the located peer node that match a connectivity of the LIF from the primary cluster. One or more ports are identified based upon one or more filtering criteria to generate a candidate port list. A port from the candidate port list is selected based at least upon a load of the port. Other embodiments are described and claimed.

Risk detection of data loss for 5G enabled devices

Embodiments of the present invention provide a computer system, a computer program product, and a method that comprises collecting data capable of being replicated from a computing device; detecting risks of the computing device, wherein detecting risks comprises detecting the computing device's surroundings, location, speed, and condition; initiating data replication on the computing device once the risks are determined to reach a predetermined threshold; and storing the replicated data within a cloud storage system using a 5G network.

Identifying valid data after a storage system recovery

Staging data on a storage element integrating fast durable storage and bulk durable storage, including: receiving, at a storage element integrating fast durable storage and bulk durable storage, a data storage operation from a host computer; storing data corresponding to the data storage operation within fast durable storage in accordance with a first data resiliency technique; and responsive to detecting a condition for transferring data between fast durable storage and bulk durable storage, transferring the data from fast durable storage to bulk durable storage in accordance with a second data resiliency technique.

Data loss recovery in a secondary storage controller from a primary storage controller

A secondary storage controller determines one or more tracks of one or more volumes in which data loss has occurred in the secondary storage controller. The secondary storage controller suspends a peer to peer remote copy operation between the secondary storage controller and a primary storage controller. Information on the one or more tracks of the one or more volumes in which the data loss has occurred is transmitted to the primary storage controller.