Patent classifications
G06F11/2094
SNAPSHOT SHIPPING TO MULTIPLE CLOUD DESTINATIONS
An apparatus comprises at least one processing device configured to identify a snapshot lineage comprising snapshots of a storage volume, the snapshot lineage comprising (i) a local snapshot lineage stored on a storage system and (ii) cloud snapshot lineages stored on cloud storage external to the storage system, to select at least one snapshot that is to be copied from the local snapshot lineage, to determine at least two of the cloud snapshot lineages as destinations for the selected snapshot, to generate a snapshot copy job for copying the selected snapshot to the at least two cloud snapshot lineages, and to process the snapshot copy job by reading data of the selected snapshot stored in the local snapshot lineage once and writing the data of the selected snapshot to the at least two cloud snapshot lineages.
ENVOY FOR MULTI-TENANT COMPUTE INFRASTRUCTURE
A data management and storage (DMS) cluster of peer DMS nodes manages data of a tenant of a multi-tenant compute infrastructure. The compute infrastructure includes an envoy connecting the DMS cluster to virtual machines of the tenant executing on the compute infrastructure. The envoy provides the DMS cluster with access to the virtual tenant network and the virtual machines of the tenant connected via the virtual tenant network for DMS services such as data fetch jobs to generate snapshots of the virtual machines. The envoy sends the snapshot from the virtual machine to a peer DMS node via the connection for storage within the DMS cluster. The envoy provides the DMS cluster with secure access to authorized tenants of the compute infrastructure while maintaining data isolation of tenants within the compute infrastructure.
FRAMEWORK FOR LIVE DATA MIGRATION
Systems and methods including a framework for migration of live data. The method may comprised, by one or more hardware processors executing program instructions, receiving, at a migration proxy of the framework, code for reading data and writing data compatible with each of a plurality of states of a migration of data in a data store, wherein a service is at least intermittently reading data from and writing data to the data store; determining, by a migration runner of the framework, to perform the migration of the data; initiating, by the migration runner, the migration of the data, wherein the migration comprises a plurality of stages; storing, as the migration progresses through the plurality of stages, and at a migration data store of the framework, a current stage of the migration; and during the migration, using the migration proxy to read data from and write data to the data store.
RESILIENT DATA STORAGE SYSTEM WITH EFFICIENT SPACE MANAGEMENT
A storage system has a plurality of storage nodes having equal non-volatile storage capacity that is subdivided into equal size cells. Host application data that is stored in the cells is protected using RAID or EC protection groups each having members stored in ones of the cells and distributed across the storage nodes such that no more than one member of any single protection group is stored by any one of the storage nodes. Spare cells are maintained for rebuilding protection group members of a failed one of the storage nodes on remaining non-failed storage nodes so full data access is possible before replacement or repair of the failed storage node.
STORAGE VOLUME SYNCHRONIZATIONS RESPONSIVE TO COMMUNICATION LINK RECOVERIES
In some examples, a system detects recovery, from an unavailable state, of a communication link between a first storage system that includes a first storage volume and a second storage system that includes a second storage volume that is to be a synchronized version of the first storage volume, where while the communication link is in the unavailable state the second storage volume is in an offline state and the first storage volume is in an online state. In response to detecting the recovery of the communication link, the system sends a first tracking metadata for the first storage volume from the first storage system to the second storage system, and in response to receipt of the first tracking metadata at the second storage system that maintains a second tracking metadata for the second storage volume, the system transitions the second storage volume from the offline state to a controlled online state, and initiates a synchronization process to synchronize the second storage volume with the first storage volume.
Automated orchestration of cyber protection of a set of storage volumes
An Orchestrated Data Recovery (ODR) Cyber Protection Automation (CPA) operates to ensure one-to-one creation of snapsets of a production site and corresponding snapsets of a cyber vault. During an initiation phase, the ODR CPA monitors synchronization of a snapset of production volumes from the production site to the cyber vault. If additional snapsets of the production volumes are created prior to completion of synchronization of the first snapset, the additional snapsets are also synchronized to the cyber vault. Once the initial synchronization of the storage volumes has been completed, the ODR CPA causes a Storage Volume Creation and Management System (SVCMS) to create a snapset of the storage volumes at the cyber vault. Subsequently, each time a snapset is created of the production site, the ODR CPA orchestrates synchronization of the snapset to the cyber vault and creation of a corresponding snapset at the cyber vault.
Virtualized file server smart data ingestion
In one embodiment, a system for managing a virtualization environment includes a set of host machines, each of which includes a hypervisor, virtual machines, and a virtual machine controller, and a data migration system configured to identify one or more existing storage items stored at one or more existing File Server Virtual Machines (FSVMs) of an existing virtualized file server (VFS). For each of the existing storage items, the data migration system is configured to identify a new FSVMs of a new VFS based on the existing FSVM, send a representation of the storage item from the existing FSVM to the new FSVM, such that representations of storage items are sent between different pairs of FSVMs in parallel, and store a new storage item at the new FSVM, such that the new storage item is based on the representation of the existing storage item received by the new FSVM.
Intelligent post-packaging repair
Techniques are provided for storing a row address of a defective row of memory cells to a bank of non-volatile storage elements (e.g., fuses or anti-fuses). After a memory device has been packaged, one or more rows of memory cells may become defective. In order to repair (e.g., replace) the rows, a post-package repair (PPR) operation may occur to replace the defective row with a redundant row of the memory array. To replace the defective row with a redundant row, an address of the defective row may be stored (e.g., mapped) to an available bank of non-volatile storage elements that is associated with a redundant row. Based on the bank of non-volatile storage elements the address of the defective row, subsequent access operations may utilize the redundant row and not the defective row.
Methods and systems for rapid failure recovery for a distributed storage system
Methods and systems are provided for rapid failure recovery for a distributed storage system for failures by one or more nodes.
Fileset partitioning for data storage and management
In one approach, filesets to be backed up are divided into partitions and snapshots are pulled for each partition. In one architecture, a data management and storage (DMS) cluster includes a plurality of peer DMS nodes and a distributed data store implemented across the peer DMS nodes. One of the peer DMS nodes receives fileset metadata for the fileset and defines a plurality of partitions for the fileset based on the fileset metadata. The peer DMS nodes operate autonomously to execute jobs to pull snapshots for each of the partitions and to store the snapshots of the partitions in the distributed data store.