G06F16/1844

Directing Operations To Synchronously Replicated Storage Systems

Managing connectivity to synchronously replicated storage systems, including: identifying a plurality of storage systems across which a dataset is synchronously replicated; identifying a host that can issue I/O operations directed to the dataset; identifying a plurality of data communications paths between the host and the plurality of storage systems across which a dataset is synchronously replicated; identifying, from amongst the plurality of data communications paths between the host and the plurality of storage systems across which a dataset is synchronously replicated, one or more optimal paths; and issuing, to the host, an identification of the one or more optimal paths.

Smart archiving and data previewing for mobile devices

Techniques for providing data preview before recalling large data files are disclosed. In one aspect, a data file is made accessible while being offline by converting the data file from a native format to a preview format, storing the data file in the preview format in a primary storage that is locally available and moving, after the conversion to the preview format, the data file in the native format to a secondary storage. When a viewing request is received for the data file, the data file in the preview format is displayed to fulfill the viewing request.

Efficient update-anywhere replication of queue operations on a replicated message queue
11494347 · 2022-11-08 · ·

Update-anywhere replication of queuing operations on a replicated message queue is performed. A dequeue ready time (“ready time”) is associated by each participating persistent storage server with a queue message to be dequeued. Unless a queue message is already locked by a distributed dequeue transaction, a participating leader PSS initiates a distributed dequeue transaction for the queue message once the ready time for the queue message is reached, subject to certain conditions. An initiator PSS is in effect designated for a queue message; the initiator PSS associates a desired ready time for the queue message. The designated PSS is referred to herein as the primary leader PSS and the ready time the primary PSS associates with queue message is referred as the primary ready time. The other participating leader PSSs are backup leader PSSs which serve as backups for dequeuing a queue message. Each backup leader PSS associates a later “backup ready time” with the queue message. In an embodiment, each backup ready time for a queue message is different. The primary ready time for a queue message together with backup ready times for the queue message form a more or less staggered set of ready times. If the primary leader PSS does not initiate dequeuing of a queue message before a successive backup ready time, the respective backup leader PSS may initiate dequeuing. This measure provides fault tolerance for dequeuing a queue message. Because the ready times are staggered, not all PSSs will initiate dequeuing at more or less the same time. Thus, initiator conflict is substantially reduced while fault tolerance is provided.

Document replication based on distributional semantics
11487709 · 2022-11-01 · ·

Embodiments of the present invention are directed toward systems, methods, and computer storage media for using a neural network language model to identify semantic relationships between file storage specifications for replication requests. By treating file storage specifications (or at least a portion thereof) as “words” in the language model, replication vectors can be determined based on the file storage specifications. Instead of determining the relationship of the file storage specifications based on ordering within a document, the relationship can be based on proximity of the replication requests in a replication session. When a replication request is received from a user, the replication vectors can be used to determine a semantic similarity between the received replication request and one or more additional replication requests.

FILE SYSTEM CLONING METHOD AND APPARATUS
20220350779 · 2022-11-03 ·

A file system cloning method and apparatus is provided. In the method, a destination storage system first receives first information from a source storage system in which a file system runs. The first information is used to indicate a data layout of the file system. Then, the destination storage system creates a cloned file system of the file system based on the first information.

Resiliency in a cloud-based storage system
11487715 · 2022-11-01 · ·

Data resiliency in a cloud-based storage system, including: receiving, for storage within a first tier of cloud storage of the cloud-based storage system, one or more segments of data; generating, for each of one or more shards of data of the one or more segments of data, self-describing information for recoverability of the one or more shards of data; and storing, within a second tier of cloud storage of the cloud-based storage system, both the one or more shards of data and the generated self-describing information for recoverability of the one or more shards of data.

Data delivery and validation in hybrid cloud environments

Data content delivery and validation in a computer environment may provide a file system in the computer environment, the file system subdivided into unique folder locations per content type, each of the unique folder locations representing a content type folder. The file system is monitored for changes to a content type folder. An occurrence of a manifest file in the content type folder may be detected, the occurrence of the manifest file ensuring that all files in a package of files associated with the manifest file have arrived. Content of the manifest file may be analyzed to check validity of the files. A content package registry may be queried to determine a base job for processing a given content type associated with the package of files, and the base job may be run to process the package of files.

METHOD AND APPARATUS FOR REPLICATING A TARGET FILE BETWEEN DEVICES

There is provided a method and apparatus for remote differential compression (RDC) and data deduplication. According to embodiments, when a sending device acquires a new target file, the following steps are performed. Initially, Jaccard segmentation is performed, followed by performing identity-based segment deduplication and similarity-based segment deduplication. The transmission of the target file in the deduplicated form to the recipient device is subsequently performed. The recipient device can then rebuild the original target file from the deduplicated form thus replicating the target file at the recipient device with the target file originally present at the sending device.

REPLICATION FOR SNAPSHOT REVERSION
20220342905 · 2022-10-27 · ·

A method for snapshot reversion, the method may include initializing a recovery of a storage system, from a failure that stopped a replication of a failed replication snapshot to an other storage system; wherein the other storage system reverted to an older snapshot not stored in the storage system; and reverting to the older snapshot, by the storage system using older snapshot metadata and data received from the storage system; wherein the older snapshot data may include a non-existing indication for a snapshot segment that existed in a snapshot that followed the older snapshot and not exists at the older snapshot, and wherein the data may include one or more older snapshot segments; wherein the reverting to the older snapshot is executed without undoing changes that were made since the older snapshot.

SYNCHRONOUS REMOTE REPLICATION OF SNAPSHOTS

Snapshots from a first LSU (R1) on a first storage system (A1) may be replicated to a second replica LSU (R2) on a second storage system (A2), for example, concurrently to remotely replicating (e.g., synchronously) write operations for R1 to R2. A process, P, on A1 executing the replication of the snapshots from R1 to R2 may be a separate process than the one or more processes on A1 executing remote replication of write operations for R1 to R2. During a consistency window on A1, outstanding write operations for R1 at the time the consistency window opened may be logged, and a pair of snapshots, SS1.sub.1 and SS1.sub.2 may be activated on R1 and R2, respectively. After the consistency window has closed, the SS1.sub.2 snapshot metadata and snapshot data may be updated based on the outstanding write operations.