G06F11/2092

Data reporting method and system

A data reporting method includes generating a data uploading token by a main network node, determining, from a plurality of data-uploading network nodes waiting for data uploading, a plurality of qualified network nodes having a data uploading qualification, allowing the qualified network nodes to compete for the data uploading token, and coordinating a data uploading operation with a winning network node of the qualified network nodes that obtained the data uploading token.

WRITING SEGMENTS AND ERASE BLOCKS BASED ON ORDERING

A method is provided. The method includes receiving a set of data blocks to be stored in a storage system. The storage system includes a plurality of non-volatile memory modules. The method also includes generating a set of segments based on the set of data blocks. A respective segment comprising portions of one or more erase blocks. The method further includes writing the set of segments to the non-volatile memory modules based on orderings of the portions of the one or more erase blocks.

TECHNIQUES FOR LIF PLACEMENT IN SAN STORAGE CLUSTER SYNCHRONOUS DISASTER RECOVERY

Improved techniques for disaster recover within storage area networks are disclosed. Embodiments include replicating a LIF of a primary cluster on a secondary cluster. LIF configuration information is extracted from the primary cluster. A peer node from a secondary cluster is located. One or more ports are located on the located peer node that match a connectivity of the LIF from the primary cluster. One or more ports are identified based upon one or more filtering criteria to generate a candidate port list. A port from the candidate port list is selected based at least upon a load of the port. Other embodiments are described and claimed.

SYSTEMS AND METHODS FOR PERFORMING A TECHNICAL RECOVERY IN A CLOUD ENVIRONMENT

A computer-implemented method for testing failover may include: determining one or more cross-regional dependencies and traffic flow of an application in a first region of a cloud environment, wherein the one or more cross-regional dependencies include a dependency of the application in the first region of the cloud environment to one or more applications in at least one other region of the cloud environment; determining a risk score associated with performing failover of the application to a second region of the cloud environment at least based on the determined one or more cross-regional dependencies and traffic flow of the application; comparing the determined risk score with a predetermined risk score; in response to determining that the determined risk score is lower than the predetermined risk score, performing failover of the application to the second region of the cloud environment; isolating the second region of the cloud environment from the first region of the cloud environment for a predetermined period of time; and monitoring operation of the application in the second region of the cloud environment during the predetermined period of time.

FACILITATING IMMEDIATE PERFORMANCE OF VOLUME RESYNCHRONIZATION WITH THE USE OF PASSIVE CACHE ENTRIES

Systems and methods for reducing delays between the time at which a need for a resynchronization of data replication between a volume of a local CG and its peer volume of a remote CG is detected and the time at which the resynchronization is triggered (Reseed Time Period) are provided. According to an example, information indicative of the direction of data replication between the volume and the peer volume is maintained within a cache of a node. Responsive to a disruptive operation (e.g., relocation of the volume from an original node to a new node), the Reseed Time Period is lessened by proactively adding a passive cache entry to a cache within the new node at the time the CG relationship is created when the new node represents an HA partner of the original node and prior to the volume coming online when the new node represents a non-HA partner.

SHARE-BASED FILE SERVER REPLICATION FOR DISASTER RECOVERY

A file server manager disclosed herein accesses information regarding a selected share of a source distributed file server for replication, where the selected share stores at least a portion of a namespace of storage items and is hosted by a first file server virtual machine of the source distributed file server. The file server manager accesses a mapping between virtual machines of the source distributed file server and virtual machines of the destination distributed file server and replicates the selected share to a second file server virtual machine of the destination distributed file server based on the mapping. The file server manager directs a request to read a first storage item to the destination distributed file server and directs a request to write to a second storage item to the source distributed file server while the destination distributed file server services the request to read the first storage item.

DATA STORAGE CLUSTER WITH QUORUM SERVICE PROTECTION
20220334934 · 2022-10-20 ·

In a protective quorum service, during an initial period of normal operation in which a clustered pair of data storage nodes provide host I/O access to a data storage object and replicate write-type requests to each other, the nodes are first registered to the quorum service. Subsequently, based on the registration and in response to a first auto promote request from a first-requesting node, a success response is returned and the service enters an auto promoted condition, the success response indicating that the first-requesting node is to continue providing the host I/O access to the data storage object without write replication. In response to receiving a subsequent auto promote request from the other node when in the auto promoted condition, a failure response is returned indicating that the other node is to cease providing host I/O access to the data storage object.

DATA CENTER STORAGE AVAILABILITY ARCHITECTURE USING RACK-LEVEL NETWORK FABRIC

Operations are monitored that involve a plurality of servers coupled to a plurality of data storage enclosures via a rack-level, storage networking fabric. The servers are operable to provide data storage services utilizing the data storage enclosures via a network. The data storage enclosures each have one or more data storage devices. The servers and the data storage enclosures are mounted within a data center rack. A failed unit determined that includes a failed one of the servers or a failed one of the data storage devices within a selected one of the data storage enclosures. A replacement is found for the failed unit within the data center rack, and the replacement is established for the failed unit within the data center rack. The establishment of the replacement involves establishing a new connection via the storage networking fabric.

Virtualized file server user views

In one embodiment, a system for managing a virtualization environment includes a plurality of host machines, wherein each of the host machines comprises a hypervisor and one or more user virtual machines (user VMs), and a virtual machine controller, one or more virtual disks comprising a plurality of storage devices, a virtualized file server (VFS) comprising a plurality of file server virtual machines (FSVMs), wherein each of the FSVMs is running on one of the host machines. The VFS may be configured to receive a request for storage system information from a user and generate and send a response to the request, wherein the response is customized according to configuration information of the VFS that is specific to the user. The storage system information requested may include a total size of storage available to the user, and the user may have an associated storage quota limit.

Customized hash algorithms
11652884 · 2023-05-16 · ·

A storage system determines source addresses, and destination addresses in a storage system, for network traffic. The storage system determines a hash algorithm, from a plurality of hash algorithms. The hash algorithm is to be used across the source addresses for load-balancing the network traffic to the destination addresses. The storage system determines that the hash algorithm more closely meets one or more load-balancing criteria than at least one other hash algorithm, of the plurality of hash algorithms. The storage system distributes the network traffic from the source addresses to the destination addresses in the storage system, with load-balancing according to the determined hash algorithm.