Patent classifications
G06F2212/285
DISASTER RECOVERY AND REPLICATION IN DISAGGREGATED DATACENTERS
Embodiments for disaster recovery in a disaggregated computing system. Memory resources are allocated at a secondary, disaster recovery site for data received from a primary site. The data from the primary site is continuously replicated to the allocated memory resources at the disaster recovery site without requiring any compute resources to be attached to the allocated memory resources. Responsive to determining a disaster recovery failover is in progress, the compute resources are assigned to the allocated memory resources for performing a failover workload, and the failover workload is executed at the disaster recovery site.
Method and Apparatus for Facilitating a Trim Process Using Auxiliary Tables
A method for processing a trim command via an input and output (I/O) command of a solid-state drive (SSD) using various tables is disclosed. The process is able to retrieve a trim node state table (TNST) from a local memory in response to the trim command. Upon identifying current node status of the TNST associated with a logical block address (LBA) referenced by the trim command, a trim operation is processed to a node if the current node status indicates a pending state. After changing the node status to a dirty state when the current node status is in a clean state, the content of a trim invalid bitmap table (TIBT) is updated to indicate the status of LBAs when the current node status is in a dirty state.
Inline compression support using discrete sized containers for backing store
A storage array presents a logical production volume that is backed by tangible data storage devices. The production volume is organized into fixed size front end allocation units. The tangible data storage devices are organized into discrete size back end allocation units of a plurality of different sizes. Data associated with each one of the front end allocation units is stored on only one of the back end allocation units. For example, compressed data may be stored on a back end allocation unit that is smaller than a front end allocation unit while maintaining a 1-to-1 relationship between the front end allocation unit and the back end allocation unit.
POPULATING A SECOND CACHE WITH TRACKS FROM A FIRST CACHE WHEN TRANSFERRING MANAGEMENT OF THE TRACKS FROM A FIRST NODE TO A SECOND NODE
Provided are a computer program product, system, and method for populating a second cache with tracks from a first cache when transferring management of the tracks from a first node to a second node. Management of a first group of tracks in the storage managed by the first node is transferred to the second node managing access to a second group of tracks in the storage. After the transferring the management of the tracks, the second node manages access to the first and second groups of tracks and caches accessed tracks from the first and second groups in the second cache of the second node. The second cache of the second node is populated with the tracks in a first cache of the first node
Distributed caching cluster client configuration
A cache or other cluster is configuration-aware such that initialization and changes to the underlying structure of the cluster can be dynamically updated for use by a client. A client may use a client driver as an intermediary that is responsible for managing the communication with the cluster. For example, a client driver may resolve an alias from a static configuration endpoint to a storage node. The client driver may request an initial configuration from the storage node and then update configuration from one or more storage nodes that store current configuration of the cluster.
Cluster-based storage device buffering
Technologies are provided for storing data by alternating the performance of data write operations using multiple clusters of storage devices. Data is written to internal buffers of storage devices in one cluster while data stored in buffers of storage devices in another cluster is transferred to the storage devices' permanent storages. When available buffer capacity in a cluster falls below a specified threshold, data write commands are no longer sent the cluster and the storage devices in the cluster transfer data stored in their buffers to their permanent storages. While the data is being transferred, data write commands are transmitted to other clusters. When the data transfer is complete, the storage devices in the cluster can be scheduled to receive data write commands again. A cluster can be selected for performing a given data write request by matching the attributes of the cluster to parameters of the data write request.
COMPUTER SYSTEM AND METHOD FOR REDUNDANTIZING NONVOLATILE MEMORY
A computer system including: a first computer including a first processor and a first nonvolatile memory; and a second computer including a second processor and a second nonvolatile memory in which the second computer is connected to the first computer. The first computer includes a redundant hardware that, on receiving a write command from the first processor, writes the write data of the write command both into the first nonvolatile memory and into the second computer.
STORAGE SYSTEM AND CONTROL METHOD THEREOF
With omission of a duplication process of compressed data, a cache access frequency is reduced to improve throughput. A storage system includes first and second control units and a storage drive. Upon receiving a data write command, the first control unit stores data to be subjected to the write command in a first cache area of the first control unit, and stores the data in a second cache area of the second control unit to perform duplication, and upon completion of the duplication, the first control unit transmits a response indicating an end of write, performs a predetermined process on the data to be subjected to the write command, stores the data in a buffer area, reads the data stored in the buffer area, and transmits the read data to the storage drive.
Data storage device including multiple memory modules and circuitry to manage communication among the multiple memory modules
A memory system includes a plurality of volatile memory modules to temporarily store data in a distributed manner, a V storing place management unit included in each of the volatile memory modules, a plurality of nonvolatile memory modules to store the data stored in each of the volatile memory modules in a distributed manner, and a NV storing place management unit included in each of the nonvolatile memory modules. Each V storing place management unit and each NV storing place management unit communicate and determine the destination nonvolatile memory module for each volatile memory module. The data is transmitted to the determined destination nonvolatile memory module and stored in the destination nonvolatile memory module.
Cache memory data management using relative access frequency
A caching system and methodology for data in a memory for faster access to commonly-used data by other applications and computer devices on a network. The memory can include a solid-state drive (SSD) array for the cache memory that has read-bias, in addition to a magnetic hard drive array. The system uses a logical set of slots to hold identifiers for specific groups of data that can be placed into cache memory and each identifier has a usage attribute that changes based upon the usage of the specific group of data and causes the identifier to move within the set of slots and potentially into cache memory.