G06F11/16

PEER-TO-PEER REDUNDANT ARRAY OF INDEPENDENT DISKS (RAID) LACKING A RAID CONTROLLER
20170286237 · 2017-10-05 ·

Programmable disk drives are configured within a peer-to-peer redundant array of independent disks (RAID) that lacks a RAID controller. A spare programmable disk drive not part of the RAID can monitor a given drive of the RAID. If the monitored drive fails, the spare drive can replace the failed drive within the RAID such that it becomes part of the RAID, and reconstruct the data on the programmable disk drive that has failed by communicating with the other drives of the RAID. For each data block to be reconstructed, each such other drive may receive just one piece of data and sends just one piece of data. Further, the spare drive may receive just one piece of data for each data block to be reconstructed, which is the data block as reconstructed from one of the other drives of the RAID.

METHODS FOR FILESYSTEM METADATA CACHING TO IMPROVE FAILOVER PERFORMANCE AND DEVICES THEREOF

A method, non-transitory computer readable medium, and device that assists with caching filesystem metadata to a partner non-volatile random-access memory (NVRAM) includes caching metadata related to an incoming data modifying operation generated by a client computing device to at least one storage controller device in a cluster. A service interruption event that makes a data block present in the storage device of a hosting storage node inaccessible to the client computing device is determined for during the caching. The requested metadata block from the at least one NVRAM is retrieved when the service interruption event is determined. The cache is warmed using the retrieved metadata block from the at least one NVRAM.

Memory erasure information in cache lines

Example implementations relate to storing memory erasure information in memory devices on a memory module. In example implementations, a memory location associated with an error in a first cache line may be identified. The first cache line may include data read from the memory location, and the memory location may be in a first memory device of a plurality of memory devices on a memory module. A device number corresponding to the first memory device may be written to one of the plurality of memory devices. When the memory location is read for a second cache line, the device number corresponding to the first memory device may be retrieved. The second cache line may include the retrieved device number and data read from the memory location.

Storage control device and storage system

A control module that manages a segment to which data is written implements write processing and resynchronization processing using a bitmap managed for each LUN. In other words, the control module stores the bitmap for the managed LUN in a bitmap storage unit. A mirror LUN control unit sets a corresponding portion of the bitmap to 1, controls data write to a target segment and a mirror segment, and resets the bitmap to 0 when the data write to both of the segments is complete. A resynchronization control unit refers to the bitmap storage unit to perform the resynchronization processing.

STORAGE CONTROL OF STORAGE MEDIA SUBJECT TO WRITE AMPLIFICATION EFFECTS
20170277608 · 2017-09-28 ·

According to one aspect of the present invention, there is provided a method for performing storage control. Member storage media and a hot spare storage medium are identified in a storage system. The member storage media are members of a storage medium array, and the hot spare storage medium is for joining in the storage medium array when a member storage medium fails. Data on a member storage medium having a write amplification effect is migrated to the hot spare storage medium. In the member storage medium having a write amplification effect, an erase operation is performed on a storage medium where the migrated data is located. Embodiments of the present invention can alleviate adverse impact caused by a write amplification effect.

CLOSING BLOCK FAMILY BASED ON SOFT AND HARD CLOSURE CRITERIA
20220050758 · 2022-02-17 ·

A system includes a memory device and a processing device, operatively coupled to the memory device. The processing device is to perform operations, including initializing a block family associated with the memory device and initializing a timer associated with the block family. Responsive to beginning to program a block residing on the memory device, the processing device associates the block with the block family. In response to the timer reaching a soft closure value, the processing device performs a soft closure of the block family; continues to program data to the block; and performs a hard closure of the block family in response to one of the timer reaching a hard closure value or the block family satisfying a hard closure criteria.

STORAGE SYSTEM, COMPUTER SYSTEM, AND CONTROL METHOD FOR STORAGE SYSTEM
20170277443 · 2017-09-28 ·

Saving a capacity of an on-premises storage apparatus, a high access performance of the on-premises storage apparatus, and resuming an operation quickly and accurately by using data on a cloud when a resource in the on-premises fails are achieved.

A processor provides a first volume which is a virtual volume, and configures a copy pair of the first volume and a second volume provided from another storage system. Write data to the first volume is transferred to the second volume via a network, based on the copy pair. The processor writes to the memory a part of data written to the second volume, and writes to the storage device the data written to the memory.

FAULT-TOLERANCE PATTERN AND SWITCHING PROTOCOL FOR MULTIPLE HOT AND COLD STANDBY REDUNDANCIES
20170277607 · 2017-09-28 ·

A method for fault tolerant controller readiness. Executing functions by a first controller operating in a primary status mode. Operating in a hot standby status mode by a second controller and mirroring the first controller by executing functions to operate as a redundant controller. Operating in a cold standby status mode by at least one backup controller under normal operating conditions. The second controller is reconfigured while operating under normal operating conditions from the hot standby status mode to the primary standby status mode if a failure occurs in the first controller. Reconfiguring the at least one backup controller operating under normal operating conditions from cold standby status mode to hot standby status mode to operate as a redundant controller in response to the reconfiguring the second controller from the hot standby status mode to the primary status mode.

Audit file generation in a dispersed storage network
11243839 · 2022-02-08 · ·

A method for execution by a dispersed storage and task (DST) processing unit includes obtaining audit records for an audit object and determining when the audit object is complete. When the audit object is complete, aggregating the audit records of the audit object within the audit object by generating the audit object to include the audit records; generating identifier (ID) information and generating integrity information. Fields of the audit object are populated with the audit records, the ID information, and the integrity information and a name of the audit object is determined for storage of the audit object and the name of the audit object in a dispersed storage network (DSN).

Self healing cluster of a content management system

Systems and methods herein provide for a clustered content management comprising at least two computing nodes. A first node comprises an instance of the content repository. The first computing node may perform content management operations on its instance of the content repository. Changes to the instance of the content repository of the first computing node are synchronized with the content repository by way of a second computing node. The second computing node is communicatively coupled to the first computing node through a network and is operable to synchronize the change with the content repository. The second computing node also determines that synchronization of the change is blocked due to an error. The second computing node identifies the error, determines that the error is correctable, and corrects the error to synchronize the change with the content repository.