G06F11/1489

METHOD, SYSTEM AND COMPUTER PROGRAM PRODUCT FOR BACKUP ASSIGNMENT OF A MOBILE ELECTRONICS DEVICE

A method, system and computer program product for backup assignment of a mobile electronics device is disclosed. In response to one or more device assignment requests received by a controller apparatus when a server is operating in an off-line mode, one or more device assignment records are created within the controller apparatus, and which indicate assignment of the at least one mobile electronics device to a respective one or more users. When the server is operating in an off-line mode such that communication between the controller apparatus and the server is off-line, the one or more device assignment records, in relation to the at least one mobile electronics device, are stored in the controller apparatus. In response to the server re-entering the on-line mode, the server is updated with the one or more device assignment records.

LOW LATENCY PARITY FOR A MEMORY DEVICE

Apparatuses, systems, and methods for low latency parity for a memory device include a controller configured to accumulate, in a memory buffer, combined parity data for a plurality of regions of memory of a memory device in response to write operations for the plurality of regions of memory. The controller is configured to perform a recovery operation for a region of memory in response to determining that a latency setting for the region satisfies a latency threshold. The controller is configured to service a read request for data from the region based on a recovery operation to satisfy the latency setting.

Read retry scratch space

Devices and techniques to recover data from a memory device using a custom Read Retry feature are disclosed herein. A memory device can receive a first read request, read data from the memory array corresponding to the read request, and determine if the read data corresponding to the first read request includes a detectable error. In response to a detected error in the received data corresponding to the first read request, the memory device can recover data corresponding to the first read request using one of a set of read retry features, and load the one of the set of read retry features used to recover data corresponding to the first read request as a custom read retry feature in the memory device for a second read request subsequent to the first read request.

METHOD FOR DETECTING AND RECOVERY FROM SOFT ERRORS IN A COMPUTING DEVICE

A method for detecting and recovery from a soft error in a computing device is provided. In examples discussed herein, the method can be performed to detect soft errors that may occur during execution of a predefined critical instruction(s) and/or has been propagated in the computing device prior to the execution of the predefined critical instruction(s). Specifically, a software compiler may be used to embed an error detector block(s) after the predefined critical instruction(s). In this regard, the error detector block(s) can be executed after the predefined critical instruction(s) to detect the soft error. Accordingly, it may be possible to invoke a diagnosis routine to determine severity of the detected soft error and take appropriate action against the detected soft error. As such, it may be possible to protect the execution of the predefined critical instruction(s) concurrent to eliminating vulnerable voting intervals and reducing soft error detection overhead.

READ RETRY SCRATCH SPACE
20230205628 · 2023-06-29 ·

Devices and techniques to recover data from a memory device are disclosed, including recovering data corresponding to a detected error in data stored on a memory array corresponding to a memory operation using one of a set of read offset values and loading the one of the set of read offset values used to recover data corresponding to the detected error in a temporary storage of the memory array as a custom read offset value for a subsequent memory operation. The temporary storage of the memory array can include a scratch space of the memory array separate from read retry offset registers of the memory device.

USING BITMAPS TO VALIDATE DATA
20230205642 · 2023-06-29 ·

An incremental backup is made of a source volume on a production host to on premise secondary storage. The incremental backup includes a bitmap specifying locations of blocks that have changed since a last backup of the source volume, and data of the changed blocks. First checksums are calculated from data of the changed blocks. The incremental backup, including the changed blocks and the bitmap, are moved to cloud storage. The changed blocks are merged, at the cloud storage, with the last backup of the source volume to generate a synthetic full backup of the source volume. The bitmap is consulted to identify locations of the changed blocks. Second checksums are calculated from data written to the locations on the synthetic full backup. The first and second checksums are compared. If the first and second checksums do not match, an indication is generated that the synthetic full backup is corrupt.

Recovering dynamic system data

Aspects of the invention include creating a collection of control blocks associated with one or more catalogs, each control block including an eye catcher value and a plurality of pointers, creating a collection of recovery control blocks based the collection of control blocks, each recovery control block including a corresponding copy of the eye catcher value and the plurality of pointers, storing a control block memory location for each control block in the collection of control blocks and a recovery memory location for each recovery control block in the collection of recovery control blocks in a recovery table, analyzing the collection of control blocks to determine that a corruption exists, determining a corruption type, the corruption type comprises an infrastructure corruption and or a control block corruption, and initiating a remedial action for the collection of control blocks based at least in part on the corruption type.

RAID DATA STORAGE DEVICE AND DATA STORAGE SYSTEM INCLUDING THE SAME
20220058086 · 2022-02-24 ·

A data storage system includes a host configured to provide a read request; a plurality of storage devices constituting a redundant array of independent disks (RAID); and a RAID controller configured to a plurality of read commands in response to the read request, the read commands being provided to the plurality of storage devices according to a RAID setting, wherein one data storage device of the plurality of data storage devices includes a nonvolatile memory device; an error handling information storage circuit to store error handling information; and an error detection and correction circuit configured to detect an error in data output from the nonvolatile memory device according to a read command and to selectively correct the error according to the error handling information.

USING OVER PROVISIONING SPACE FOR SELECTIVELY STORING BLOCK PARITY

Methods and apparatus for storing parity bits in an available over provisioning (OP) space to recover data lost from an entire memory block. For example, a data storage device may receive data from a host device, write the data to a block, and generate a corresponding block parity. The device may then determine a bit error rate (BER) of the block and an average programming duration to write the data written to the block, calculate a probability of the block becoming defective based on the BER and the average programming duration, and comparing the probability of the block to a set of probabilities respectively corresponding to a set of worst-performing blocks in a NVM. Thereafter, the device may write the block parity to an available over provisioning (OP) space in the NVM responsive to the probability of the block being greater than any probability in the set of probabilities.

MANAGER ELECTION FOR ERASURE CODING GROUPS

To ensure that there is an elected manager among storage nodes of an erasure coding group (“ECG”), an ECG manager (“ECGM”) election process is periodically performed among available storage nodes that are configured with the software to perform the services of an ECGM. When a storage node is activated, an ECGM process of the storage node begins executing and is assigned a process identifier (“PID”). A storage node can utilize a service query framework to identify other available storage nodes and retrieve their ECGM PIDs. The storage node then selects a PID according to a criterion and elects the storage node corresponding to the selected PID to be the acting ECGM. This process is performed periodically, so even if the acting ECGM storage node fails, a new ECGM is eventually selected from the available storage nodes.