G06F2212/281

System and method for protecting GPU memory instructions against faults

A system and method for protecting memory instructions against faults are described. The system and method include converting the slave instructions to dummy operations, modifying memory arbiter to issue up to N master and N slave global/shared memory instructions per cycle, sending master memory requests to memory system, using slave requests for error checking, entering master requests to the GM/LM FIFO, storing slave requests in a register, and comparing the entered master requests with the stored slave requests.

CACHE PROGRAM OPERATION OF THREE-DIMENSIONAL MEMORY DEVICE WITH STATIC RANDOM-ACCESS MEMORY

Embodiments of three-dimensional (3D) memory devices with a 3D NAND memory array having a plurality of pages, an on-die cache coupled to the memory array on a same chip and configured to cache a plurality of batches of program data between a host and the memory array, the on-die cache having SRAM cells, and a controller coupled to the on-die cache on the same chip. The controller is configured to check a status of an (N2).sup.th batch of program data, N being an integer equal to or greater than 2, program an (N1).sup.th batch of program data into respective pages in the 3D NAND memory array, and cache an N.sup.th batch of program data in respective space in the on-die cache as a backup copy of the N.sup.th batch of program data.

DATA UPDATING TECHNOLOGY
20200356270 · 2020-11-12 · ·

A storage system includes a management node and a plurality of storage nodes forming a redundant array of independent disks (RAID). When the management node determines that not all data in an entire stripe is updated based on a received write request, the management node sends update data chunk obtained from to-be-written data to corresponding storage node. The storage node do not directly update, based on the received update data chunks, data block stored in storage device of the storage node, but store the update data chunk into non-volatile memories (NVM) cache of the storage node and send the update data chunk to another storage node to backup. According to the data updating method, write amplification problems caused in a stripe update process can be reduced, thereby improving update performance of the storage system.

Power saving mechanisms for a dynamic mirror service policy

Described is storage system and method for reducing power consumption. The storage system has first and second physical disks configured to provide mirroring. The first physical disk is placed into a power-saving mode of operation, while the second physical disk is in an active mode of operation responding to read and write requests. The first physical disk transitions from the power-saving mode of operation to an active mode of operation for destaging writes pending from cache to the first physical disk, while the second physical disk responds to read and write requests. The second physical disk transitions from the active mode of operation to the power-saving mode of operation, while the first physical disk responds to read and write requests.

MAGNETIC DISK DEVICE AND METHOD
20200294540 · 2020-09-17 ·

According to one embodiment, a magnetic disk device includes a magnetic disk, a first magnetic head, a second magnetic head, a buffer memory, and a control circuit. The magnetic disk includes a plurality of first storage regions. The control circuit controls the first magnetic head to read first data from a second storage region of the first storage regions to the buffer memory. The control circuit controls the second magnetic head to write second data to a third storage region concurrently with the reading of the first data. The third region is of the first storage regions, different from the second storage region. The second data corresponds to the first data stored in the buffer memory.

Method and apparatus for caching data in an solid state disk (SSD) of a hybrid drive that includes the SSD and a hard disk drive (HDD)
10761737 · 2020-09-01 · ·

A system includes a read/write module and a caching module. The read/write module is configured to access a first portion of a recording surface of a rotating storage device. Data is stored on the first portion of the recording surface of the rotating storage device at a first density. The caching module is configured to cache data on a second portion of the recording surface of the rotating storage device at a second density. The second portion of the recording surface of the rotating storage device is separate from the first portion of the recording surface of the rotating storage device. The second density is less than the first density.

Information processing device and method for managing file

A media drive is loaded with a ROM medium on which a plurality of files constituting an application are recorded. A readout block reads out data from the media drive. A memory temporarily stores the read-out data. A providing block provides the data stored in the memory to game software. A recording processing block records the data stored in the memory in an auxiliary storage device.

THREAD PREFETCH MECHANISM

An apparatus to facilitate data prefetching is disclosed. The apparatus includes a memory, one or more execution units (EUs) to execute a plurality of processing threads and prefetch logic to prefetch pages of data from the memory to assist in the execution of the plurality of processing threads.

Sensing device data caching
10740242 · 2020-08-11 · ·

Data is received from a sensing device of a plurality of sensing devices in communication with a device for storage in at least one memory of the device. A first cache memory or a second cache memory of the device is selected for caching the received data based at least in part on the sensing device sending the data. According to another aspect, data is received from a sensing device for storage in at least one memory of a device. It is determined whether to cache the received data based on at least one of the sensing device sending the data and information related to the received data. A cache memory is selected from among a plurality of cache memories of the device for caching the received data based at least in part on the sensing device sending the data.

Shingled magnetic recording storage system with reduced time to recover
10733097 · 2020-08-04 · ·

The disclosed technology provides a method that reduces time to recover in storage systems. In one implementation, the method comprises entering an idle status, determining if there is an incomplete band update operation, invalidating a media scratch pad (MSP) by clearing headers responsive to determining there is no incomplete band update operation, performing a power cycle, reading an MSP header, and determining if an MSP header is valid. If a rude power cycle occurs and the MSP header is determined to be valid, an MSP is examined, and restored if required. If a safe power cycle occurs, an MSP restore operation is not required, reducing time to recover.