Patent classifications
G06F2212/224
Method and storage array for processing a write data request
According to a write data request processing method and a storage array provided in the embodiments of the present invention, a controller is connected to a cache device via a switching device, an input/output manager is connected to the controller via the switching device, and the input/output manager is connected to a cache device via the switching device. The controller obtains a cache address from the cache device for to-be-written data according to the write data request, the controller sends an identifier of the cache device and the cache address to the input/output manager via the switching device, and the input/output manager writes the to-be-written data to the cache address via the switching device.
Gateway for cloud-based secure storage
The systems and methods disclosed herein transparently provide an improved scalable cloud-based dynamically adjustable or configurable storage volume. In one aspect, a gateway provides a dynamically or configurably adjustable storage volume, including a local cache. The storage volume may be transparently adjusted for the amount of data that needs to be stored using available local or cloud-based storage. The gateway may use caching techniques and block clustering to provide gains in access latency compared to existing gateway systems, while providing scalable off-premises storage.
METHODS AND SYSTEMS FOR DIRECTLY MAPPING A BACKEND BLOCK ADDRESS INTO A PHYSICAL ADDRESS OF A CACHING DEVICE
A storage device made up of multiple storage media is configured such that one such media serves as a cache for data stored on another of such media. The device includes a controller configured to manage the cache by consolidating information concerning obsolete data stored in the cache with information concerning data no longer desired to be stored in the cache, and erase segments of the cache containing one or more of the blocks of obsolete data and the blocks of data that are no longer desired to be stored in the cache to produce reclaimed segments of the cache.
Tape-managed partition support for effective workload allocation and space management
In one embodiment, a system includes a disk cache that includes a plurality of hard disk drives (HDDs) and a controller. The controller is configured to create one or more tape-managed partitions in the disk cache, each of the one or more tape-managed partitions being configured to store data that is subject to hierarchical storage management (HSM). The controller is also configured to create a premigration queue configured to service premigration data for all of the one or more tape-managed partitions. Moreover, the controller is configured to receive a premigration delay value for a first tape-managed partition, the premigration delay value defining a time period that elapses prior to queuing the premigration data for the first tape-managed partition to the premigration queue. The premigration delay value is based on a volume creation time. Other systems, methods, and computer program products are described in accordance with more embodiments.
Persistent caching for operating a persistent caching system
A persistent caching system is provided. The persistent caching system includes a storage system having a caching server for storing data, and a client for accessing the data through a network. The caching server is configured to store the data in a number of virtual memory blocks. The virtual memory blocks refer to an associated memory-mapped file in a file system of the caching server. The caching server is configured to export addresses of the virtual memory blocks to the client. The client is configured to access at least some of the virtual memory blocks through RDMA using the exported addresses. The caching server is configured to page virtual memory blocks being accessed by one or more clients through RDMA to and/or from the memory-mapped files associated with the accessed virtual memory blocks.
CONSIDERING A DENSITY OF TRACKS TO DESTAGE IN GROUPS OF TRACKS TO SELECT GROUPS OF TRACKS TO DESTAGE
Provided are a computer program product, system, and method for considering a density of tracks to destage in groups of tracks to select groups of tracks to destage. Groups of tracks in the cache are scanned to determine whether they are ready to destage. A determination is made as to whether the tracks in one of the groups are ready to destage in response to scanning the tracks in the group. A density for the group is increased in response to determining that the group is not ready to destage. The group is destaged in response to determining that the density of the group exceeds a density threshold.
Multi-Speed Data Storage Device with Media Cache for High Speed Writes
Apparatus and method for managing data transfers in a data storage device with rotational media that can be rotated at different speeds. In some embodiments, a non-volatile main memory is formed on a rotatable medium accessed by a moveable data transducer. A media cache provides a non-volatile data storage area. A control circuit directs writes to the main memory as the medium is rotated at a first speed and directs reads from the main memory as the medium is rotated at a higher, second speed. Writes during the rotation of the medium at the second speed are directed to the media cache instead of to the main memory so that no data are written to the main memory at the second speed. The media cache may also be located on the medium or may be formed from solid-state semiconductor memory.
MASS STORAGE CACHE IN NON VOLATILE LEVEL OF MULTI-LEVEL SYSTEM MEMORY
An apparatus is described. The apparatus includes a memory controller comprising logic circuitry to implement a mass storage cache in a non volatile region of a system memory. The non volatile region of the system memory is to support execution of program code directly out of the non volatile region system memory.
Data storage management in a memory device
The disclosure is related to systems and methods of managing data storage in a memory device. In a particular embodiment, a method is disclosed that includes receiving, in a data storage device, at least one data packet that has a size that is different from an allocated storage capacity of at least one physical destination location on a data storage medium in the data storage device for the at least one data packet. The method also includes storing the at least one received data packet in a non-volatile cache memory prior to transferring the at least one received data packet to the at least one physical destination location.
Flushing host cache data before rebuilding degraded redundant virtual disk
A storage management method includes receiving a degrade signal indicating a degraded state of a virtual disk associated with a host system. Rebuild-flush operations may be performed. The operations may include writing, to the virtual disk and also to a hot spare drive (HSP) associated with the virtual disk, valid-modified data, stored in a host storage cache, associated with the virtual disk. In contrast, valid-unmodified storage cache data associated with the virtual disk, may be written to the HSP only. After the rebuild-flush completes, the virtual disk may be rebuilt. During rebuild, however, any cached-LBA may be skipped where traditional RAID rebuild operations are performed for un-cached LBAs only.