Patent classifications
G06F2212/224
Dynamic premigration throttling for tiered storage
A dynamic premigration protocol is implemented in response to a secondary tier returning to an operational state and an amount of data associated with a premigration queue of a primary tier exceeding a first threshold. The dynamic premigration protocol can comprise at least a temporary premigration throttling level. An original premigration protocol is implemented in response to an amount of data associated with the premigration queue decreasing below the first threshold.
Dynamic cache management in hard drives
Technologies are provided for dynamically changing a size of a cache region of a storage device. A storage device controller writes data to the cache region of the storage device using a particular storage format. The storage device controller then migrates the cached data to a storage region of the device, where the data is written using a different storage format. A dynamic cache manager monitors input and output activity for the storage device and dynamically adjusts a size of the cache region to adapt to changes in the input and/or output activity. The dynamic cache manager can also adjust a size of the storage region. The storage device controller can automatically detect that the storage device has dynamic cache support and configure the storage device by creating the cache region and the storage region on the device.
Sense flags in a memory device
Memory devices might be configured to perform methods including reading a first page of memory cells and flag data wherein the flag data indicates whether a second page of memory cells adjacent to the first page is programmed, and determining from the flag data whether to re-read the first page of memory cells with an adjusted read voltage; performing a sense operation on memory cells coupled to first data lines of a first array of memory cells and memory cells coupled to data lines of a second array of memory cells, and determining a program indication of memory cells coupled to second data lines from the sense operation performed on the memory cells coupled to the data lines of the second array of memory cells; and/or programming memory cells coupled to first data lines in a first array of memory cells, and programming memory cells coupled to second data lines in the first array of memory cells while programming memory cells coupled to data lines in a second array of memory cells with flag data indicative of the memory cells coupled to the second data lines in the first array of memory cells being programmed.
System and method for a hit-based ratio write cache operation mode in a storage controller
An object is to suppress a process of evicting cached data so as to improve a throughput of an entire system. A storage controller includes an access request section and an operation management section. The access request section requests access to a first storage and to a second storage that is higher in response speed than the first storage, the second storage storing part of data stored in the first storage. The operation management section manages, based on a usage state of the second storage, whether or not to transfer from the first storage to the second storage data targeted for access but not stored in the second storage.
EFFICIENT CLONING OF LOGICAL STORAGE DEVICES
A no-copy clone of a logical storage unit is created. A define process is initiated for defining a target logical storage unit as the clone before activation of the target logical storage unit. By initiating the define process before activating the logical storage unit, there is a greater likelihood that, when a write operation is received for a data portion on the source logical storage unit or target logical storage unit after activation of the target LSU, the data portion will already be defined and not need to be defined when performing the write operation. When a write operation is received at the source logical storage unit, if the target logical storage unit is not active yet, the data of the write operation may be written to an allocated physical location for the data portion shared between the source and target logical storage units without updating any clone metadata.
Provisioning Virtual Machines with a Single Identity and Cache Virtual Disk
A virtual disk is provided to a computing environment. The virtual disk includes identity information to enable identification of a virtual machine within the computing environment. A size of the virtual disk is increased within the computing environment to enable the virtual disk to act as a storage for the identity information and as a cache of other system data to operate the virtual machine. The virtual machine is booted within the computing environment. The virtual machine is configured to at least access the virtual disk that includes both identity information and caches other system data to operate the virtual machine. Related apparatus, systems, techniques and articles are also described.
SYSTEM AND METHOD OF DISAGGREGATED TIERED CACHE FOR CLOUD CONTENT STORAGE
A content provider system includes: a repository to store a catalog of content; a storage device pool to load content from among the catalog of content from the repository into one or more storage devices of the storage device pool; a first hosted device communicably connected to the storage device pool, and to execute the content stored in the storage device pool to provide the content to a first user device; a second hosted device communicably connected to the storage device pool, and to execute the content stored in the storage device pool to provide the content to a second user device; and one or more processing circuits to identify an available storage device from among the one or more storage devices of the storage device pool for serving a requested content to a requesting device from among the first and second hosted devices.
Accelerated computer system and method for writing data into discrete pages
The instant disclosure provides an accelerated computer system and an accelerated method for writing data into discrete pages. The accelerated method includes executing write commands, with each write command including write data and a write address such that the write address corresponds to a write page of the first pages in a sector of a hard drive, identifying whether the write pages are successive according to the write addresses, acquiring stored data by reading the sector according to the write addresses if the write pages are discrete, writing the data stored in the first pages into the second pages of a memory, writing write data bit by bit into the second pages according to the write addresses, and writing the data stored in the second pages into the first pages.
Distributed safe data commit in a data storage system
In one embodiment, a safe data commit process manages the allocation of task control blocks (TCBs) as a function of the type of task control block (TCB) to be allocated for destaging and as a function of the identity of the RAID storage rank to which the data is being destaged. For example, the allocation of background TCBs is prioritized over the allocation of foreground TCBs for destage operations. In addition, the number of background TCBs allocated to any one RAID storage rank is limited. Once the limit of background TCBs for a particular RAID storage rank is reached, the distributed safe data commit logic switches to allocating foreground TCBs. Further, the number of foreground TCBs allocated to any one RAID storage rank is also limited. Other features and aspects may be realized, depending upon the particular application.
Shingled magnetic recording drive that updates media cache data in-place
When a shingled magnetic recording (SMR) hard disk drive (HDD) receives a write command that references one or more target logical block addresses (LBAs) and determines that one or more target LBAs are included in a range of LBAs for which data are stored in a memory of the drive, additional data are written to the media cache of the SMR HDD along with the write data during the same disk access. The additional data include data that are stored in the volatile memory and are associated with one or more LBAs that are adjacent in LBA space to the target LBAs. The one or more LBAs that are adjacent in LBA space to the target LBAs may include a first group of LBAs that is adjacent to and follows the target LBAs and a second group of LBA that is adjacent to and precedes the target LBAs.