Patent classifications
G06F12/122
PROCESS RUNNING METHOD AND APPARATUS
A process running method and apparatus is disclosed. The method is: selecting a code page from a candidate process, storing only a code page of the candidate process into a swap partition, and releasing memory space occupied by the code page; updating a physical address that is of the code page and that is stored into a page entry; and when it is determined that the candidate process to which the code page belongs needs to be run, if the code page needs to be executed, directly executing the code page in the swap partition.
PROCESS RUNNING METHOD AND APPARATUS
A process running method and apparatus is disclosed. The method is: selecting a code page from a candidate process, storing only a code page of the candidate process into a swap partition, and releasing memory space occupied by the code page; updating a physical address that is of the code page and that is stored into a page entry; and when it is determined that the candidate process to which the code page belongs needs to be run, if the code page needs to be executed, directly executing the code page in the swap partition.
READ CACHE MANAGEMENT
Implementations disclosed herein provide for a storage system including an on-disk read cache and a variety of read cache management techniques. According to one implementation, a storage device controller time-sequentially reads a series of non-contiguous data blocks storing a data sequence in a read cache of a magnetic disk, the data sequence identified by a requested sequence of logical block addresses (LBAs). The controller determines that read requests for the data sequence satisfy at least one predetermined access frequency criterion and, responsive to the determination, the controller re-writes data of the data sequence to a series of contiguous data blocks in the read cache.
Managing processing tasks in storage systems
A method is used in managing processing tasks in storage systems. A set of tasks is received for processing. A type of the set of tasks is identified. Based on the type of the set of tasks, a determination is made as to whether to add data objects associated with the set of tasks to a used object list managed in conjunction of a cache of the system for accessing the data objects associated with the set of tasks. The cache is configured to store the data objects of the storage system. A portion of a memory of the storage system is reserved as the cache.
SCALED SET DUELING FOR CACHE REPLACEMENT POLICIES
A processing system includes a cache that includes a cache lines that are partitioned into a first subset of the cache lines and a second subsets of the cache lines. The processing system also includes one or more counters that are associated with the second subsets of the cache lines. The processing system further includes a processor configured to modify the one or more counters in response to a cache hit or a cache miss associated with the second subsets. The one or more counters are modified by an amount determined by one or more characteristics of a memory access request that generated the cache hit or the cache miss.
ADAPTIVE METHOD FOR SELECTING A CACHE LINE REPLACEMENT ALGORITHM IN A DIRECT-MAPPED CACHE
A method of managing a direct-mapped cache is provided. The method includes a direct-mapped cache receiving memory references indexed to a particular cache line, using a first cache line replacement algorithm to select a main memory block as a candidate for storage in the cache line in response to each memory reference, and using a second cache line replacement algorithm to select a main memory block as a candidate for storage in the cache line in response to each memory reference. The method further includes identifying, over a plurality of most recently received memory references, which one of the algorithms has selected a main memory block that matches a next memory reference a greater number of times, and storing a block of main memory in the cache line, wherein the block of main memory stored in the cache line is the main memory block selected by the identified algorithm.
ADAPTIVE METHOD FOR SELECTING A CACHE LINE REPLACEMENT ALGORITHM IN A DIRECT-MAPPED CACHE
A method of managing a direct-mapped cache is provided. The method includes a direct-mapped cache receiving memory references indexed to a particular cache line, using a first cache line replacement algorithm to select a main memory block as a candidate for storage in the cache line in response to each memory reference, and using a second cache line replacement algorithm to select a main memory block as a candidate for storage in the cache line in response to each memory reference. The method further includes identifying, over a plurality of most recently received memory references, which one of the algorithms has selected a main memory block that matches a next memory reference a greater number of times, and storing a block of main memory in the cache line, wherein the block of main memory stored in the cache line is the main memory block selected by the identified algorithm.
SYSTEM AND METHOD FOR MANAGING STORAGE SPACE
Disclosed herein is a technique for managing storage space in a user device. Users are provided with options to manage storage space usage in an organized and efficient manner. The options can include recommendations to the user regarding automatically and/or manually purging data from the user device to free up a particular amount of storage space that is needed to carry out a particular task.
INVOKING DEMOTE THREADS ON PROCESSORS TO DEMOTE TRACKS INDICATED IN DEMOTE READY LISTS FROM A CACHE WHEN A NUMBER OF FREE CACHE SEGMENTS IN THE CACHE IS BELOW A FREE CACHE SEGMENT THRESHOLD
Provided are a computer program product, system, and method for invoking demote threads on processors to demote tracks from a cache. A plurality of demote ready lists indicate tracks eligible to demote from the cache. In response to determining that a number of free cache segments in the cache is below a free cache segment threshold, a determination is made of a number of demote threads to invoke on processors based on the number of free cache segments and the free cache segment threshold. The determined number of demote threads are invoked to demote tracks in the cache indicated in the demote ready lists, wherein each invoked demote thread processes one of the demote ready lists to select tracks to demote from the cache to free cache segments in the cache.
INVOKING DEMOTE THREADS ON PROCESSORS TO DEMOTE TRACKS INDICATED IN DEMOTE READY LISTS FROM A CACHE WHEN A NUMBER OF FREE CACHE SEGMENTS IN THE CACHE IS BELOW A FREE CACHE SEGMENT THRESHOLD
Provided are a computer program product, system, and method for invoking demote threads on processors to demote tracks from a cache. A plurality of demote ready lists indicate tracks eligible to demote from the cache. In response to determining that a number of free cache segments in the cache is below a free cache segment threshold, a determination is made of a number of demote threads to invoke on processors based on the number of free cache segments and the free cache segment threshold. The determined number of demote threads are invoked to demote tracks in the cache indicated in the demote ready lists, wherein each invoked demote thread processes one of the demote ready lists to select tracks to demote from the cache to free cache segments in the cache.