Patent classifications
G06F2212/65
IMPLEMENTING MAPPING DATA STRUCTURES TO MINIMIZE SEQUENTIALLY WRITTEN DATA ACCESSES
A system includes a memory device, and a processing device, operatively coupled to the memory device, to perform operations including receiving a request to sequentially write data to a block of a memory device, in response to receiving the request, writing the data to the block to obtain sequentially written data, initiating accumulation of logical-to-physical (L2P) mapping data corresponding to the sequentially written data, determining that a criterion for terminating the accumulation of the L2P mapping data is satisfied, in response to determining that the criterion is satisfied, terminating the accumulation of the L2P mapping data to obtain accumulated L2P mapping data, and updating an L2P mapping data structure based on the accumulated L2P mapping data.
APPARATUS AND METHOD FOR PERFORMING ADDRESS TRANSLATION
An apparatus, system, and method for address translation are provided. Physical address information corresponding to virtual addresses is prefetched and stored, where at least some sequences of the virtual addresses are in a predefined order. The physical address information is prefetched based on identification information provided by a data processing activity, comprising at least a segment identifier and a portion of a virtual address to be translated. The storage has segments of entries, wherein each segment stores physical address information which corresponds to virtual addresses in a predefined order. This predefined order means that it is not necessary to store virtual addresses in the storage. Storage capacity and response speed are therefore gained.
ADJUSTING ACTIVE CACHE SIZE BASED ON CACHE USAGE
Provided are a computer program product, system, and method for adjusting active cache size based on cache usage. An active cache in at least one memory device caches tracks in a storage during computer system operations. An inactive cache in the at least one memory device is not available to cache tracks in the storage during the computer system operations. During caching operations in the active cache, information is gathered on cache hits to the active cache and cache hits that would occur if the inactive cache was available to cache data during the computer system operations. The gathered information is used to determine whether to configure a portion of the inactive cache as part of the active cache for use during the computer system operations.
Retaining cache entries of a processor core during a powered-down state
A processor core associated with a first cache initiates entry into a powered-down state. In response, information representing a set of entries of the first cache are stored in a retention region that receives a retention voltage while the processor core is in a powered-down state. Information indicating one or more invalidated entries of the set of entries is also stored in the retention region. In response to the processor core initiating exit from the powered-down state, entries of the first cache are restored using the stored information representing the entries and the stored information indicating the at least one invalidated entry.
APPLICATION PROGRAMMING INTERFACE TO DISASSOCIATE A VIRTUAL ADDRESS
Apparatuses, systems, and techniques to manage memory arrays. In at least one embodiment an application programming interface (API) is performed to disassociate a virtual address indicated by the API from a corresponding physical address.
N-WAY ACTIVE-ACTIVE STORAGE CONFIGURATION TECHNIQUES
A stretched volume may be configured from N volumes of N data storage systems configured as a cluster. N may be an integer greater than two. Each of the N volumes may be included in a different one of the N data storage systems. The N volumes may be exposed to a host as a logical volume having a unique identifier over a plurality of paths from the N data storage systems. The N volumes may be configured for multidirectional synchronous replication. At a first system of the cluster, a write operation may be received that writes to a target logical address of the stretched volume. Servicing the first write may include synchronously replicating the first write to every other one of the N data storage systems of the cluster. Also described are techniques for handling lock contention and avoiding deadlock in connection with processing writes to the stretched volume.
N-way active-active storage configuration techniques
A stretched volume may be configured from N volumes of N data storage systems configured as a cluster. N may be an integer greater than two. Each of the N volumes may be included in a different one of the N data storage systems. The N volumes may be exposed to a host as a logical volume having a unique identifier over a plurality of paths from the N data storage systems. The N volumes may be configured for multidirectional synchronous replication. At a first system of the cluster, a write operation may be received that writes to a target logical address of the stretched volume. Servicing the first write may include synchronously replicating the first write to every other one of the N data storage systems of the cluster. Also described are techniques for handling lock contention and avoiding deadlock in connection with processing writes to the stretched volume.
SERDES link training
Aspects of the embodiments are directed to systems and methods for performing link training using stored and retrieved equalization parameters obtained from a previous equalization procedure. As part of a link training sequence, links interconnecting an upstream port with a downstream port and with any intervening retimers, can undergo an equalization procedure. The equalization parameter values from each system component, including the upstream port, downstream port, and retimer(s) can be stored in a nonvolatile memory. During a subsequent link training process, the equalization parameter values stored in the nonvolatile memory can be written to registers associated with the upstream port, downstream port, and retimer(s) to be used to operate the interconnecting links. The equalization parameter values can be used instead of performing a new equalization procedure or can be used as a starting point to reduce latency associated with equalization procedures.
Complex I/O value prediction for multiple values with physical or virtual addresses
An apparatus, and corresponding method, for input/output (I/O) value determination, generates an I/O instruction for an I/O device, the I/O device including a state machine with state transition logic. The apparatus comprises a controller that includes a simplified state machine with a reduced version of the state transition logic of the state machine of the I/O device. The controller is configured to improve instruction execution performance of a processor core by employing the simplified state machine to predict at least one state value of at least one I/O device true state value to be affected by the I/O instruction at the I/O device.
Implementing mapping data structures to minimize sequentially written data accesses
A system includes a memory device, and a processing device, operatively coupled to the memory device, to perform operations including receiving a request to sequentially write data to a block of a memory device, in response to receiving the request, writing the data to the block to obtain sequentially written data, initiating accumulation of logical-to-physical (L2P) mapping data corresponding to the sequentially written data, determining that a criterion for terminating the accumulation of the L2P mapping data is satisfied, in response to determining that the criterion is satisfied, terminating the accumulation of the L2P mapping data to obtain accumulated L2P mapping data, and updating an L2P mapping data structure based on the accumulated L2P mapping data.