G06F2212/254

Data block transfer with extended read buffering

Various embodiments enable read buffering in connection with data block transfer on a memory device. For some embodiments, read buffering from a set of cache blocks is enabled during a period of wait time after data is copied (e.g., data is transferred, such as part of a compaction operation) from the set of cache blocks to a set of non-cache blocks. In various embodiments, after the wait time, data stored on the set of cache blocks is erased (e.g., the set of cache blocks is released) and read buffering from the set of cache blocks is disabled.

Storage aggregator controller with metadata computation control

This disclosure describes a storage aggregator controller with metadata computation control. The storage aggregator controller communicates, via a host interface, over a computer network with one or more remote hosts, and also communicates, via a storage device interface, with a plurality of local storage devices, which are separate from the remote host(s) and which have respective non-volatile memories. The storage aggregator controller manages the local storage devices for storage or retrieval of media objects. The storage aggregator controller also governs a selective computation, at aggregator control circuitry or at a storage device controller of one or more of the storage devices, of metadata that defines content characteristics of the media objects that are retrieved from the plurality of storage devices or that are received from the one or more hosts over the computer network for storage in the plurality of storage devices.

Storage edge controller with a metadata computational engine

Embodiments described herein provide improved methods and systems for generating metadata for media objects at a computational engine (such as an artificial intelligence engine) within the storage edge controller, and for storing and using such metadata, in data processing systems.

STORAGE DEVICE OPERATING DATA COUNTER SYSTEM
20230350796 · 2023-11-02 ·

A counter system includes a counter management subsystem, a storage subsystem having storage elements, and a non-volatile memory system storing a first counter including a first value field/first bitmap field combination for each storage element, and a second counter including a second value field/second bitmap field combination for each storage element. The counter management subsystem resets the first counter and, following each storage operation on a storage element, updates a bit in the first bitmap field for that storage element. When one of the first bitmap fields is filled, the counter management subsystem converts each first value field/first bitmap field combination to a respective first value, resets the second counter, updates the second value field for each storage element with the respective first value for each storage element and, following each storage operation on a storage element, updates a bit in the second bitmap field for that storage element.

MANAGING AND RANKING MEMORY RESOURCES

The present disclosure relates to systems, methods, and computer-readable media for tracking memory usage data on a memory controller system and providing a mechanism whereby one or multiple accessing agents (e.g., computing nodes, applications, virtual machines) can access memory usage data for a memory resource managed by a memory controller. Indeed, the systems described herein facilitate generation of and access to heatmaps having memory usage data thereon. The systems described herein describe features and functionality related to generating and maintaining the heatmaps as well as providing access to the heatmaps to a variety of accessing agents. This memory tracking and accessing is performed using low processing overhead while providing useful information to accessing agents in connection with memory resources managed by a memory controller.

MULTI-PLATFORM DATA STORAGE SYSTEM SUPPORTING PEER-TO-PEER SHARING OF CONTAINERS

A multi-platform data storage system that facilitates sharing of containers including one or more virtual storage resources. The multi-platform data storage system can, for example, include a storage interface configured to enable access to a plurality of storage platforms that use different storage access and/or management protocols, the plurality of storage platforms storing data objects in physical data storage; and a storage mobility and management layer providing virtual management of virtual storage resources corresponding to one or more data objects stored in the plurality of storage platforms, the storage mobility and management layer including at least a transfer module coupled to at least one network and configured to transfer at least one of the data objects. The transfer module can transfer the at least one of the data objects between the multi-platform data storage system and another data storage system.

SYSTEMS AND METHODS FOR DYNAMIC IN-MEMORY CACHING OF MAPPINGS INTO PARTITIONS
20220229780 · 2022-07-21 · ·

Systems and methods for routing requests to dynamically cached mappings are disclosed. A system may comprise a memory storing instructions and at least one processor configured to execute instructions to perform operations including: receiving an access request to access first item data from a user device, the access request having an access key; based on the access key, routing the request to a server having an in-memory cache with multiple partitions associated with different access key types and storing the first item data and second item data; receiving, from the server, the first item data; and transmitting the first item data to the user device.

Systems and methods for dynamic in-memory caching of mappings into partitions
11379369 · 2022-07-05 · ·

Systems and methods for routing requests to dynamically cached mappings are disclosed. A system may comprise a memory storing instructions and at least one processor configured to execute instructions to perform operations including: receiving an access request to access first item data from a user device, the access request having an access key; based on the access key, routing the request to a server having an in-memory cache with multiple partitions associated with different access key types and storing the first item data and second item data; receiving, from the server, the first item data; and transmitting the first item data to the user device.

SYSTEMS AND METHODS FOR DYNAMIC IN-MEMORY CACHING OF MAPPINGS INTO PARTITIONS
20220276961 · 2022-09-01 · ·

Systems and methods for routing requests to dynamically cached mappings are disclosed. A system may comprise a memory storing instructions and at least one processor configured to execute instructions to perform operations including: receiving an access request to access first item data from a user device, the access request having an access key; based on the access key, routing the request to a server having an in-memory cache with multiple partitions associated with different access key types and storing the first item data and second item data; receiving, from the server, the first item data; and transmitting the first item data to the user device.

Managing and ranking memory resources

The present disclosure relates to systems, methods, and computer-readable media for tracking memory usage data on a memory controller system and providing a mechanism whereby one or multiple accessing agents (e.g., computing nodes, applications, virtual machines) can access memory usage data for a memory resource managed by a memory controller. Indeed, the systems described herein facilitate generation of and access to heatmaps having memory usage data thereon. The systems described herein describe features and functionality related to generating and maintaining the heatmaps as well as providing access to the heatmaps to a variety of accessing agents. This memory tracking and accessing is performed using low processing overhead while providing useful information to accessing agents in connection with memory resources managed by a memory controller.