G06F12/00

SYSTEM AND METHOD FOR ACCELERATED DATA SEARCH OF DATABASE STORAGE SYSTEM
20230029029 · 2023-01-26 ·

Embodiments of the present disclosure provide a system for accelerated data search of a database storage system. The system includes a host device including a database storage engine; and a memory system including a controller and a memory device, which includes a plurality of pages storing multiple records. The controller includes a page processing accelerator configured to: read, from the plurality of pages, multiple pages in response to a filtered read command; filter particular pages among the multiple pages based on a column full search condition, the filtered pages including entries satisfying the column full search condition; and transfer, to the host device, information regarding the filtered pages.

MEMORY MODULE

In a memory module according to an embodiment of the present disclosure, a controller, upon reception of a read command including a logical address, converts the logical address included in the read command into a physical address using address lookup information. The controller further inputs a first physical address, which is a portion of the physical address obtained by the conversion, to a non-volatile memory via a first address bus terminal, and then inputs a second physical address, which is a rest of the physical address obtained by the conversion, to the non-volatile memory via a second address bus terminal, to thereby read data corresponding to the physical address obtained by the conversion from the non-volatile memory.

Queue Versioning
20230229696 · 2023-07-20 ·

Embodiments described herein may involve queue versioning. An example implementation may involve a playback device initiating playback of a queue including one or more first audio streams. A queue identification token stored in data storage represents a current version of the queue. The playback device receives, from a computing system, data representing instructions to add one or more second audio streams to the queue, the instructions including an indication of the one or more second audio streams and a first token representing an expected queue version. The playback device determines whether the expected queue version represented by the first token matches the current version of the queue represented by the queue identification token. If not, the playback device foregoes adding the one or more second audio streams to the queue.

Queue Versioning
20230229696 · 2023-07-20 ·

Embodiments described herein may involve queue versioning. An example implementation may involve a playback device initiating playback of a queue including one or more first audio streams. A queue identification token stored in data storage represents a current version of the queue. The playback device receives, from a computing system, data representing instructions to add one or more second audio streams to the queue, the instructions including an indication of the one or more second audio streams and a first token representing an expected queue version. The playback device determines whether the expected queue version represented by the first token matches the current version of the queue represented by the queue identification token. If not, the playback device foregoes adding the one or more second audio streams to the queue.

Integration of application indicated minimum time to cache to least recently used track demoting schemes in a cache management system of a storage controller

A minimum retention time in cache is indicated for a first plurality of tracks, where no minimum retention time is indicated for a second plurality of tracks. A cache management application demotes a track of the first plurality of tracks from the cache, in response to determining that the track is a least recently used (LRU) track in a LRU list of tracks in the cache and the track has been in the cache for a time that exceeds the minimum retention time.

Method and system of managing a file cache by selective delete-on-read operations

Disclosed is a computer implemented method to manage a cache, the method comprising, determining that a primary application opens a first file, wherein opening the first file includes reading the first file into a file cache from a storage. The method also includes, setting a first monitoring variable in the primary application process proc structure, wherein the first monitoring variable is set in response to the primary application opening the first file, and the first monitoring variable records a set of operations completed on the first file by the primary application. The method comprises a first read of the first file being at a beginning of the first file. The method includes identifying that the first file is read according to a pattern that includes reading the first file sequentially and reading the first file entirely and removing the first file from the file cache.

Selecting read voltage using write transaction data
11561734 · 2023-01-24 · ·

A system includes a memory component; and a processing device, operatively coupled with the memory component. The processing device is to perform operations including receiving a read request with respect to data stored at a physical address of the memory component; determining whether an indicator of the physical address is stored in a write transaction catalog; in response to determining that the physical address is stored in the write transaction catalog, determining a time difference between when the read request was received and when the data was written; reading the data stored at the physical address using a first read voltage level in response to determining that the time difference is less than a threshold criterion; and reading the data stored at the physical address using a second read voltage level in response to determining that the time difference is equal to or greater than the threshold criterion.

Solid state storage device with variable logical capacity based on memory lifecycle
11704025 · 2023-07-18 · ·

Several embodiments of memory devices and systems having a variable logical memory capacity are disclosed herein. In one embodiment, a memory device can include a plurality of memory regions that collectively define a physical memory capacity and a controller operably coupled to the plurality of memory regions. The controller is configured to advertise a first logical memory capacity to a host device, determine that at least one of the memory regions is at or near end of life, and in response to the determination—send a notification to the host device that a logical memory capacity of the memory device will be reduced and then retire the at least one of the memory regions.

Systems and methods for use in segregating data blocks to distributed storage

Systems and methods are provided for storing data blocks in distributed storage. One example computer-implemented method includes, in response to receipt of a data block comprising data, generating a value N for the data block, wherein the value N includes a variable integer greater than one and dividing the data block into N segments, wherein each segment includes a portion of the data. The method also includes generating a value M for the data block, wherein the value M includes a variable integer greater than or equal to one, and adding M segments of chaff to the N segments. The method then includes encrypting the N segments and the M segments of chaff and distributing the M segments and the N segments in distributed storage, wherein the N segments and the M segments of chaff are stored in multiple different storage devices included in the distributed storage.

Deduplication decision based on metrics

Systems and method for implementing deduplication process based on performance analyses. The system may include a processing device to determine a first performance metric associated with retrieving a second stored data block that is within a specified range of a duplicate of the first data block and a second performance metric associated with retrieving a hash value corresponding to the second stored data block. The processing device further to retrieve the second stored data block within a specified range of the duplicate of the first data block in response to the first performance metric not exceeding the second performance metric.