Patent classifications
G06F16/1847
Garbage collection for hash-based data structures
Examples disclosed herein are relevant to garbage collection for data structures, such as hash tables. The data structure can store garbage collection values for use during a garbage collection process. The garbage collection values can have a value indicating the occurrence of a poisoned status. Disclosed configurations can be suited for use in high-performance computing applications.
Security for Data at Rest in a Remote Network Management Platform
An embodiment may involve persistent storage including a parent filesystem and a pre-configured amount of free space within the parent filesystem that is dedicated for shared use. The embodiment may also involve one or more processors configured to, for each of a plurality of child filesystems: create a sparse file with an apparent size equivalent to the pre-configured amount of free space; create a virtual mapped device associated with the sparse file; establish one or more cryptographic keys for the virtual mapped device; create an encrypted virtual filesystem for the virtual mapped device and within the sparse file, wherein the encrypted virtual filesystem uses the cryptographic keys for application-transparent encryption and decryption of data stored by way of the encrypted virtual filesystem; and mount the encrypted virtual filesystem within the parent filesystem as one of the child filesystems.
Data management system, method, terminal and medium based on hybrid storage
This application provides a data management system, method, terminal, and medium based on hybrid storage. The data management system includes: a first file system mount module, to mount at least one user-mode file system; a second file system mount module, to mount at least two independent back-end file systems based on the user-mode file system for storing hot data and cold data respectively; a data label module, to label the hot or cold attribute of the data in a user data request; a file system selection module, to store the data in the corresponding back-end file system and/or take the data out from the corresponding back-end file system according to the hot or cold attribute of the data.
FINGERPRINTS FOR COMPRESSED COLUMNAR DATA SEARCH
The present disclosure involves systems, software, and computer implemented methods for compressed columnar data search using fingerprints. One example method includes compressing columnar data that includes dividing the columnar data into multiple data blocks and generating a fingerprint for each data block, storing the compressed columnar data and the generated fingerprints in an in-memory database, receiving a query for the columnar data, for each in-memory data block stored in the in-memory database, determining whether the in-memory data block satisfies the query and in response to a determination that the in-memory data block does not satisfy the query, pruning the in-memory data block from the multiple data blocks to generate an unpruned set of data blocks, decompressing the unpruned set of data blocks, and performing a query search on the decompressed unpruned set of data blocks for the received query.
Electronic storage system
Methods and systems for electronic storage are provided. A storage system comprises a plurality of storage system front ends, a plurality of storage system back ends, and a plurality of solid state drive (SSD) agents. Each storage system front end resides on a server of a plurality of servers. Each server of the plurality of servers comprises one or more storage system back ends of the plurality of storage system back ends. Each storage system front end is able to receive I/O requests and relay information associated with the I/O requests to a relevant storage system back end. The relevant storage system back end communicates metadata associated with the I/O request to an SSD via an SSD agent.
Apparatus and method for managing data storage
Provided are an apparatus and method for managing data storage. A first log structured array stores data in a storage device. A second log structured array in the storage device stores metadata for the data in the first log structured array, wherein the second log structured array storing the metadata for the first log structured data storage system is nested within the first log structured array, and wherein the first and second log structured arrays comprise separate instances of log structured arrays. Address space is allocated in the second log structured array for metadata when the allocation of address space is required for metadata for data stored in the first log structured array.
Efficient resource upgrade
In one implementation, a method includes establishing a connection between a new frontend system resource and an existing frontend system resource for a client network. The method further includes transferring, by a processing device, a frontend system resource role from the existing frontend system resource to the new frontend system resource to enable the existing frontend system resource to operate as a backend system resource.
TRIM SETTING DETERMINATION ON A MEMORY DEVICE
The present disclosure includes apparatuses and methods related to determining trim settings on a memory device. An example apparatus can determine a set of trim settings for the array of memory cells based on the operational characteristics of the array of memory cells, wherein the set of trim settings are associated with desired operational characteristics for the array of memory cells.
Application level SD card space management
An apparatus includes a removable media interface circuit and a processor. The removable media interface may be configured to read and write files to a non-volatile storage medium. The processor may be configured to generate encoded data and manage file operations involving storing the encoded data on the non-volatile storage medium to minimize a number of file fragments.
Nonvolatile/persistent memory drive with address subsections configured for respective read bandwidths
This disclosure provides for improvements in managing multi-drive, multi-die or multi-plane NAND flash memory. In one embodiment, the host directly assigns physical addresses and performs logical-to-physical address translation in a manner that reduces or eliminates the need for a memory controller to handle these functions, and initiates functions such as wear leveling in a manner that avoids competition with host data accesses. A memory controller optionally educates the host on array composition, capabilities and addressing restrictions. Host software can therefore interleave write and read requests across dies in a manner unencumbered by memory controller address translation. For multi-plane designs, the host writes related data in a manner consistent with multi-plane device addressing limitations. The host is therefore able to “plan ahead” in a manner supporting host issuance of true multi-plane read commands.