Patent classifications
G06F2212/224
DYNAMIC RESULT SET CACHING WITH A DATABASE ACCELERATOR
According to one embodiment of the present invention, a system for processing a database query stores one or more result sets for one or more first database queries in a data store. The system receives a second database query and compares the second database query to the one or more first database queries to determine presence of a corresponding result set in the data store for the second database query. The system provides the corresponding result set from the data store for the second database query based on the comparison. Embodiments of the present invention further include a method and computer program product for processing a database query in substantially the same manners described above.
DYNAMIC RESULT SET CACHING WITH A DATABASE ACCELERATOR
According to one embodiment of the present invention, a system for processing a database query stores one or more result sets for one or more first database queries in a data store. The system receives a second database query and compares the second database query to the one or more first database queries to determine presence of a corresponding result set in the data store for the second database query. The system provides the corresponding result set from the data store for the second database query based on the comparison. Embodiments of the present invention further include a method and computer program product for processing a database query in substantially the same manners described above.
Methods and systems for performing a copy forward operation
A storage device made up of multiple storage media is configured such that one such media serves as a cache for data stored on another of such media. The device includes a controller configured to manage the cache by consolidating information concerning obsolete data stored in the cache with information concerning data no longer desired to be stored in the cache, and erase segments of the cache containing one or more of the blocks of obsolete data and the blocks of data that are no longer desired to be stored in the cache to produce reclaimed segments of the cache.
Tiered data storage system
A storage media comprising a first storage subset with a first value of a storage media characteristic and a second storage subset with a second value of the storage media characteristic, the first value of a storage media characteristic being substantially different than the second value of the storage media characteristic and a storage controller configured to allocate a plurality of logical block addresses (LBAs) between the first storage subset and the second storage subset based on a predetermined criterion in view of the first value of the storage media characteristic and the second value of the storage media characteristic.
STORAGE SYSTEM AND A METHOD USED BY THE STORAGE SYSTEM
Performing failover processing between a production host and a backup host, a storage system is connected to the production host and the backup host. In response to a failure of the production host, metadata is obtained of data blocks that have been cached from an elastic space located in a fast disk of the storage system. A storage capacity of the elastic space is expanded. Data blocks are obtained to which the metadata corresponds according to the metadata and the storage capacity of the expanded elastic space, and storing the same in the expanded elastic space. In response the backup host requesting the data blocks to which the metadata corresponds, and the data blocks to which the metadata corresponds have already been stored in the expanded elastic space, data blocks are obtained to which the metadata corresponds from the expanded elastic space and transmitting the same to the backup host.
System and method for managing a cache pool
In one embodiment, a system includes a processor and a memory communicatively coupled to the processor. The processor is configured to receive a write request associated with a cache pool, which comprises a plurality of disks. The write request comprises data associated with the write request. The processor is additionally configured to select a first disk from the plurality of disks using a life parameter associated with the first disk. The processor is further configured to cause the data associated with the write request to be written to the first disk.
Systems and Methods for Rapid Processing and Storage of Data
Systems and methods of building massively parallel computing systems using low power computing complexes in accordance with embodiments of the invention are disclosed. A massively parallel computing system in accordance with one embodiment of the invention includes at least one Solid State Blade configured to communicate via a high performance network fabric. In addition, each Solid State Blade includes a processor configured to communicate with a plurality of low power computing complexes interconnected by a router, and each low power computing complex includes at least one general processing core, an accelerator, an I/O interface, and cache memory and is configured to communicate with non-volatile solid state memory.
SYSTEM AND METHOD FOR USING A MEMORY BUFFER TO STREAM DATA FROM A TAPE TO MULTIPLE CLIENTS
Embodiments described herein provide systems and methods for streaming data to multiple reading clients. More particularly, embodiments described herein provide systems and methods for using a memory buffer to stream file data from a mounted tape file system volume to multiple reading clients.
Dynamic result set caching with a database accelerator
According to one embodiment of the present invention, a system for processing a database query stores one or more result sets for one or more first database queries in a data store. The system receives a second database query and compares the second database query to the one or more first database queries to determine presence of a corresponding result set in the data store for the second database query. The system provides the corresponding result set from the data store for the second database query based on the comparison. Embodiments of the present invention further include a method and computer program product for processing a database query in substantially the same manners described above.
Dynamic result set caching with a database accelerator
According to one embodiment of the present invention, a system for processing a database query stores one or more result sets for one or more first database queries in a data store. The system receives a second database query and compares the second database query to the one or more first database queries to determine presence of a corresponding result set in the data store for the second database query. The system provides the corresponding result set from the data store for the second database query based on the comparison. Embodiments of the present invention further include a method and computer program product for processing a database query in substantially the same manners described above.