G06F2212/163

Optimized read access to shared data via monitoring of mirroring operations
09830088 · 2017-11-28 · ·

A method and system for optimized read access to shared data via monitoring of mirroring operations are described. A data storage system performs operations that include one controller in a dual-controller host storage appliance in an asymmetric active/active configuration receiving a request from the host for data on a logical unit number owned by the partner controller. The receiving controller, which has a mirror cache of the partner controller's memory for failure recovery, accesses the mirror cache using a data structure that was populated during previous mirror operations. If the data is found in the mirror cache, it is read from the cache and returned to the requesting host without having to contact the partner controller for the data.

Fault tolerant cluster data handling

The described technology is generally directed towards fault tolerant cluster data handling techniques, as well as devices and computer readable media configured to perform the disclosed fault tolerant cluster data handling techniques. Nodes in a computing cluster can be configured to generate wire format resources corresponding to operating system resources. A wire format resource can comprise a cache key and a hint information to locate data, such as a file, corresponding to the operating system resource. The wire format resource can be stored in a resource cache along with a pointer that points to the operating system resource. The wire format resource can also be provided to client devices. Nodes in the computing cluster can be configured to receive and process client instructions that include wire format resources, as well as to use hint information to re-allocate data associated with a wire format resource.

PREFETCHING OPERATIONS IN STORAGE DEVICES
20220365878 · 2022-11-17 ·

Provided is a method of adjusting prefetching operations, the method including setting a prefetching distance, accessing a prefetching-trigger key, determining a target key is outside of the prefetching distance from the prefetching-trigger key, increasing the prefetching distance, and successfully fetching a subsequent target key of a subsequent prefetching-trigger key from a prefetching read-ahead buffer.

DETERMINISTIC MULTIFACTOR CACHE REPLACEMENT
20170329720 · 2017-11-16 ·

Some embodiments modify caching server operation to evict cached content based on a deterministic and multifactor modeling of the cached content. The modeling produces eviction scores for the cached items. The eviction scores are derived from two or more factors of age, size, cost, and content type. The eviction scores determine what content is to be evicted based on the two or more factors included in the eviction score derivation. The eviction scores modify caching server eviction operation for specific traffic or content patterns. The eviction scores further modify caching server eviction operation for granular control over an item's lifetime on cache.

DISTRIBUTED CLIENT BASED CACHE FOR KEYS USING DEMAND FAULT INVALIDATION
20170329740 · 2017-11-16 ·

A method and system of storing and retrieving data in a computer system are disclosed. In an embodiment, the method comprises storing a data object at a specified location in a first data store, storing in a second data store first information identifying said specified location, retrieving the data object from a defined location in the first data store, obtaining second information identifying the defined location in the first data store, and comparing the second information to the first information to determine if the first information and the second information identify a same location in the first data store. In embodiments of the invention, the method further comprises when the second information and the first information identify different locations in the first data store, generating a fault signal to indicate that the location of the data object in the first data store has changed.

SYSTEM AND METHOD FOR IMPROVED PERFORMANCE IN A MULTIDIMENSIONAL DATABASE ENVIRONMENT
20220350819 · 2022-11-03 ·

In accordance with an embodiment, described herein is a system and method for improving performance within a multidimensional database computing environment. A multidimensional database, utilizing a block storage option, performs numerous input/output (I/O) operations when executing calculations. To separate I/O operations from calculations, a background task queue is created to identify data blocks requiring I/O. The background task queue is utilized by background writer threads to execute the I/O operations in parallel with calculations.

TRANSACTIONS ON NON-TRANSACTIONAL DATABASE
20220350793 · 2022-11-03 ·

Disclosed are systems, methods, and non-transitory computer-readable media for an improved database management system that provides database transactions on a non-transactional database. The database management system executes garbage collection on data stored in a database to remove data values written to the database as part of uncommitted transactions. Each uncommitted transaction is associated with a respective transaction identifier that is not included in a list of committed transaction identifiers. The list of committed transaction identifiers lists, in sequential order, transaction identifiers for committed transaction. After removing each data value written to the database as part of an uncommitted transaction, the database management system modifies the list of committed transaction identifiers to include the transaction identifier for the uncommitted transaction.

COMPUTER SYSTEM
20170308472 · 2017-10-26 · ·

A computer system, comprising first computers, an application operate on each of the first computers; the each of the first computers is coupled to a second computer for providing a storage area; the each of the first computers includes a processor, a memory, a cache device to which a cache area, and a interface; the memory includes a program for realizing an operating system; the operating system includes a cache driver; and a cooperation control module configured to issue a control I/O request for instructing arrangement control; and the cooperation control module generate the control I/O request from a detected I/O request based on a analysis result of the detected I/O request in a case where an issuance of the I/O request from the cache driver is detected; and transfer the control I/O request to an apparatus different from an apparatus of a transfer destination of the detected I/O request.

Memory management for serverless databases

A solution is disclosed for memory management of serverless databases that includes: based at least on detecting a trigger event, determining whether memory is to be reclaimed; based at least on determining that memory is to be reclaimed, determining an amount of memory to be reclaimed; identifying memory to be reclaimed; and reclaiming the identified memory. Disclosed solutions are flexible, enabling customization of the aggressiveness and manner of memory reclamation. This permits users to specify a tailored balance point between performance and cost, for arrangements that bill users based on resource usage (e.g., memory consumed by a serverless database). In some examples, users specify a ramp-down parameter that is used to determine how much memory can be evicted in a particular reclamation event, time intervals (or another criteria) for triggering a reclamation event, and a definition for whether a cache is active.

Single Double Cuckoo Hash
20170286292 · 2017-10-05 ·

In a network element a decision apparatus has a plurality of multi-way hash tables of single size and double size associative entries. A logic pipeline extracts a search key from each of a sequence of received data items. A hash circuit applies first and second hash functions to the search key to generate first and second indices. A lookup circuit reads associative entries in the hash tables that are indicated respectively by the first and second indices, matches the search key against the associative entries in all the ways. Upon finding a match between the search key and an entry key in an indicated associative entry. A processor uses the value of the indicated associative entry to insert associative entries from a stash of associative entries into the hash tables in accordance with a single size and a double size cuckoo insertion procedure.