G06F2212/60

Destaging multiple cache slots in a single back-end track in a RAID subsystem

A data service layer running on a storage director node generates a request to destage host data from a plurality of cache slots in a single back-end track. The destage request includes pointers to addresses of the cache slots and indicates an order in which the host application data in the cache slots is to be included in the back-end track. A back-end redundant array of independent drives (RAID) subsystem running on a drive adapter is responsive to the request to calculate parity information using the host application data in the cache slots. The back-end RAID subsystem assembles the single back-end track comprising the host application data from the plurality of cache slots of the request, and destages the single back-end track to a non-volatile drive in a single back-end input-output (IO) operation.

UPWARD EVICTION OF CACHE LINES

The disclosure relates to technology for up-evicting cache lines. An apparatus comprises a hierarchy of caches comprising a first cache having a first cache controller and a second cache having a second cache controller. The first cache controller is configured to store cache lines evicted from a first processor group to the first cache and to down-evict cache lines from the first cache to the second cache. The second cache controller is configured to store cache lines evicted from a second processor group into the second cache, to up-evict a first cache line from the second cache to the first cache in response to an eviction of a second cache line from the second processor group to the second cache, and to provide the up-evicted first cache line from the first cache to the second processor group in response to a request from the second processor group.

System and method for multi-node buffer transfer

A method, computer program product, and computing system for receiving, at a local node, a request to buffer data on a remote persistent cache memory system of a remote node. A target memory address within the remote persistent cache memory system may be sent from the local node via a remote procedure call (RPC). The data may be sent from the local node to the target memory address within the remote persistent cache memory system via a remote direct memory access (RDMA) command.

AIR FREIGHT RATE DATA CACHING METHOD AND SYSTEM

An air freight rate data caching method and system. The method includes converting air freight rate data into a data format of a first-level cache, and storing same in the first-level cache; performing, on the basis of a flight origin city and a flight destination city, data fragmentation on the air freight rate data stored in the first-level cache so as to generate fragmented data; and storing the fragmented data, after same is validated, in a second-level cache. Each data node of the fragmented data cached in the second-level cache only includes part of the air freight rate data on which a fragmentation algorithm can be performed, and therefore, the horizontal expansion capacity of a cache system is improved relative to the case where cached data copies are all complete sets.

METHOD OF CACHE MANAGEMENT BASED ON FILE ATTRIBUTES, AND CACHE MANAGEMENT DEVICE OPERATING BASED ON FILE ATTRIBUTES
20220365876 · 2022-11-17 ·

A method of cache management includes: acquiring attributes of at least one file currently opened; extracting a first attribute subset related to a predetermined analysis target from among the attributes of the at least one file; determining region numbers of corresponding universal flash storage (UFS) device storage regions to be cached, based on the first attribute subset; and caching data of the corresponding UFS device storage regions to be cached into a cache memory, based on the region numbers.

Automatic span context propagation to worker threads in rich-client applications

Disclosed techniques relate to tracing a subsidiary thread of a web page application. In an example, a method includes providing a web page application to a web browser. The method further includes creating, from the web page application, a global tracer configured to log tracing data for the web page application and a span. The method further includes instantiating, from the web page application, a wrapper for a subsidiary thread. The wrapper is configured to execute shim code before executing the subsidiary thread. The method further includes passing, from the wrapper to the shim code, configuration data of the global tracer. The method further includes creating a subsidiary tracer from the shim code and the configuration data. The subsidiary tracer configured to log tracing data for the subsidiary thread and a thread span.

METHODS AND SYSTEMS FOR HANDLING DATA RECEIVED BY A STATE MACHINE ENGINE
20230048032 · 2023-02-16 ·

A data analysis system to analyze data. The data analysis system includes a data buffer configured to receive data to be analyzed. The data analysis system also includes a state machine lattice. The state machine lattice includes multiple data analysis elements and each data analysis element includes multiple memory cells configured to analyze at least a portion of the data and to output a result of the analysis. The data analysis system includes a buffer interface configured to receive the data from the data buffer and to provide the data to the state machine lattice.

MULTI-LEVEL INDEPENDENT SECURITY ARCHITECTURE
20230049021 · 2023-02-16 · ·

A system includes a plurality of data input ports, each port corresponding to one of a plurality of different levels of security classification; a security device, configured for cryptographic processing, coupled to receive incoming data from each of the plurality of input ports, wherein the incoming data includes first data having a first classification level; a key manager configured to select and tag-identified first set of keys from a plurality of key sets, each of the key sets corresponding to one of the different levels of security classification, wherein the first set of keys is used by the security device to encrypt the first data; and a common encrypted data storage, coupled to receive the encrypted first data from the security device for storage.

Streaming Network Monitoring Caching Infrastructure

Systems and methods of network telemetry caching and distribution are provided. The system can receive network telemetry data and store it as a plurality of data nodes. The system can maintain a node pointer map and a node pointer queue. If the system receives an update to a data node having a corresponding node pointer not already present in the node pointer map, the system can add the node pointer to the node pointer queue and to the node pointer map with a count of zero. If the node pointer is already present in the node pointer map, the system can increment the node count for the node pointer in the node pointer map and not add the node pointer to the node pointer queue. The system can transmit data values and node counts to the client device for each node pointer in the node pointer queue.

Data Storage Device and Method for Host-Initiated Cached Read to Recover Corrupted Data Within Timeout Constraints

A data storage device and method for host-initiated cached read to recover corrupted data within timeout constraints are provided. In one embodiment, a data storage device is provided comprising a volatile memory, a non-volatile memory, and a controller. The controller is configured to receive a read look-ahead command from a host to perform a read look-ahead of a first logical address; receive a read command from the host to read a second logical address; and execute the read look-ahead command by performing the following as background operations while executing the read command: read data for a location in the non-volatile memory that corresponds to the first logical address; correct an error in the data; and cache the corrected data in the volatile memory. The cached corrected data can be sent back to the host in response to the host requesting a read of the same logical address. Other embodiments are provided.