G06F2212/60

Phase change memory in a dual inline memory module

Subject matter disclosed herein relates to management of a memory device.

Streaming network monitoring caching infrastructure

Systems and methods of network telemetry caching and distribution are provided. The system can receive network telemetry data and store it as a plurality of data nodes. The system can maintain a node pointer map and a node pointer queue. If the system receives an update to a data node having a corresponding node pointer not already present in the node pointer map, the system can add the node pointer to the node pointer queue and to the node pointer map with a count of zero. If the node pointer is already present in the node pointer map, the system can increment the node count for the node pointer in the node pointer map and not add the node pointer to the node pointer queue. The system can transmit data values and node counts to the client device for each node pointer in the node pointer queue.

Using Epoch Counter Values for Controlling the Retention of Cache Blocks in a Cache
20230095461 · 2023-03-30 ·

An electronic device includes a cache, a memory, and a controller. The controller stores an epoch counter value in metadata for a location in the memory when a cache block evicted from the cache is stored in the location. The controller also controls how the cache block is retained in the cache based at least in part on the epoch counter value when the cache block is subsequently retrieved from the location and stored in the cache.

SUPPRESSING CACHE LINE MODIFICATION
20230096563 · 2023-03-30 · ·

Disclosed is a system and method for use in a cache for suppressing modification of cache line. The system and method includes a processor and a memory operating cooperatively with a cache controller. The memory includes a coherence directory stored within a cache created to track at least one cache line in the cache via the cache controller. The processor instructs a cache controller to store a first data in a cache line in the cache. The cache controller tags the cache line based on the first data. The processor instructs the cache controller to store a second data in the cache line in the cache causing eviction of the first data from the cache line. The processor compares based on the tagging the first data and the second data and suppresses modification of the cache line based on the comparing of the first data and the second data.

RE-REFERENCE INTERVAL PREDICTION (RRIP) WITH PSEUDO-LRU SUPPLEMENTAL AGE INFORMATION
20230102891 · 2023-03-30 · ·

Systems and methods for cache replacement are disclosed. Techniques are described that determine a re-reference interval prediction (RRIP) value of respective data blocks in a cache, where an RRIP value represents a likelihood that a respective data block will be re-used within a time interval. Upon an access, by a processor, to a data segment in a memory, if the data segment is not stored in the cache, a data block in the cache to be replaced by the data segment is selected, utilizing a binary tree that tracks recency of data blocks in the cache.

Deep-learning-based image processing method and system

An image processing method is provided, which is applied to a deep learning model. A cache queue is provided in front of each layer of the deep learning model; a plurality of computation tasks are preset for each layer of the deep learning model in advance, and are configured for computing weight parameters and corresponding to-be-processed data in a plurality of channels in each corresponding layer in parallel, and storing a computation result into a cache queue behind the corresponding layer thereof; in addition, as long as the cache queue in front of the layer includes the computation result stored in the previous layer, the layer can obtain the to-be-processed data from the computation result, subsequent computation is performed, and a parallel pipeline computation mode is also formed between the layers. By means of the mode, the throughput rate during image processing is remarkably improved, and the image processing parallelism degree and speed and the computation performance of the deep learning model are improved. Further provided are an image processing device and system, which have the same beneficial effects as the above image processing method.

Data storage system with adaptive, memory-efficient cache flushing structure

In a method of flushing cached data in a data storage system, instances of a working-set structure (WSS) are used over a succession of operating periods to organize cached data for storing to the persistent storage. In each operating period, leaf structures of the WSS are associated with respective address ranges of a specified size. Between operating periods, a structure-tuning operation is performed to adjust the specified size and thereby dynamically adjust a PD-to-leaf ratio of the WSS, including (1) comparing a last-period PD-to-leaf ratio to a predetermined ratio range, (2) when the ratio is below the predetermined ratio range, increasing the specified size for use in a next operating period, and (3) when ratio is above the predetermined ratio range, then decreasing the specified size for use in the next operating period.

METHODS FOR CACHING AND READING CONTENT, CLIENT, AND STORAGE MEDIUM
20220350740 · 2022-11-03 ·

A method for caching content, a method for reading content, a client, and a storage medium are provided. The method for caching content includes: acquiring JSON data corresponding to content to be delivered, and determining identification information corresponding to the JSON data; grouping the JSON data and storing the grouped JSON data according to the identification information to obtain a memory list corresponding to the identification information, and writing the JSON data to a target disk according to the identification information and the memory list; performing video preloading processing on the content to be delivered according to the JSON data to obtain preloaded video data, and determining address information corresponding to the preloaded video data; and storing the address information in the memory list and the target disk according to the identification information, to complete caching of the JSON data and the preloaded video data.

SMART SPAN PRIORITIZATION BASED ON INGESTION SERVICE BACKPRESSURE

Disclosed techniques relate to automatically instrumenting a web application. In an aspect, a method identifies that a web application includes an event that is triggered by a user interaction. The method associates the event with a tracer that is configured to log tracing information based on an execution of a first set of operations caused by the event and to obtain a first measurement of performance of a first span. The method identifies, in the code, that the execution of a first set of operations causes a request to be made to a server. The method associates the request with the tracer. The tracer is configured to log tracing information based on an execution of a second set of operations caused by the request and to obtain a second measurement of performance of a second span that is a child span of the first span.

SELECTION OF VICTIM ENTRY IN A DATA STRUCTURE
20230086723 · 2023-03-23 ·

In an embodiment, a processor may include an execution engine to execute a plurality of instructions, a memory to store a tagged data structure comprising a plurality of entries, and an eviction circuit. The eviction circuit may be to: generate a pseudo-random number responsive to an eviction request for the tagged data structure; in response to a determination that the pseudo-random number is outside of a valid eviction range for the plurality of entries, generate an alternative identifier by rotating through the valid eviction range, the valid eviction range comprising a range of numbers that are valid to identify victim entries of the tagged data structure; and evict a victim entry from the tagged data structure, the victim entry associated with the alternative identifier. Other embodiments are described and claimed.