G06F2212/60

Mechanism for interrupting and resuming execution on an unprotected pipeline processor

Techniques related to executing a plurality of instructions by a processor comprising receiving a first instruction for execution on an instruction execution pipeline, beginning execution of the first instruction, receiving one or more second instructions for execution on the instruction execution pipeline, the one or more second instructions associated with a higher priority task than the first instruction, storing a register state associated with the execution of the first instruction in one or more registers of a capture queue associated with the instruction execution pipeline, copying the register state from the capture queue to a memory, determining that the one or more second instructions have been executed, copying the register state from the memory to the one or more registers of the capture queue, and restoring the register state to the instruction execution pipeline from the capture queue.

Domain-based access in a memory device

Methods, systems, and devices related to domain-based access in a memory device are described. In one example, a memory device in accordance with the described techniques may include a memory array, a sense amplifier array, and a signal development cache configured to store signals (e.g., cache signals, signal states) associated with logic states (e.g., memory states) that may be stored at the memory array (e.g., according to various read or write operations). The memory array may be organized according to domains, which may refer to various configurations or collections of access lines, and selections thereof, of different portions of the memory array. In various examples, a memory device may determine a plurality of domains for a received access command, or an order for accessing a plurality of domains for a received access command, or combinations thereof, based on an availability of the signal development cache.

Smart span prioritization based on ingestion service backpressure

Disclosed techniques relate to automatically instrumenting a web application. In an aspect, a method identifies that a web application includes an event that is triggered by a user interaction. The method associates the event with a tracer that is configured to log tracing information based on an execution of a first set of operations caused by the event and to obtain a first measurement of performance of a first span. The method identifies, in the code, that the execution of a first set of operations causes a request to be made to a server. The method associates the request with the tracer. The tracer is configured to log tracing information based on an execution of a second set of operations caused by the request and to obtain a second measurement of performance of a second span that is a child span of the first span.

STREAMING ENGINE WITH STREAM METADATA SAVING FOR CONTEXT SWITCHING
20230004391 · 2023-01-05 ·

A streaming engine employed in a digital data processor specifies a fixed read only data stream defined by plural nested loops. An address generator produces addresses of data elements. A steam head register stores data elements next to be supplied to functional units for use as operands. Stream metadata is stored in response to a stream store instruction. Stored stream metadata is restored to the stream engine in response to a stream restore instruction. An interrupt changes an open stream to a frozen state discarding stored stream data. A return from interrupt changes a frozen stream to an active state.

METHOD AND SYSTEM FOR IMPROVING CACHE MANAGEMENT IN A 5G NETWORK FUNCTION

A method and system for improving cache management in a 5th generation (5G) network function (NF) is provided. The method includes receiving a list of requests from a plurality of user equipment (UE) requiring access to their corresponding user context. The method further includes determining a priority value for each of the plurality of UEs in the received list based on at least one of analytics data received from a network and data analysis function (NWDAF), paging data, or cache eviction policy. The method further includes prioritizing storage of user contexts in a local cache unit based on the priority value for each of the plurality of UEs. The method further includes receiving a request for accessing a user context from a UE of the plurality of UEs and accessing the user context from the cache unit based on the received request.

Handling memory requests

A converter module is described which handles memory requests issued by a cache (e.g. an on-chip cache), where these memory requests include memory addresses defined within a virtual memory space. The converter module receives these requests, issues each request with a transaction identifier and uses that identifier to track the status of the memory request. The converter module sends requests for address translation to a memory management unit and where there the translation is not available in the memory management unit receives further memory requests from the memory management unit. The memory requests are issued to a memory via a bus and the transaction identifier for a request is freed once the response has been received from the memory. When issuing memory requests onto the bus, memory requests received from the memory management unit may be prioritized over those received from the cache.

STORAGE DEVICE AND OPERATING METHOD THEREOF
20220405009 · 2022-12-22 ·

A storage device may include: a memory device; a cache memory device including a first cache memory which caches first data among data stored in the plurality of pages and a second cache memory which caches second data among the data stored in the plurality of pages; and a memory controller for counting a number of times that each of the plurality of pages is read and a number of times that each of the plurality of pages is written, based on a read request or a write request which are received from a host, and, moving the first data from the first cache memory to the second cache memory when the first data is stored in a first page and a number of times that the first page is read and a number of times that the first page is written satisfy a predetermined condition.

INTELLIGENT CACHE WITH READ DESTRUCTIVE MEMORY CELLS

A data storage system can employ a read destructive memory configured to fill a first cache with a first data set from a data repository prior to populating a second cache with a second data set describing the first data set with the first and second cache each having non-volatile ferroelectric memory cells. An entirety of the first cache may be read in response to a cache hit in the second cache with the cache hit responsive to a data read command from a host and with the first cache being read without a refresh operation restoring the data of the first cache.

APPARATUSES, SYSTEMS, AND METHODS FOR CONFIGURING COMBINED PRIVATE AND SHARED CACHE LEVELS IN A PROCESSOR-BASED SYSTEM

Apparatuses, systems, and methods for configuring combined private and shared cache levels in a processor-based system. The processor-based system includes a processor that includes a plurality of processing cores each including execution circuits which are coupled to respective cache(s) and a configurable combined private and shared cache, and which may receive instructions and data on which to perform operations from the cache(s) and the combined private and shared cache. A shared cache portion of each configurable combined private and shared cache can be treated as an independently-assignable portion of the overall shared cache, which is effectively the shared cache portions of all of the processing cores. Each independently-assignable portion of the overall shared cache can be associated with a particular client running on the processor as an example. This approach can provide greater granularity of cache partitioning of a shared cache between particular clients running on a processor.

Hybrid memory systems with cache management
11526441 · 2022-12-13 · ·

In a general aspect, a hybrid memory system with cache management is disclosed. In some aspects, a memory module includes volatile memory, non-volatile memory, and an internal cache. The internal cache is communicably coupled with the volatile memory and the non-volatile memory. Whether to execute a memory access request is determined by operation of the memory module. In response to the inability of the memory access request to be executed, a data transferring process is performed to copy data between the volatile memory and the non-volatile memory via the internal cache.