G06F12/0877

Quasi-volatile system-level memory

A high-capacity system memory may be built from both quasi-volatile (QV) memory circuits, logic circuits, and static random-access memory (SRAM) circuits. Using the SRAM circuits as buffers or cache for the QV memory circuits, the system memory may achieve access latency performance of the SRAM circuits and may be used as code memory. The system memory is also capable of direct memory access (DMA) operations and includes an arithmetic logic unit for performing computational memory tasks. The system memory may include one or more embedded processor. In addition, the system memory may be configured for multi-channel memory accesses by multiple host processors over multiple host ports. The system memory may be provided in the dual-in-line memory module (DIMM) format.

Method, electronic device and computer program product for managing cache
11593268 · 2023-02-28 · ·

Techniques for cache management involve accessing, when a first data block to be accessed is missing in a first cache, the first data block from a storage device storing the first data block; selecting, when the first cache is full and based on a plurality of parameters associated with a plurality of eviction policies, an eviction policy for evicting a data block in the first cache from the plurality of eviction policies, the plurality of parameters indicating corresponding possibilities that the plurality of eviction policies are selected; evicting a second data block in the first cache to a second cache based on the selected eviction policy, the second cache being configured to record the data block evicted from the first cache; and caching the accessed first data block in the first cache. Such techniques can improve the cache hit rate, thereby improving the access performance of a system.

Method, electronic device and computer program product for managing cache
11593268 · 2023-02-28 · ·

Techniques for cache management involve accessing, when a first data block to be accessed is missing in a first cache, the first data block from a storage device storing the first data block; selecting, when the first cache is full and based on a plurality of parameters associated with a plurality of eviction policies, an eviction policy for evicting a data block in the first cache from the plurality of eviction policies, the plurality of parameters indicating corresponding possibilities that the plurality of eviction policies are selected; evicting a second data block in the first cache to a second cache based on the selected eviction policy, the second cache being configured to record the data block evicted from the first cache; and caching the accessed first data block in the first cache. Such techniques can improve the cache hit rate, thereby improving the access performance of a system.

Sector cache for compression

In an example, an apparatus comprises a plurality of execution units, and a cache memory communicatively coupled to the plurality of execution units, wherein the cache memory is structured into a plurality of sectors, wherein each sector in the plurality of sectors comprises at least two cache lines. Other embodiments are also disclosed and claimed.

Sector cache for compression

In an example, an apparatus comprises a plurality of execution units, and a cache memory communicatively coupled to the plurality of execution units, wherein the cache memory is structured into a plurality of sectors, wherein each sector in the plurality of sectors comprises at least two cache lines. Other embodiments are also disclosed and claimed.

SYSTEM AND METHOD FOR REDUCING ECC OVERHEAD AND MEMORY ACCESS BANDWIDTH
20180011758 · 2018-01-11 · ·

A system, and corresponding method, is described for updating or calculating ECC where the transaction volume is significantly reduced from a read-modify-write to a write, which is more efficient and reduces demand on the data access bandwidth. The invention can be implemented in any chip, system, method, or HDL code that perform protection schemes and require ECC calculation, of any kind. Embodiments of the invention enable IPs that use different protections schemes to reduce power consumption and reduce bandwidth access to more efficiently communicate or exchange information.

MEMORY BUILT-IN DEVICE, PROCESSING METHOD, PARAMETER SETTING METHOD, AND IMAGE SENSOR DEVICE
20230236984 · 2023-07-27 · ·

A memory built-in device according to the present disclosure is a memory built-in device including a processor; a memory access controller; and a memory to be accessed in accordance with a process by the memory access controller, wherein the memory access controller is configured to read and write data to be used in an operation of a convolution arithmetic circuit from and to the memory according to designation of a parameter.

MEMORY BUILT-IN DEVICE, PROCESSING METHOD, PARAMETER SETTING METHOD, AND IMAGE SENSOR DEVICE
20230236984 · 2023-07-27 · ·

A memory built-in device according to the present disclosure is a memory built-in device including a processor; a memory access controller; and a memory to be accessed in accordance with a process by the memory access controller, wherein the memory access controller is configured to read and write data to be used in an operation of a convolution arithmetic circuit from and to the memory according to designation of a parameter.

Oldest operation wait time indication input into set-dueling
11561895 · 2023-01-24 · ·

Systems, apparatuses, and methods for dynamically adjusting cache policies to reduce execution core wait time are disclosed. A processor includes a cache subsystem. The cache subsystem includes one or more cache levels and one or more cache controllers. A cache controller partitions a cache level into two test portions and a remainder portion. The cache controller applies a first policy to the first test portion and applies a second policy to the second test portion. The cache controller determines the amount of time the execution core spends waiting on accesses to the first and second test portions. If the measured wait time is less for the first test portion than for the second test portion, then the cache controller applies the first policy to the remainder portion. Otherwise, the cache controller applies the second policy to the remainder portion.

DATA PROCESSING DEVICE AND RELATED PRODUCT
20230214327 · 2023-07-06 ·

A data processing device and related products are provided. The data processing device includes: a decoding unit, a discrete address determining unit, a continuous data caching unit, a data read/write unit, and a storage unit. Through the data processing device, the processing instruction may be decoded and executed, and the discrete data may be transferred to a continuous data address, or the continuous data may be stored to a plurality of discrete data addresses. As such, a vector computation of the discrete data and vector data restoration after the vector computation may be implemented, which may simplify a processing process, thereby reducing data overheads. In addition, according to the embodiments of the disclosure, when the discrete data is read, by caching a storage address corresponding to a read request, a read request of each piece of data may be merged to read one or more pieces of discrete data, thereby improving reading efficiency of the data.