Patent classifications
G06F12/0853
Random tag setting instruction for a tag-guarded memory system
An apparatus has processing circuitry (4); memory access circuitry (15) to perform a guard tag check for a tag checking target address having an associated address tag, the guard tag check comprising comparing the address tag with a guard tag stored in a memory system in association with a block of one or more memory locations comprising an addressed location identified by the target address; and an instruction decoder (6) responsive to a random tag setting instruction specifying a tag setting target address, to control the processing circuitry (4) to set the address tag associated with the tag setting target address to a random tag value randomly selected from a set of candidate tag values.
DATA PROCESSOR
A data processor comprises a memory-management-unit for receiving external-operation-data from a CPU. The memory-management-unit sets a deterministic-quantity value for the external-operation-data based on the external-operation-data. The deterministic-quantity value may be either an active-value or an inactive-value. The data processor has a non-deterministic-processor-block for receiving a memory-signal from the memory-management-unit, and has a control-block configured to (i) send the memory-signal to an NDP-output-terminal if the deterministic-quantity value is the active-value, thereby bypassing a performance-enhancement-block, or (ii) send at least a portion of the memory-signal that is representative of the request for response-data to the performance-enhancement-block if the deterministic-quantity value is the inactive-value.
DATA PROCESSOR
A data processor comprises a memory-management-unit for receiving external-operation-data from a CPU. The memory-management-unit sets a deterministic-quantity value for the external-operation-data based on the external-operation-data. The deterministic-quantity value may be either an active-value or an inactive-value. The data processor has a non-deterministic-processor-block for receiving a memory-signal from the memory-management-unit, and has a control-block configured to (i) send the memory-signal to an NDP-output-terminal if the deterministic-quantity value is the active-value, thereby bypassing a performance-enhancement-block, or (ii) send at least a portion of the memory-signal that is representative of the request for response-data to the performance-enhancement-block if the deterministic-quantity value is the inactive-value.
Methods and apparatus to facilitate an atomic operation and/or a histogram operation in cache pipeline
Methods, apparatus, systems and articles of manufacture to facilitate an atomic operation and/or a histogram operation in cache pipeline are disclosed. An example system includes a cache storage coupled to an arithmetic component; and a cache controller coupled to the cache storage, wherein the cache controller is operable to: receive a memory operation that specifies a set of data; retrieve the set of data from the cache storage; utilize the arithmetic component to determine a set of counts of respective values in the set of data; generate a vector representing the set of counts; and provide the vector.
Methods and apparatus to facilitate an atomic operation and/or a histogram operation in cache pipeline
Methods, apparatus, systems and articles of manufacture to facilitate an atomic operation and/or a histogram operation in cache pipeline are disclosed. An example system includes a cache storage coupled to an arithmetic component; and a cache controller coupled to the cache storage, wherein the cache controller is operable to: receive a memory operation that specifies a set of data; retrieve the set of data from the cache storage; utilize the arithmetic component to determine a set of counts of respective values in the set of data; generate a vector representing the set of counts; and provide the vector.
Caching device, cache, system, method and apparatus for processing data, and medium
A caching device, an instruction cache, a system for processing an instruction, a method and apparatus for processing data and a medium are provided. The caching device includes a first queue, a second queue, a write port group, a read port, a first pop-up port, a second pop-up port and a press-in port. The is configured to write cache data into a set storage address in the first queue and/or the second queue; the read port is configured to read all cache data from the first queue and/or the second queue at one time; the press-in port is configured to press cache data into the first queue and/or the second queue; the first pop-up port is configured to pop up cache data from the first queue; and the second pop-up port is configured to pop up cache data from the second queue.
Method for implementing a line speed interconnect structure
A method for line speed interconnect processing. The method includes receiving initial inputs from an input communications path, performing a pre-sorting of the initial inputs by using a first stage interconnect parallel processor to create intermediate inputs, and performing the final combining and splitting of the intermediate inputs by using a second stage interconnect parallel processor to create resulting outputs. The method further includes transmitting the resulting outputs out of the second stage at line speed.
Data migration in a storage array that includes a plurality of storage devices
Migrating data in a storage array that includes a plurality of storage devices and a plurality of write buffer devices, including: detecting, by the storage array, an occurrence of a write buffer device evacuation event associated with one or more source write buffer devices; responsive to detecting the occurrence of the write buffer device evacuation event, determining, by the storage array, whether the storage array includes at least a predetermined amount of write buffer resources in addition to the one or more source write buffer devices; and responsive to determining that the storage array includes at least a predetermined amount of write buffer resources in addition to the one or more source write buffer devices, reducing, by the storage array, write access to the one or more source write buffer devices.
Data migration in a storage array that includes a plurality of storage devices
Migrating data in a storage array that includes a plurality of storage devices and a plurality of write buffer devices, including: detecting, by the storage array, an occurrence of a write buffer device evacuation event associated with one or more source write buffer devices; responsive to detecting the occurrence of the write buffer device evacuation event, determining, by the storage array, whether the storage array includes at least a predetermined amount of write buffer resources in addition to the one or more source write buffer devices; and responsive to determining that the storage array includes at least a predetermined amount of write buffer resources in addition to the one or more source write buffer devices, reducing, by the storage array, write access to the one or more source write buffer devices.
SYSTEM AND METHOD FOR PREVENTING CACHE CONTENTION
A system and method for preventing cache contention for a cache including a plurality of ways and a separate port for each way, the method including: obtaining, in a core of a processor, a multidimensional coefficient array of a multidimensional filter, and pointers to data elements from a plurality of rows of a multidimensional data array, and loading the plurality of rows into the cache, where each row is stored in a different way of the cache.