Patent classifications
G06F2212/622
SYSTEMS AND METHODS FOR SIMULATING WORST-CASE CONTENTION TO DETERMINE WORST-CASE EXECUTION TIME OF APPLICATIONS EXECUTED ON A PROCESSOR
Techniques for determining worst-case execution time for at least one application under test are disclosed using memory thrashing. Memory thrashing simulates shared resource interference. Memory that is thrashed includes mapped memory, and optionally shared cache memory.
Multiple location index
Systems and methods for accessing data stored in multiple locations. A cache and a storage system are associated with an index. Entries in the index identify locations of data in both the cache and the storage system. When an index lookup occurs and an entry in the index identifies at least two locations for the data, the locations are ordered based on at least one factor and the data stored in the optimal location as determined from the at least one factor is returned.
STACKED MEMORY DEVICE SYSTEM INTERCONNECT DIRECTORY-BASED CACHE COHERENCE METHODOLOGY
A system includes a plurality of host processors and a plurality of hybrid memory cube (HMC) devices configured as a distributed shared memory for the host processors. An HMC device includes a plurality of integrated circuit memory die including at least a first memory die arranged on top of a second memory die, and at least a portion of the memory of the memory die is mapped to include at least a portion of a memory coherence directory; and a logic base die including at least one memory controller configured to manage three-dimensional (3D) access to memory of the plurality of memory die by at least one second device, and logic circuitry configured to implement a memory coherence protocol for data stored in the memory of the plurality of memory die.
High performance interconnect physical layer
Re-initialization of a link can take place without termination of the link, where the link includes, a transmitter and a receiver are to be coupled to each lane in the number of lanes, and re-initialization of the link is to include transmission of a pre-defined sequence on each of the lanes.
Page-based memory operation with hardware initiated secure storage key update
Methods and systems for secure storage protection for memory operations are provided. Aspects include providing a drawer comprising a plurality of clusters, each of the plurality of clusters comprising a plurality of processors, wherein each of the plurality of clusters share a first cache memory, providing a cluster shared cache integrated circuit to manage a second cache memory shared among the plurality of clusters, providing a system memory associated with each of the plurality of clusters, receiving, by a memory controller, a memory operation request from one of the plurality of processors, wherein the memory operation includes a store command, and wherein the memory controller is configured to perform the memory operation and atomically write a secure storage key for the memory operation with the store command of the memory operation.
REGION BASED SPLIT-DIRECTORY SCHEME TO ADAPT TO LARGE CACHE SIZES
Systems, apparatuses, and methods for maintaining region-based cache directories split between node and memory are disclosed. The system with multiple processing nodes includes cache directories split between the nodes and memory to help manage cache coherency among the nodes' cache subsystems. In order to reduce the number of entries in the cache directories, the cache directories track coherency on a region basis rather than on a cache line basis, wherein a region includes multiple cache lines. Each processing node includes a node-based cache directory to track regions which have at least one cache line cached in any cache subsystem in the node. The node-based cache directory includes a reference count field in each entry to track the aggregate number of cache lines that are cached per region. The memory-based cache directory includes entries for regions which have an entry stored in any node-based cache directory of the system.
Stacked memory device system interconnect directory-based cache coherence methodology
A system includes a plurality of host processors and a plurality of hybrid memory cube (HMC) devices configured as a distributed shared memory for the host processors. An HMC device includes a plurality of integrated circuit memory die including at least a first memory die arranged on top of a second memory die, and at least a portion of the memory of the memory die is mapped to include at least a portion of a memory coherence directory; and a logic base die including at least one memory controller configured to manage three-dimensional (3D) access to memory of the plurality of memory die by at least one second device, and logic circuitry configured to implement a memory coherence protocol for data stored in the memory of the plurality of memory die.
HIGH PERFORMANCE INTERCONNECT
- Robert J. Safranek ,
- Robert G. Blankenship ,
- Venkatraman Iyer ,
- Jeff Willey ,
- Robert Beers ,
- Darren S. Jue ,
- Arvind A. Kumar ,
- Debendra Das Sharma ,
- Jeffrey C. Swanson ,
- Bahaa Fahim ,
- Vedaraman Geetha ,
- Aaron T. Spink ,
- Fulvio Spagna ,
- Rahul R. Shah ,
- Sitaraman V. Iyer ,
- William Harry Nale ,
- Abhishek Das ,
- Simon P. Johnson ,
- Yuvraj S. Dhillon ,
- Yen-Cheng Liu ,
- Raj K. Ramanujan ,
- Robert A. Maddox ,
- Herbert H. Hum ,
- Ashish Gupta
A physical layer (PHY) is coupled to a serial, differential link that is to include a number of lanes. The PHY includes a transmitter and a receiver to be coupled to each lane of the number of lanes. The transmitter coupled to each lane is configured to embed a clock with data to be transmitted over the lane, and the PHY periodically issues a blocking link state (BLS) request to cause an agent to enter a BLS to hold off link layer flit transmission for a duration. The PHY utilizes the serial, differential link during the duration for a PHY associated task selected from a group including an in-band reset, an entry into low power state, and an entry into partial width state
Configuration based cache coherency protocol selection
Topology of clusters of processors of a computer configuration, configured to support any of a plurality of cache coherency protocols, is discovered at initialization time to determine which one of the plurality of cache coherency protocols is to be used to handle coherency requests of the configuration.
Using post-cache edge computing to re-populate dynamic content in cached content
A method that receives, at a first compute server of a plurality of compute servers, a request from a client device, wherein the request is a request for a network resource. The method locates at least one content item in response to the request for the network resource, detects possible dynamic content in the at least one content item, compares a first dynamic value in a first copy of the at least one content item to a second dynamic value in a second copy of the at least one content item from an origin server, and stores the at least one content item in a local cache as a safe content item to be returned for subsequent requests of the at least one content item in response to the first dynamic value matching the second dynamic value.