G06F12/0615

Systems and methods in a graphics environment for providing shared virtual memory addressing support for a host system

Systems and methods for providing shared virtual memory addressing support for a host system are disclosed. In one embodiment, a graphics processor includes processing resources to perform graphics operations. A memory management unit (MMU) is coupled to the processing resources. The MMU to support a first virtual address size for managing allocation of non-shared virtual memory and to support a second virtual address size for managing allocation of shared virtual memory that is shared between the graphics processor and a host.

Solid state storage device with variable logical capacity based on memory lifecycle
11704025 · 2023-07-18 · ·

Several embodiments of memory devices and systems having a variable logical memory capacity are disclosed herein. In one embodiment, a memory device can include a plurality of memory regions that collectively define a physical memory capacity and a controller operably coupled to the plurality of memory regions. The controller is configured to advertise a first logical memory capacity to a host device, determine that at least one of the memory regions is at or near end of life, and in response to the determination—send a notification to the host device that a logical memory capacity of the memory device will be reduced and then retire the at least one of the memory regions.

Multi-ring shared, traversable, and dynamic advanced database
11556468 · 2023-01-17 · ·

Examples of the present disclosure describe systems and methods for sharing memory using a multi-ring shared, traversable and dynamic database. In aspects, the database may be synchronized and shared between multiple processes and/or operation mode protection rings of a system. The database may also be persisted to enable the management of information between hardware reboots and application sessions. The information stored in the database may be view independent, traversable, and resizable from various component views of the database. In some aspects, an event processor is additionally described. The event processor may use the database to allocate memory chunks of a shared heap to components/processes in one or more protection modes of the operating system.

ARTIFICIAL INTELLIGENCE CHIP AND DATA OPERATION METHOD

An artificial intelligence chip and a data operation method are provided. The artificial intelligence chip receives a command carrying first data and address information and includes a chip memory, a computing processor, a base address register, and an extended address processor. The base address register is configured to access an extended address space in the chip memory. The extended address processor receives the command. The extended address processor determines an operation mode of the first data according to the address information. When the address information points to a first section of the extended address space, the extended address processor performs a first operation on the first data. When the address information points to a section other than the first section of the extended address space, the extended address processor notifies the computing processor of the operation mode and the computing processor performs a second operation on the first data.

MULTI-RING SHARED, TRAVERSABLE, AND DYNAMIC ADVANCED DATABASE
20230129140 · 2023-04-27 ·

Examples of the present disclosure describe systems and methods for sharing memory using a multi-ring shared, traversable and dynamic database. In aspects, the database may be synchronized and shared between multiple processes and/or operation mode protection rings of a system. The database may also be persisted to enable the management of information between hardware reboots and application sessions. The information stored in the database may be view independent, traversable, and resizable from various component views of the database. In some aspects, an event processor is additionally described. The event processor may use the database to allocate memory chunks of a shared heap to components/processes in one or more protection modes of the operating system.

METHODS AND APPARATUS FOR CONTENT DELIVERY VIA BROWSER CACHE EXTENSION

Embodiments include methods, systems, and apparatuses for content delivery using shared caching, and more specifically, a browser cache extension (BCE) between a local browser cache and a remote cache located on an edge server. In an embodiment, a remote BCE function on the edge server may create a shared cache folder containing a remote cache and an remote cache index file. A local BCE function in the local browser may be able to access the shared cache folder via a network file system (NFS) protocol. The local BCE function may merge the remote index file with a local index file from the local browser and retrieve the remote cache at the local browser.

Indirect register access method and system

Systems and methods are provided for managing access to registers. In one embodiment, a system may include a processor and a plurality of registers. The processor and the plurality of registers may be integrated into a single device, or may be in separate devices. The plurality of registers may include a first set of registers that are directly accessible by the processor, and a second set of registers that are not directly accessible by the processor. The second set of registers may, however, be accessed indirectly by the processor via the first set of registers. In one embodiment, the first set of registers may include a register for selecting a register bank from the second set of registers, and a register for selecting a particular address within the register bank, to allow indirect access by the processor to the registers of the second set.

Semiconductor device capable of performing software lock-step

A semiconductor device performs a software lock-step. The semiconductor device includes a first circuit group including a first Intellectual Property (IP) to be operated in a first address space, a first bus, and a first memory, a second circuit group including a second IP to be operated in a second address space, a second bus, and a second memory, a third bus connectable to a third memory, and a transfer control circuit coupled to the first to third buses. when the software lock-step is performed, the second circuit group converts an access address from the second IP to the second memory such that an address assigned to the second memory in the second address space is a same as an address assigned to the first memory in the first address space.

PRIORITY-BASED ACCESS OF COMPRESSED MEMORY LINES IN MEMORY IN A PROCESSOR-BASED SYSTEM

Aspects disclosed relate to a priority-based access of compressed memory lines in a processor-based system. In an aspect, a memory access device in the processor-based system receives a read access request for memory. If the read access request is higher priority, the memory access device uses the logical memory address of the read access request as the physical memory address to access the compressed memory line. However, if the read access request is lower priority, the memory access device translates the logical memory address of the read access request into one or more physical memory addresses in memory space left by the compression of higher priority lines. In this manner, the efficiency of higher priority compressed memory accesses is improved by removing a level of indirection otherwise required to find and access compressed memory lines.

FUNCTION ANALYSIS METHOD AND MEMORY DEVICE
20170262366 · 2017-09-14 · ·

A function analysis method applied to a memory device includes analyzing the mapping relationship between at least one API function and at least one normal function, analyzing a calling relationship between the at least one normal function through the correspondence relationship, and developing a two-dimensional array to analyze whether there is a loop in the calling relationship or not. When one of the at least one normal function calls another normal function, whether or not the name of the called normal function is the same as the name of the normal function and names of the normal functions which call the normal function is checked.