Patent classifications
G06F2212/45
Operating a FIFO memory
The present invention concerns a method of operating a first-in first-out memory (9) arranged to store measurement data samples measured by a plurality of data measurement sensors (1, 3, 5), which can operate at various sampling rates. The oldest measurement data sample in the memory (9) is arranged to be read first before the newer measurement data samples. The method comprises: receiving measurement data samples from at least two data measurement sensors (1, 3, 5); and saving the received measurement data samples in the memory (9). Each of the measurement data samples saved in the memory is associated with a tag which is also saved in the memory (9) and which identifies the data measurement sensor (1, 3, 5) which measured the respective measurement data sample.
Selecting a resource from a set of resources for performing an operation
The described embodiments comprise a selection mechanism that selects a resource from a set of resources in a computing device for performing an operation. In some embodiments, the selection mechanism performs a lookup in a table selected from a set of tables to identify a resource from the set of resources. When the resource is not available for performing the operation and until another resource is selected for performing the operation, the selection mechanism identifies a next resource in the table and selects the next resource for performing the operation when the next resource is available for performing the operation.
SYSTEMS AND METHODS FOR BOOTSTRAP MANAGEMENT
The present disclosure is directed techniques for bootstrap management. A method includes: upon an initial launch of an application on a client device, fetching, from a server and using a native component of the application, content for loading a web component of the application on the client device; determining whether a bootstrap management mode is enabled on the client device; and responsive to the bootstrap management mode is enabled and in response to the web component being launched: receiving, at the native component and from the web component, a manifest and a request for bootstrapping resources; caching, by the native component, the manifest from the web component; fetching, from the server and using the native component, the bootstrapping resources requested by the web component; caching, by the native component, the fetched bootstrapping resources in the memory; and providing, by the native component, the fetched bootstrapping resources to the web component.
INCREMENTAL STACK WALKING
Techniques for incremental stack walking are disclosed, including: performing a stack walk of a runtime stack, at least by traversing the runtime stack from a current frame to a root frame, to obtain a set of stack walking results; storing a cache of the set of stack walking results; and installing, on the runtime stack, a marker frame that marks a boundary of stack frames represented by the set of stack walking results.
Installing, updating or reconfiguring an application without disabling a write filter
An application can be installed, updated or reconfigured without disabling a write filter. When a package is to be deployed to a client terminal to install, update or reconfigure an application, an overlay optimizer can be instructed to start a session. During the session, the overlay optimizer can analyze I/O requests to identify any I/O request that pertains to the deployment of the package. The overlay optimizer can then redirect the identified I/O requests to a session overlay that the overlay optimizer has created for the session rather than passing the I/O requests to the write filter. As a result, the artifacts that are affected by the deployment of the package will be stored in the session overlay rather than the write filter's overlay. Once the session is completed, the overlay optimizer can copy the artifacts from the session overlay to the write filter's overlay and commit them.
Columnar Techniques for Big Metadata Management
A method for managing big metadata using columnar techniques includes receiving a query request requesting data blocks from a data table that match query parameters. The data table is associated with system tables that each includes metadata for a corresponding data block of the data table. The method includes generating, based on the query request, a system query to return a subset of rows that correspond to the data blocks that match the query parameters. The method further includes generating, based on the query request and the system query, a final query to return a subset of data blocks from the data table corresponding to the subset of rows. The method also includes determining whether any of the data blocks in the subset of data blocks match the query parameters, and returning the matching data blocks when one or more data blocks match the query parameters.
Configuration or data caching for programmable logic device
An integrated circuit device may cache configuration data to enable rapid configuration from fabric cache memory. The integrated circuit device may include programmable logic fabric having configuration memory and programmable logic elements controlled by the configuration memory, and sector-aligned memory apart from the programmable logic fabric. A first sector of the configuration memory may be programmed with first configuration data. The sector-aligned memory may include a first sector of sector-aligned memory that may cache the first configuration data while the configuration memory is programmed with the first configuration data a first time. A second sector of sector-aligned memory may cache second configuration data for a second sector of the configuration memory in parallel while the first sector of sector-aligned memory caches the first configuration data for the first sector of the configuration memory.
COMPILE TIME LOGIC FOR INSERTING A BUFFER BETWEEN A PRODUCER OPERATION UNIT AND A CONSUMER OPERATION UNIT IN A DATAFLOW GRAPH
A dataflow graph for an application has operation units that are configured to be producers and consumers of tensors. A write access pattern of a particular producer specifies an order in which the particular producer generates elements of a tensor, and a read access pattern of a corresponding consumer specifies an order in which the corresponding consumer processes the elements of the tensor. The technology disclosed detects conflicts between the producers and the corresponding consumers that have mismatches between the write access patterns and the read access patterns. A conflict occurs when the order in which the particular producer generates the elements of the tensor is different from the order in which the corresponding consumer processes the elements of the tensor. The technology disclosed resolves the conflicts by inserting buffers between the producers and the corresponding consumers.
BLOCKCHAIN MACHINE NETWORK ACCELERATION ENGINE
Embodiments herein describe a hardware accelerator (e.g., a network acceleration engine) for a blockchain machine or node. The hardware accelerator parses packets containing separate components of a block of transactions to generate data to perform a validation process. To avoid the latency that comes with using software, the embodiments herein describe a protocol processor in the hardware accelerator that parses the packets and prepares the data so it can be consumed by downstream components in the accelerator without software intervention. These downstream components can then perform a validation operation to validate one or more transactions before those transactions are committed (i.e., added) to a ledger of a permissioned blockchain.
Memory Circuit and Cache Circuit Configuration
A memory circuit includes a stack of first dies including multiple sets of memory cells of a first type, a second die including multiple sets of memory cells of a second type, a third die, and an interposer carrying the first, second, and third dies. The second die includes a first set of input/output (I/O) terminals on a top surface of the second die and a second set of I/O terminals on a bottom surface of the second die. The stack of first dies is coupled to the second die through the first set of I/O terminals. The interposer is coupled to the second die through the second set of I/O terminals. The third die is positioned aside the second die and in communication with the second die through the interposer.