Patent classifications
G06F16/1865
SYSTEM AND METHOD FOR AN ULTRA HIGHLY AVAILABLE, HIGH PERFORMANCE, PERSISTENT MEMORY OPTIMIZED, SCALE-OUT DATABASE
A shared-nothing database system is provided in which parallelism and workload balancing are increased by assigning the rows of each table to “slices”, and storing multiple copies (“duplicas”) of each slice across the persistent storage of multiple nodes of the shared-nothing database system. When the data for a table is distributed among the nodes of a shared-nothing system in this manner, requests to read data from a particular row of the table may be handled by any node that stores a duplica of the slice to which the row is assigned. For each slice, a single duplica of the slice is designated as the “primary duplica”. All DML operations (e.g. inserts, deletes, updates, etc.) that target a particular row of the table are performed by the node that has the primary duplica of the slice to which the particular row is assigned. The changes made by the DML operations are then propagated from the primary duplica to the other duplicas (“secondary duplicas”) of the same slice.
A SYSTEM AND A METHOD FOR PERFORMING ATOMIC SWAP TRANSACTIONS OF DIGITIAL RECORDS AMONG A PLURALITY OF DITRIBUTED DATABASES
The present invention relates to a system and a method for performing exchanges of digital data records among a plurality of distributed databases. More specifically, the present invention relates to a technology agnostic atomic swap system platform configured to perform atomic swap transactions of digital data records among a plurality of distributed ledger technology platforms.
Method and apparatus for resolving source database precommitted transactions that are replicated to a target database of a database replication system
Source database precommitted transactions are resolved in a target database of a database replication system when selected source database precommitted transactions are subsequently aborted in the source database.
Preventing DBMS deadlock by eliminating shared locking
A DBMS receives a database-access request that includes an instruction to non-destructively read a database table row. The DBMS assigns the request a TSN identifier and creates a TSN image that identifies all TSNs assigned to transactions that are not yet committed. The DBMS traverses a linked list of log entries that identifies a chronological history of transactions performed on the same row. The DBMS infers that the table row currently contains data stored by the most recently logged transaction that is not contained in the TSN image and that has thus been committed. The DBMS then continues to process statements of the transaction based on the assumption that the row contains the inferred value. The DBMS performs this procedure without acquiring a shared lock on the data page or on the index leaf page that points to the table row.
Transaction-enabled systems and methods for royalty apportionment and stacking
Transaction-enabled systems and methods for royalty apportionment and stacking are disclosed. An example system may include a plurality of royalty generating elements (a royalty stack) each related to a corresponding one or more of a plurality of intellectual property (IP) assets (an aggregate stack of IP). The system may further include a royalty apportionment wrapper to interpret IP licensing terms and apportion royalties to a plurality of owning entities corresponding to the aggregate stack of IP in response to the IP licensing terms and a smart contract wrapper. The smart contract wrapper is configured to access a distributed ledger, interpret an IP description value and IP addition request, to add an IP asset to the aggregate stack of IP, and to adjust the royalty stack.
Outputting map-reduce jobs to an archive file
Method and system are provided for writing output from map-reduce jobs to an archive file. The method may include providing an archive manager and exposing an interface to be called from map-reduce jobs to output to an archive file in a map-reduce distributed file system. The method may also include using a buffering database as a temporary cache to buffer updates to the archive file. Handling by the archive manager calls from map-reduce jobs may allow: reading directly from an archive file or from a job index at the buffering database; writing to a job index at the buffering database used as a temporary cache to buffer updates; and serializing updates from the buffering database to the archive file.
Transaction-enabling systems and methods for customer notification regarding facility provisioning and allocation of resources
The present disclosure describes transaction-enabling systems and methods. A system can include a facility including a core task including a customer relevant output and a controller. The controller may include a facility description circuit to interpret a plurality of historical facility parameter values and corresponding facility outcome values and a facility prediction circuit to operate an adaptive learning system, wherein the adaptive learning system is configured to train a facility production predictor in response to the historical facility parameter values and the corresponding outcome values. The facility description circuit also interprets a plurality of present state facility parameter values, wherein the trained facility production predictor determines a customer contact indicator in response to the plurality of present state facility parameter values and a customer notification circuit provides a notification to a customer in response.
System and method for an ultra highly available, high performance, persistent memory optimized, scale-out database
A shared-nothing database system is provided in which parallelism and workload balancing are increased by assigning the rows of each table to “slices”, and storing multiple copies (“duplicas”) of each slice across the persistent storage of multiple nodes of the shared-nothing database system. When the data for a table is distributed among the nodes of a shared-nothing system in this manner, requests to read data from a particular row of the table may be handled by any node that stores a duplica of the slice to which the row is assigned. For each slice, a single duplica of the slice is designated as the “primary duplica”. All DML operations (e.g. inserts, deletes, updates, etc.) that target a particular row of the table are performed by the node that has the primary duplica of the slice to which the particular row is assigned. The changes made by the DML operations are then propagated from the primary duplica to the other duplicas (“secondary duplicas”) of the same slice.
Hybrid approach to performing a lazy pull of container images
Embodiments of the present disclosure provide a hybrid approach to performing a lazy pull of a container image. A file system in user space (FUSE) is utilized to lazy pull the container image, and manage file requests from the container while the container image is being fetched locally. During the retrieving, the FUSE may receive from the container, one or more file requests, and may temporarily block each of the one or more file requests until it can process them. Once the container image is fully fetched locally, the overlay structure of the container image expected by a file system in the kernel (e.g., Overlay FS) is created and control is passed to the file system in the kernel. The FUSE may then unmount itself, to expose the container to the underlying mount point.
DATA PROCESSING METHOD AND APPARATUS FOR VIRTUAL RESOURCE
Aspects of the disclosure provide a data processing apparatus having interface circuitry and processing circuitry. The interface circuitry is configured to obtain data related to at least two resource accounts, and related to resource transfer of the at least two resource account. The processing circuitry configured to construct resource transfer relationships of the at least two resource accounts according to the data, calculate, based on the resource transfer relationships, similarity measurements between the at least two resource accounts according to a preset calculation rule, and output, according to the calculated similarity measurements, a resource account with a similarity measurement satisfying a preset condition.