Patent classifications
G06F16/2272
Data Layout Model Generation System
A data layout model generation system generates, with reinforcement learning, a node configuration and a data layout key in a distributed parallel database. This system includes a sample acquisition processor that acquires, on the basis of a predetermined acquisition method, sample data from data stored in the distributed parallel database, a data layout estimator having, as states in the reinforcement learning, the node configuration and the data layout key including information regarding an order of sorting columns that constitute the data and information regarding a method for distribution between nodes, the data layout estimator estimating layout of the data on the basis of the state and the sample data, a reward calculator that calculates a reward in the reinforcement learning on the basis of a result obtained by estimating the layout of the data, the node configuration, and a processing cost of a query executed on the distributed parallel database.
Computer system and method of presenting information related to basis of predicted value output by predictor
There is provided is a computer system that outputs a predicted value of data to be evaluated using a predictor generated using learning data. The computer system includes the predictor; an index calculation unit that calculates an interpretation index of the data to be evaluated; and an extraction unit that selects the learning data useful for a user to interpret the predicted value of the data to be evaluated, wherein index management information for managing an interpretation index of the learning data is stored, the index calculation unit calculates the interpretation index of the data to be evaluated, and the extraction unit calculates a selection index based on the interpretation index of the data to be evaluated and the interpretation index of the learning data, selects the learning data based on the selection index, and outputs display information for presenting information indicating a processing result.
Resolving opaqueness of complex machine learning applications
Computing systems and technical methods that transform data structures and pierce opacity difficulties associated with complex machine learning modules are disclosed. Advances include a framework and techniques that include: i) global diagnostics; ii) locally interpretable models LIME-SUP-R and LIME-SUP-D; and iii) explainable neural networks. Advances also include integrating LIME-SUP-R and LIME-SUP-D approaches that create a transformed data structure and replicated modeling over local and global effects and that yield high interpretability along with high accuracy of the replicated complex machine learning modules that make up a machine learning application.
Component-based synchronization of digital assets
The present disclosure relates to a digital asset synchronization system that provides improved local and remote synchronization of digital assets. In particular, the digital asset synchronization system manages digital assets by separating each digital asset into multiple components stored as a set of distributed individual files. Employing individual components for a digital asset rather than single monolithic file enables the digital asset synchronization system to provide safe concurrent access to the digital asset from multiple applications on the same device and across different devices. In addition, using components for a digital asset provides the digital asset synchronization system with the ability to efficiently store and synchronize multiple versions of the digital asset, both locally and remotely.
Efficient in-memory multi-version concurrency control for a trie data structure based database
The invention describes a method for determining a storage location of a database object of a specific version, wherein indexes for each version of the database object are stored in a trie having a root node corresponding to the specific version, the method comprising: determining a trie corresponding to the specific version by accessing the root node of the trie corresponding to the specific version; determining an object identifier of the database object by traversing the trie corresponding to the specific version using a secondary key related to the database object as search key; determining the storage location of the database object by traversing the trie corresponding to the specific version using the determined object identifier as search key.
System and method for data visualization using machine learning and automatic insight of outliers associated with a set of data
In accordance with various embodiments, described herein are systems and methods for use of computer-implemented machine learning to automatically determine insights of facts, segments, outliers, or other information associated with a set of data, for use in generating visualizations of the data. In accordance with an embodiment, the system can use a machine learning process to automatically determine one or more outliers or findings within the data, based on, for example, determining a plurality of combinations representing pairs of attribute dimensions within a data set, from which a general explanation or pattern can be determined for one or more attributes, and then comparing particular values for attributes, with the determined pattern for those attributes. Information describing such outliers or findings can be graphically displayed at a user interface, as text, graphs, charts, or other types of visualizations, and used as a starting point for further analysis of the data set.
Independent datastore in a network routing environment
Systems, methods, and devices for offloading network data to a datastore. A system includes a publisher device in a network computing environment. The system includes a subscriber device in the network computing environment. The system includes a datastore independent of the publisher device and the subscriber device, the datastore comprising one or more processors in a processing platform configurable to execute instructions stored in non-transitory computer readable storage media. The instructions includes receiving data from the publisher device. The instructions include storing the data across one or more of a plurality of shared storage devices. The instructions include providing the data to the subscriber device.
Generating compact data structures for monitoring data processing performance across high scale network infrastructures
A compact data structure generation engine can be used to generate a compact data structure that represents performance data for high-scale networks. The compact data structure representing the performance data can be used to monitor the operation performed on or by a computer system to identify potentially anomalous conditions. In response, a corrective action can be taken to address the issue. This can be useful, for example, in improving the efficiency, effectiveness, and reliability of the computer system during operation.
Implementing a type restriction that restricts to a non-polymorphic layout type or a maximum value
A type restriction contextually modifies an existing type descriptor. The type restriction is imposed on a data structure to restrict the values that are assumable by the data structure. The type restriction does not cancel or otherwise override the effect of the existing type descriptor on the data structure. Rather the type restriction may declare that a value of the data structure's type is forbidden for the data structure. Additionally or alternatively, the type restriction may declare that an element count allowable for a data structure's type is forbidden for the data structure. Type restriction allows optionality (where only a singleton value for a data structure is allowed), empty sets (where no value for a data structure is allowed), and multiplicity (where only a limited element count for a data structure) to be injected into a code set independent of data type. Type restriction allows certain optimizations to be performed.
VERIFYING DATA CONSISTENCY USING VERIFIERS IN A CONTENT MANAGEMENT SYSTEM FOR A DISTRIBUTED KEY-VALUE DATABASE
A consistency verification system that verifies data consistency in a content item management system. The system maintains a plurality of verifiers for checking data consistency, each verifier comprising instructions for verifying consistency for a type of requests. The system may verify the original request by selecting a verifier from the plurality of verifiers based on a type of the original request. Using the selected verifier, the consistency verification system may send verification requests to the content item management system at a second timestamp that is a period of time after the original timestamp. The system may determine whether the second response is consistent with the first response. If the responses are not consistent, the system may output information describing data inconsistency. If the responses are consistent, the system may output results and perform further verifications to help ensure data accuracy.