G06F16/254

Artificially-intelligent, continuously-updating, centralized-database-identifier repository system

A centralized database identifier repository may identify databases using a unique identifier, or key tag, for each database. Each identified database may include data relating to one or more specific data elements. The repository may include a variety of data elements. Each data element may be associated with one or more database keys. The repository may be a repository of reference pointers. The repository may facilitate data viewing and data retrieval. A requestor may search for a data element using the centralized repository. The repository may retrieve data relating to a specific data element, from all databases identified by unique identifiers, that include data relating to the data element. The databases' unique identifiers may be encrypted tokens.

Insight expansion in smart data retention systems

A computer-implemented method applies insights from a variety of data sources to each of the data sources. The method includes identifying a set of data sources, wherein each of the data sources are associated with a domain. The method includes analyzing documentation for each of the data sources. The method further includes extracting a set of attributes for each data source, and determining a data schema associated with each data source. The method includes mapping each data schema to a common domain schema. The method also includes linking, based on the mapping and on the set of attributes for each data source, common features across each data source. The method includes generating, in response to the linking, a knowledge graph. The method further includes preparing a visual display for a set of domain insights; and forking the set of domain insights into a first data source.

Techniques for unifying ETL filter operators

Techniques are provided for unifying filter operators in exchange, transform, load (ETL) plans. Such a technique includes a method that may include receiving, by a computer system, an ETL plan including a split operator and a plurality of filter operators. The may include identifying, by the computer system, that the plurality of filter operators are configured to act on data output by the split operator in the ETL plan. The method may include generating, by the computer system, a unified filter operator using the plurality of filter operators. The method may include generating, by the computer system, an updated ETL plan comprising the unified filter operator providing filtered data to the split operator. The method may also include storing the updated ETL plan in a data store.

Key pair platform and system to manage federated trust networks in distributed advertising

Systems and methods are provided for object identifier translation using a key pairs platform in a virtualized or cloud-based computing system. A key pair refers to a pair of identifiers held by an entity. Each key pair includes at least one anonymized object identifier. Advantageously, the key pair system protects privacy and provides anonymity for objects by not disclosing the identity of the objects or the underlying data associated with the objects.

Dynamic updating of query result displays

Described are methods, systems and computer readable media for dynamic updating of query result displays.

A NOVEL ANTI-CD3/ANTI-EGFR BISPECIFIC ANTIBODY AND USES THEREOF
20230008090 · 2023-01-12 ·

Provided are bispecific antibodies against CD3 and EGFR, the nucleic acid molecules encoding the antibodies, expression vectors and host cells used for the expression of the antibodies. The antibodies provide a potent agent for the treatment of CD3-related and/or EGFR-related diseases via modulating immune functions.

Time-based partitioning to avoid in-place updates for data set copies

Time-based partitioning of a data set is applied to capture updates to the data set in a copy of the data set. Items that have been updated in a data set with in a time period are identified. Partitions of the data set that include the updated items are created according to a partitioning scheme. The created partitions are grouped in a storage location for the time period in a file structure that stores a copy of the database. A latest version of the copy of the data set may be accessed according to latest partitions of the data set stored in the different locations of the file structure.

Methods and systems for the execution of analysis and/or services against multiple data sources while maintaining isolation of original data source
11573973 · 2023-02-07 ·

Methods and systems for data are disclosed. A system implementation includes a data module for storing data received from an external source. The data module includes a file system for unstructured data, a database for structured data, a transform for operating upon unstructured or structured data, a data broker for receiving data having a first format and providing the data in a second format, a data network for communications within the data module, and a processing module for performing operations upon data. The processing module further includes a process broker and a process container. The process container is for providing one or more instances of processes during a runtime operation. The system further includes an inter-process network for communications within the processing module and an internal gateway for the data module to communicate with the processing module.

System and method for automatic correction/rejection in an analysis applications environment

Systems and methods for automatic error rejection are provided. Systems and methods described herein bypass the creation of a staging table at the outset and, instead, attempt a direct merge from a source data location to a target data location. In the event that the merge fails, then a temporary/staging table can be loaded where errors can be logged, validations can be performed, and erroneous data can be corrected.

Systems and methods for providing automated integration and error resolution of records in complex data systems

A claim editing engine for automated integration and error resolution of claim records is provided. The processor of the engine is configured to extract a set of claim components of a plurality of claim components. The processor is further configured to transform the set of claim components to conform to a standardized data format. The processor is also configured to integrate the set of transformed claim components into a set of unified claims by unifying each of the set of transformed claim components having matching claim identifiers into a unified claim. The processor is configured to apply a rule set to the set of unified claims to generate a simulation of execution of the set of claims and identify errors in the simulated execution. The processor is configured to transmit an instruction to resolve each identified error. The processor is configured to cause each resolved unified claim to be processed.