G06N3/0895

SYSTEM AND METHOD FOR UNSUPERVISED LEARNING OF SEGMENTATION TASKS
20230050573 · 2023-02-16 ·

Apparatuses and methods are provided for training a feature extraction model determining a loss function for use in unsupervised image segmentation. A method includes determining a clustering loss from an image; determining a weakly supervised contrastive loss of the image using cluster pseudo labels based on the clustering loss; and determining the loss function based on the clustering loss and the weakly supervised contrastive loss.

DATA PROCESSING METHOD AND APPARATUS
20230048031 · 2023-02-16 ·

Relating to the field of artificial intelligence, and specifically relating to the field of natural language processing, a data processing method includes and an apparatus performs: determining original text samples, where masking processing is not performed on the original text samples; and performing mask processing on the original text samples to obtain mask training samples, where the mask processing makes mask proportions of the mask training samples unfixed, and the mask training samples each are used to train a pretrained language model PLM. Training the PLM by using the mask training samples whose mask proportions are unfixed can enhance mode diversity of the training samples of the PLM. Therefore, features learned by the PLM are also diversified, a generalization capability of the PLM can be improved, and a natural language understanding capability of the PLM obtained through training can be improved.

MOLECULAR GRAPH REPRESENTATION LEARNING METHOD BASED ON CONTRASTIVE LEARNING
20230052865 · 2023-02-16 ·

The present invention is a molecular graph representation learning method based on contrastive learning, the method comprising: obtaining a molecular fingerprint representation of each molecule, and calculating a similarity between each two molecular fingerprints; collecting a full amount of chemical functional group information, and matching a corresponding functional group for each atom in the molecule; using a heterogeneous graph to model a molecular graph; using a RGCN in the structure-aware molecular encoder to encode the representation of each atom in the molecule and the representation of the functional group to which the atom belongs, and mapping the molecule to a feature space through an aggregation function to obtain a structure-aware feature representation; according to the fingerprint similarity between molecules, selecting positive and negative samples, and carrying out a comparative learning in the feature space; obtaining the structure-aware molecular encoder by using the contrastive learning method for training on a large-sample molecular dataset, and applying the structure-aware molecular encoder to a prediction task of downstream molecular attributes. The present invention helps to capture more abundant molecular structure information and solve the problem on molecular property prediction.

MOVEMENT DATA FOR FAILURE IDENTIFICATION

Configurations for data center component monitoring are disclosed. In at least one embodiment, movement of a server component is determined based on sensor data and the movement is used to diagnose a root cause for a server component failure.

DATA RETRIEVAL USING REINFORCED CO-LEARNING FOR SEMI-SUPERVISED RANKING
20230053009 · 2023-02-16 ·

A computer-implement method comprises: training a classifier with labeled data from a dataset; classifying, by the trained classifier, unlabeled data from the dataset; providing, by the classifier to a policy gradient, a reward signal for each data/query pair; transferring, by the classifier to a ranker, learning; training, by the policy gradient, the ranker; ranking data from the dataset based on a query; and retrieving data from the ranked data in response to the query.

COUNTERFACTUAL SELF-TRAINING
20230045950 · 2023-02-16 ·

A method of using a computing device to self-train a machine learning model with an incomplete dataset including original observational data. The method includes receiving a labeled training data, the labeled training data for training a machine learning model. Counterfactual unlabeled training data is received. One or more labels are predicted for the counterfactual unlabeled training data. The machine learning model is trained based upon the labeled training data, the counterfactual unlabeled training data, and the predicted one or more labels for the unlabeled training data. The machine learning model reduces bias in original observational data. An evaluation of the predicted one or more labels is received based on corresponding artificial intelligence explanations provided by an artificial intelligence explainability model.

Performance of Complex Optimization Tasks with Improved Efficiency Via Neural Meta-Optimization of Experts
20230040793 · 2023-02-09 ·

Example systems perform complex optimization tasks with improved efficiency via neural meta-optimization of experts. In particular, provided is a machine learning framework in which a meta-optimization neural network can learn to fuse a collection of experts to provide a predicted solution. Specifically, the meta-optimization neural network can learn to predict the output of a complex optimization process which optimizes over outputs from the collection of experts to produce an optimized output. In such fashion, the meta-optimization neural network can, after training, be used in place of the complex optimization process to produce a synthesized solution from the experts, leading to orders of magnitude faster and computationally more efficient prediction or problem solution.

RAY CLUSTERING LEARNING METHOD BASED ON WEAKLY-SUPERVISED LEARNING FOR DENOISING THROUGH RAY TRACING
20230037418 · 2023-02-09 ·

Disclosed is a ray clustering learning method based on weakly-supervised learning for denoising using ray tracing. The ray clustering learning method is for learning a denoising model for removing noise from a rendered image through ray tracing, and includes extracting a feature of a simulated ray through the ray tracing and clustering the ray through contrastive learning for the feature.

METHOD, DEVICE, AND COMPUTER PROGRAM PRODUCT FOR IMAGE RECOGNITION
20230038047 · 2023-02-09 ·

Embodiments of the present disclosure relate to a method, a device, and a computer program product for image recognition. In some embodiments, characterization information for a first reference image in a reference image set is generated in an image recognition engine by using a Gaussian mixture model. First reference label information for the first reference image is generated based on the characterization information for the first reference image, the first reference label information being associated with a category of a first object in the first reference image. The image recognition engine is updated by determining the accuracy of the first reference label information for the first reference image. In this way, good characterization of images and generation of reference label information for the images can be achieved, thus both improving the robustness of the generated reference label information and significantly improving the accuracy of image recognition.

SELF-SUPERVISED LEARNING WITH MODEL AUGMENTATION
20230042327 · 2023-02-09 ·

A method for providing a neural network system includes performing contrastive learning to the neural network system to generate a trained neural network system. The performing the contrastive learning includes performing first model augmentation to a first encoder of the neural network system to generate a first embedding of a sample, performing second model augmentation to the first encoder to generate a second embedding of the sample, and optimizing the first encoder using a contrastive loss based on the first embedding and the second embedding. The trained neural network system is provided to perform a task.