Patent classifications
G06N3/126
Adaptive sampling for imbalance mitigation and dataset size reduction in machine learning
According to an embodiment, a method includes generating a first dataset sample from a dataset, calculating a first validation score for the first dataset sample and a machine learning model, and determining whether a difference in validation score between the first validation score and a second validation score satisfies a first criteria. If the difference in validation score does not satisfy the first criteria, the method includes generating a second dataset sample from the dataset. If the difference in validation score does satisfy the first criteria, the method includes updating a convergence value and determining whether the updated convergence value satisfies a second criteria. If the updated convergence value satisfies the second criteria, the method includes returning the first dataset sample. If the updated convergence value does not satisfy the second criteria, the method includes generating the second dataset sample from the dataset.
Machine learning to identify individuals for a therapeutic intervention provided using digital devices
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for using machine learning to generate precision predictions of readiness. In some implementations, a database is accessed to obtain status data that indicates activities or attributes of a subject. A set of feature scores is derived from the status data for the subject, the set of feature scores including values indicative of attributes or activities of the subject. The set of feature scores to one or more models that have been configured to predict readiness of subjects to satisfy one or more readiness criteria. The one or models can be models configured using machine learning training. Based on processing performed using the one or more machine learning models and the set of feature scores, a prediction regarding the subject's ability to achieve readiness to satisfy the one or more readiness criteria is generated.
Resource-aware automatic machine learning system
The present disclosure relates to a system, a method, and a product for optimizing hyper-parameters for generation and execution of a machine-learning model under constraints. The system includes a memory storing instructions and a processor in communication with the memory. When executed by the processor, the instructions cause the processor to obtain input data and an initial hyper-parameter set; for an iteration, to build a machine learning model based on the hyper-parameter set, evaluate the machine learning model based on the target data to obtain a performance metrics set, and determine whether the performance metrics set satisfies the stopping criteria set. If yes, the instructions cause the processor to perform an exploitation process to obtain an optimal hyper-parameter set, and exit the iteration; if no, perform an exploration process to obtain a next hyper-parameter set, and perform a next iteration with using the next hyper-parameter set as the hyper-parameter set.
Learning method, management device, and management program
There is provided a learning method. The method includes performing preprocessing on light emission data in a chamber of a plasma processing apparatus, setting a constraint for generating a regression equation representing a relationship between an etching rate of the plasma processing apparatus and the light emission data, selecting a learning target wavelength from the light emission data subjected to the preprocessing, and receiving selection of other sensor data different from the light emission data. The method further includes generating a regression equation based on the set constraint while using, as learning data, the selected wavelength, the received other sensor data, and the etching rate, and outputting the generated regression equation.
METHOD FOR AUTOMATED ENSEMBLE MACHINE LEARNING USING HYPERPARAMETER OPTIMIZATION
A method for a hyperparameter optimization for an automated ensemble machine learning model includes: generating an initial population of a plurality of machine learning (ML) models with a plurality of randomly chosen hyperparameters; calculating a loss function for each of the plurality of machine learning models; creating a new population of ML models and generating a base learner model using the hyperparameters of the best model. The method for creating the new population include the steps of: (a) selecting multiple best models with least errors as parents from a previous generation; (b) creating an offspring of the new population of ML models with a crossover probability and a mutation probability; and (c) repeating the steps (a) and (b) until a number of generations is reached and reporting the hyperparameters of the best model.
METHOD FOR AUTOMATED ENSEMBLE MACHINE LEARNING USING HYPERPARAMETER OPTIMIZATION
A method for a hyperparameter optimization for an automated ensemble machine learning model includes: generating an initial population of a plurality of machine learning (ML) models with a plurality of randomly chosen hyperparameters; calculating a loss function for each of the plurality of machine learning models; creating a new population of ML models and generating a base learner model using the hyperparameters of the best model. The method for creating the new population include the steps of: (a) selecting multiple best models with least errors as parents from a previous generation; (b) creating an offspring of the new population of ML models with a crossover probability and a mutation probability; and (c) repeating the steps (a) and (b) until a number of generations is reached and reporting the hyperparameters of the best model.
System and methods for search engine parameter tuning using genetic algorithm
A method for operating a search engine may include determining a multi-dimensional search parameter space comprising a set of possible weight values for each of a plurality of search parameters and dividing the search parameter space into a grid of evenly-spaced values that is a subset of the set of possible values. The method may further include defining one or more initial populations of search parameter weight values, wherein each population of search parameter weight values comprises a plurality of initial individuals, wherein each initial individual comprises a respective one of the evenly-spaced values for each of the search parameters. The method may further include executing one or more genetic algorithms based on the one or more initial populations to select a final set of search parameter weight values, and returning results of a user search in the search engine according to the final set of search parameter weight values.
System and methods for search engine parameter tuning using genetic algorithm
A method for operating a search engine may include determining a multi-dimensional search parameter space comprising a set of possible weight values for each of a plurality of search parameters and dividing the search parameter space into a grid of evenly-spaced values that is a subset of the set of possible values. The method may further include defining one or more initial populations of search parameter weight values, wherein each population of search parameter weight values comprises a plurality of initial individuals, wherein each initial individual comprises a respective one of the evenly-spaced values for each of the search parameters. The method may further include executing one or more genetic algorithms based on the one or more initial populations to select a final set of search parameter weight values, and returning results of a user search in the search engine according to the final set of search parameter weight values.
Machine learning for computing enabled systems and/or devices
Aspects of the disclosure generally relate to computing enabled systems and/or devices and may be generally directed to machine learning for computing enabled systems and/or devices. In some aspects, the system captures one or more digital pictures, receives one or more instruction sets, and learns correlations between the captured pictures and the received instruction sets.
Automatic feature selection and model generation for linear models
Methods, systems, and devices for automated feature selection and model generation are described. A device (e.g., a server, user device, database, etc.) may perform model generation for an underlying dataset and a specified outcome variable. The device may determine relevance measurements (e.g., stump R-squared values) for a set of identified features of the dataset and can reduce the set of features based on these relevance measurements (e.g., according to a double-box procedure). Using this reduced set of features, the device may perform a least absolute shrinkage and selection operator (LASSO) regression procedure to sort the features. The device may then determine a set of nested linear models—where each successive model of the set includes an additional feature of the sorted features—and may select a “best” linear model for model generation based on this set of models and a model quality criterion (e.g., an Akaike information criterion (AIC)).