Patent classifications
G06V10/7747
METHOD AND SYSTEM FOR EXTRACTING SENTIMENTS OR MOOD FROM ART IMAGES
A method for extracting sentiments or mood from art images includes: receiving at least one of the art images as an input image; preprocessing the input image; extracting features from the preprocessed input image, the extracting including predicting a color label corresponding to a dominant perceptual color detected from the preprocessed input image a dominant subject from the preprocessed input image, detecting low-level image features from the preprocessed input image, and extracting mood feature information based on a description information included in the input image; classifying the extracted features into a plurality of mood/sentiments classes, using an artificial neural network; and predicting at least one of a mood or a sentiment that is present in the input image based on the dominant perceptual color and the plurality of mood/sentiments classes.
AR BODY PART TRACKING SYSTEM
Aspects of the present disclosure involve a system for presenting AR items. The system performs operations including: receiving an image that includes a depiction of a first real-world body part in a real-world environment; applying a machine learning technique to the image to generate a plurality of dense outputs each associated with a respective pixel of a plurality of pixels in the image; applying a first task-specific decoder to the plurality of dense outputs to identify a pixel corresponding to a center of the first real-world body part; applying a second task-specific decoder using the identified pixel to retrieve a 3D rotation, translation and scale of first real-world body part from the plurality of dense outputs; modifying an AR object based on the 3D rotation, translation, and scale of first real-world body part; and modifying the image to include a depiction of the modified AR object.
INFORMATION PROCESSING DEVICE AND INFORMATION PROCESSING METHOD
In an information processing device (a server device), a first acquirer acquires multiple captured images of the outside of a vehicle. A detector detects, from acquired multiple captured images, multiple images related to unsafe driving using a learned model. An image extractor extracts, from detected multiple images related to unsafe driving, an image to be a candidate for relearning data of the learned model.
DOMAIN BATCH BALANCING FOR TRAINING A MACHINE LEARNING ALGORITHM WITH DATA FROM MULTIPLE DATASETS
The subject disclosure relates to techniques for maintaining an optimized mix of samples selected from a smaller data set and a larger data set for use in training a machine learning algorithm. A process of the disclosed technology can include creating a plurality of chunks of samples, wherein each chunk contains a predetermined proportion of samples from the smaller data set and the larger data set, while also including substantially all samples in the larger dataset are distributed across the plurality of chunks of samples, loading a first chunk of samples of the plurality of chunks of samples into a memory, randomizing the order of samples in the first chunk of samples, and providing the samples in the first chunk of samples in the randomized order into the machine learning algorithm.
IDENTIFYING OVERFILLED CONTAINERS
Among other things, the techniques described herein include a method for receiving a plurality of images of one or more containers while the one or more containers are being emptied, the plurality of images comprising a training set of images and a validation set of images; labeling each image of the plurality of images as including either an overfilled container or a not-overfilled container; processing each image of the plurality of images to reduce bias of a machine learning model; training, and based on the labeling, the machine learning model using the plurality of images; and optimizing the machine learning model by performing learning against the validation set, the optimized machine learning model being used to generate a prediction for a new image of a container, the prediction indicating whether the container in the new image was overfilled prior to the new container being emptied.
Prediction error scenario mining for machine learning models
Provided are methods for prediction error scenario mining for machine learning methods, which can include determining a prediction error indicative of a difference between a planned decision of an autonomous vehicle and an ideal decision of the autonomous vehicle. The prediction error is associated with an error-prone scenario for which a machine learning model of an autonomous vehicle is to make planned movements. The method includes searching a scenario database for the error-prone scenario based on the prediction error. The scenario database includes a plurality of datasets representative of data received from an autonomous vehicle sensor system in which the plurality of datasets is marked with at least one attribute of the set of attributes. The method further includes obtaining the error-prone scenario from the scenario database for inputting into the machine learning model for training the machine learning model. Systems and computer program products are also provided.
METHOD AND APPARATUS FOR UPDATING OBJECT RECOGNITION MODEL
This application provides a method and apparatus for updating an object recognition model in the field of artificial intelligence. In the technical solution provided in this application, a target image and first voice information of a user are obtained. The first voice information indicates a first category of a target object in the target image. A feature library of a first object recognition model is updated based on the target image and the first voice information. The updated first object recognition model includes a feature of the target object and a first label indicating the first category, and the feature of the target object corresponds to the first label. A recognition rate of an object recognition model can be improved more easily according to the technical solution provided in this application.
CROWD MOTION SIMULATION METHOD BASED ON REAL CROWD MOTION VIDEOS
A crowd motion simulation method is provided based on real crowd motion videos. The method includes framing the videos and storing the framed videos into continuous high-definition images, generating a crowd density map of each image, and accurately positioning an individual in each density map to obtain the accurate position of each individual. The method also includes correlating the positions of each individual in different images to form a complete motion trajectory, and extracting motion trajectory data; and quantifying motion trajectory data, defining training data and data labels, and calculating data correlation. The method further includes building a deep convolutional neural network, and inputting the motion trajectory data for training to learn crowd motion behaviors; and randomly placing a plurality of simulation individuals in a two-dimensional space, testing a prediction effect of the deep convolutional neural network, adjusting parameters for simulation, and drawing a crowd motion trajectory.
IMAGE REPRESENTATION LEARNING IN DIGITAL PATHOLOGY
Described herein are systems, methods, and programming for analyzing and classifying digital pathology images. Some embodiments include receiving whole slide images (WSIs) and dividing each of the WSIs into tiles. For each WSI, a random subset of the tiles may be selected and augmented views of each of the selected tiles may be generated. For each of the selected tiles, a first convolutional neural network (CNN) may be trained to: generate, using a first one of the augmented views corresponding to the selected tile, a first representation of the selected tile, and predict a second representation of the selected tile to be generated by a second CNN, wherein the second representation is generated based on a second one of the augmented views of the selected tile.
METHOD AND SYSTEM FOR UNIQUE, PROCEDURALLY GENERATED DIGITAL OBJECTS VIA FEW-SHOT MODEL
Disclosed herein is digital object generator that makes uses a one-way function to generate unique digital objects based on the user specific input. Features of the input are first extracted via a few-shot convolutional neural network model, then evaluated weight and integrated fit. The resulting digital object includes a user decipherable output such as a visual representation, an audio representation, or a multimedia representation that includes recognizable elements from the user specific input.