Patent classifications
G06V10/7753
Surveillance system with activity recognition
A computer-implemented method, system, and computer program product are provided for activity recognition in a surveillance system. The method includes receiving a plurality of unlabeled videos from one or more cameras. The method also includes classifying an activity in each of the plurality of unlabeled videos. The method additionally includes controlling an operation of a processor-based machine to react in accordance with the activity.
Creative GAN generating art deviating from style norms
A method and system for generating art uses artificial intelligence to analyze existing art forms and then creates art that deviates from the learned styles. Known art created by humans is presented in digitized form along with a style designator to a computer for analysis, including recognition of artistic elements and association of particular styles. A graphics processor generates a draft graphic image for similar analysis by the computer. The computer ranks such draft image for correlation with artistic elements and known styles. The graphics processor modifies the draft image using an iterative process until the resulting image is recognizable as art but is distinctive in style.
Minimum-Example/Maximum-Batch Entropy-Based Clustering with Neural Networks
A computing system can include an embedding model and a clustering model. The computing system input each of the plurality of inputs into the embedding model and receiving respective embeddings for the plurality of inputs as outputs of the embedding model. The computing system can input the respective embeddings for the plurality of inputs into the clustering model and receiving respective cluster assignments for the plurality of inputs as outputs of the clustering model. The computing system can evaluate a clustering loss function that evaluates a first average, across the plurality of inputs, of a respective first entropy of each respective probability distribution; and a second entropy of a second average of the probability distributions for the plurality of inputs. The computing system can modify parameter(s) of one or both of the clustering model and the embedding model based on the clustering loss function.
Adversarial Defense Platform For Automated Dental Image Classification
Dental images are processed according to a first machine learning model to determine teeth labels. The teeth labels and image are concatenated and processed using a second machine learning model to label anatomy including CEJ, JE, GM, and Bone. The anatomy labels, teeth labels, and image are concatenated and processed using a third machine learning model to obtain feature measurements, such as pocket depth and clinical attachment level. The feature measurements, anatomy labels, teeth labels, and image may be concatenated and input to a fourth machine learning model to obtain a diagnosis for a periodontal condition. Feature measurements and/or the diagnosis may be processed according to a diagnosis hierarchy to determine whether a treatment is appropriate. Machine learning models may further be used to reorient, decontaminate, and restore the image prior to processing. A machine learning model may be made resistant to deception by images including added adversarial noise.
METHODS, SYSTEMS, AND MEDIA FOR DISCRIMINATING AND GENERATING TRANSLATED IMAGES
Methods, systems, and media for discriminating and generating translated images are provided. In some embodiments, the method comprises: identifying a set of training images, wherein each image is associated with at least one domain from a plurality of domains; training a generator network to generate: i) a first fake image that is associated with a first domain; and ii) a second fake image that is associated with a second domain; training a discriminator network, using as inputs to the discriminator network: i) an image from the set of training images; ii) the first fake image; and iii) the second fake image; and using the generator network to generate, for an image not included in the set of training images at least one of: i) a third fake image that is associated with the first domain; and ii) a fourth fake image that is associated with the second domain.
METHODS AND APPARATUSES FOR CORNER DETECTION USING NEURAL NETWORK AND CORNER DETECTOR
An apparatus configured to be head-worn by a user, includes: a screen configured to present graphics for the user; a camera system configured to view an environment in which the user is located; and a processing unit coupled to the camera system, the processing unit configured to: obtain locations of features for an image of the environment, wherein the locations of the features are identified by a neural network; determine a region of interest for one of the features in the image, the region of interest having a size that is less than a size of the image; and perform a corner detection using a corner detection algorithm to identify a corner in the region of interest.
Deep Learning Based Training of Instance Segmentation via Regression Layers
Novel tools and techniques are provided for implementing digital microscopy imaging using deep learning-based segmentation and/or implementing instance segmentation based on partial annotations. In various embodiments, a computing system might receive first and second images, the first image comprising a field of view of a biological sample, while the second image comprises labeling of objects of interest in the biological sample. The computing system might encode, using an encoder, the second image to generate third and fourth encoded images (different from each other) that comprise proximity scores or maps. The computing system might train an AI system to predict objects of interest based at least in part on the third and fourth encoded images. The computing system might generate (using regression) and decode (using a decoder) two or more images based on a new image of a biological sample to predict labeling of objects in the new image.
User Interface Configured to Facilitate User Annotation for Instance Segmentation Within Biological Sample
Novel tools and techniques are provided for implementing digital microscopy imaging using deep learning-based segmentation via multiple regression layers, implementing instance segmentation based on partial annotations, and/or implementing user interface configured to facilitate user annotation for instance segmentation. In various embodiments, a computing system might generate a user interface configured to collect training data for predicting instance segmentation within biological samples, and might display, within a display portion of the user interface, the first image comprising a field of view of a biological sample. The computing system might receive, from a user via the user interface, first user input indicating a centroid for each of a first plurality of objects of interest and second user input indicating a border around each of the first plurality of objects of interest. The computing system might train an AI system to predict instance segmentation of objects of interest in images of biological samples.
System for simplified generation of systems for broad area geospatial object detection
A system for simplified generation of systems for analysis of satellite images to geolocate one or more objects of interest. A plurality of training images labeled for a study object or objects with irrelevant features loaded into a preexisting feature identification subsystem causes automated generation of models for the study object. This model is used to parameterize pre-engineered machine learning elements that are running a preprogrammed machine learning protocol. Training images with the study are used to train object recognition filters. This filter is used to identify the study object in unanalyzed images. The system reports results in a requestor's preferred format.
ELECTRONIC DEVICE AND CONTROLLING METHOD THEREOF
An electronic device and a controlling method thereof are provided. A controlling method of an electronic device according to the disclosure includes: performing first learning for a neural network model for acquiring a video sequence including a talking head of a random user based on a plurality of learning video sequences including talking heads of a plurality of users, performing second learning for fine-tuning the neural network model based on at least one image including a talking head of a first user different from the plurality of users and first landmark information included in the at least one image, and acquiring a first video sequence including the talking head of the first user based on the at least one image and pre-stored second landmark information using the neural network model for which the first learning and the second learning were performed.