G06V10/7753

METHODS AND APPARATUSES FOR CORNER DETECTION USING NEURAL NETWORK AND CORNER DETECTOR

An apparatus configured to be head-worn by a user, includes: a screen configured to present graphics for the user; a camera system configured to view an environment in which the user is located; and a processing unit coupled to the camera system, the processing unit configured to: obtain locations of features for an image of the environment, wherein the locations of the features are identified by a neural network; determine a region of interest for one of the features in the image, the region of interest having a size that is less than a size of the image; and perform a corner detection using a corner detection algorithm to identify a corner in the region of interest.

Using high definition maps for generating synthetic sensor data for autonomous vehicles
11774250 · 2023-10-03 · ·

According to an aspect of an embodiment, operations may comprise accessing high definition (HD) map data of a region, presenting, via a user interface, information describing the HD map data, receiving instructions, via the user interface, for modifying the HD map data by adding one or more synthetic objects to locations in the HD map data, modifying the HD map data based on the received instructions, and generating a synthetic track in the modified HD map data comprising, for each of one or more vehicle poses, generated synthetic sensor data based on the one or more synthetic objects in the modified HD map data.

Occupancy prediction neural networks

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating a future occupancy prediction for a region of an environment. In one aspect, a method comprises: receiving sensor data generated by a sensor system of a vehicle that characterizes an environment in a vicinity of the vehicle as of a current time point, wherein the sensor data comprises a plurality of sensor samples characterizing the environment that were each captured at different time points; processing a network input comprising the sensor data using a neural network to generate an occupancy prediction output for a region of the environment, wherein: the occupancy prediction output characterizes, for one or more future intervals of time after the current time point, a respective likelihood that the region of the environment will be occupied by an agent in the environment during the future interval of time.

Unsupervised representation learning with contrastive prototypes
11776236 · 2023-10-03 · ·

The system and method are directed to a prototypical contrastive learning (PCL). The PCL explicitly encodes the hierarchical semantic structure of the dataset into the learned embedding space and prevents the network from exploiting low-level cues for solving the unsupervised learning task. The PCL includes prototypes as the latent variables to help find the maximum-likelihood estimation of the network parameters in an expectation-maximization framework. The PCL iteratively performs an E-step for finding prototypes with clustering and M-step for optimizing the network on a contrastive loss.

REALISTIC NEURAL NETWORK BASED IMAGE STYLE TRANSFER

A mobile device can implement a neural network-based style transfer scheme to modify an image in a first style to a second style. The style transfer scheme can be configured to detect an object in the image, apply an effect to the image, and blend the image using color space adjustments and blending schemes to generate a realistic result image. The style transfer scheme can further be configured to efficiently execute on the constrained device by removing operational layers based on resources available on the mobile device.

System and method for vehicle occlusion detection
11745736 · 2023-09-05 · ·

A system and method for vehicle occlusion detection is disclosed. A particular embodiment includes: receiving training image data from a training image data collection system; obtaining ground truth data corresponding to the training image data; performing a training phase to train a plurality of classifiers, a first classifier being trained for processing static images of the training image data, a second classifier being trained for processing image sequences of the training image data; receiving image data from an image data collection system associated with an autonomous vehicle; and performing an operational phase including performing feature extraction on the image data, determining a presence of an extracted feature instance in multiple image frames of the image data by tracing the extracted feature instance back to a previous plurality of N frames relative to a current frame, applying the first trained classifier to the extracted feature instance if the extracted feature instance cannot be determined to be present in multiple image frames of the image data, and applying the second trained classifier to the extracted feature instance if the extracted feature instance can be determined to be present in multiple image frames of the image data.

Deep learning based instance segmentation via multiple regression layers

Novel tools and techniques are provided for implementing digital microscopy imaging using deep learning-based segmentation and/or implementing instance segmentation based on partial annotations. In various embodiments, a computing system might receive first and second images, the first image comprising a field of view of a biological sample, while the second image comprises labeling of objects of interest in the biological sample. The computing system might encode, using an encoder, the second image to generate third and fourth encoded images (different from each other) that comprise proximity scores or maps. The computing system might train an AI system to predict objects of interest based at least in part on the third and fourth encoded images. The computing system might generate (using regression) and decode (using a decoder) two or more images based on a new image of a biological sample to predict labeling of objects in the new image.

Apparatus and method of labeling for object detection

An apparatus of labeling for object detection according to an embodiment of the present disclosure includes an image selector that determines a plurality of labeling target images from among a plurality of unlabeled images, and determines a labeling order of the plurality of labeling target images, a feedback obtainer that obtains label inspection information on the plurality of labeling target images from a user, and a model trainer that learns the label inspection information input from the user by using the labeling target images, obtains a pseudo label for supervised learning based on a learning result using the label inspection information, and re-determines the labeling order of the labeling target images based on the pseudo label.

DATA LABELING METHOD AND APPARATUS, COMPUTING DEVICE, AND STORAGE MEDIUM

A data labeling method and apparatus, a computing device, and a storage medium are provided, which belong to the field of artificial intelligence technologies. The method includes: an AI platform determines a plurality of hard examples in an unlabeled image set and hard example attributes of the plurality of hard examples, where the hard example attribute includes a hard example coefficient; the AI platform displays at least one hard example in the plurality of hard examples and a corresponding hard example attribute to a user by using a display interface; and the AI platform obtains a labeling result obtained after the user confirms the at least one hard example in the display interface based on hard example coefficients of the plurality of hard examples. According to this application, data labeling efficiency is improved, and AI model optimization efficiency is improved.

Method, device, and computer program product for self-supervised learning of pixel-wise anatomical embeddings in medical images

The present disclosure provides a method, a device, and a computer program product using a self-supervised anatomical embedding (SAM) method. The method includes randomly selecting a plurality of images; for each image of the plurality of images, performing random data augmentation to obtain a patch pair, generating global and local embedding tensors for each patch of the patch pair, and selecting positive pixel pairs from the patch pair and obtaining positive embedding pairs; for each positive pixel pair, computing global and local similarity maps, finding global hard negative embeddings, selecting global random negative embeddings, pooling the global hard negative embeddings and the global random negative embeddings to obtain final global negative embeddings, and finding local hard negative embeddings using the global and local similarity maps, and randomly sampling final local negative embeddings from the local hard negative embeddings; and minimizing a final info noise contrastive estimation (InfoNCE) loss.