G06V10/7753

Systems and methods for semi-supervised depth estimation according to an arbitrary camera

System, methods, and other embodiments described herein relate to semi-supervised training of a depth model using a neural camera model that is independent of a camera type. In one embodiment, a method includes acquiring training data including at least a pair of training images and depth data associated with the training images. The method includes training the depth model using the training data to generate a self-supervised loss from the pair of training images and a supervised loss from the depth data. Training the depth model includes learning the camera type by generating, using a ray surface model, a ray surface that approximates an image character of the training images as produced by a camera having the camera type. The method includes providing the depth model to infer depths from monocular images in a device.

Active learning-based data labeling service using an augmented manifest

Techniques for active learning-based data labeling are described. An active learning-based data labeling service enables a user to build and manage large, high accuracy datasets for use in various machine learning systems. Machine learning may be used to automate annotation and management of the datasets, increasing efficiency of labeling tasks and reducing the time required to perform labeling. Embodiments utilize active learning techniques to reduce the amount of a dataset that requires manual labeling. As subsets of the dataset are labeled, this label data is used to train a model which can then identify additional objects in the dataset without manual intervention. The label data can be added to an augmented manifest, the augmented manifest can be used to filter the dataset to perform further labeling jobs on the same or different subsets of the dataset.

Transform disentangling auto-encoder and related methods
11461594 · 2022-10-04 · ·

Discussed herein are devices, systems, and methods for disentangling static and dynamic features of content. A method can include encoding by a transform disentangling autoencoder (AE), first content to generate first static features and first dynamic features and second content to generate second static features and second dynamic features, and constructing, by the AE, third content based on a combination of third static features and the first dynamic features and fourth content based on a combination of fourth static features and the second dynamic features, the third and fourth static features being determined based on the first static features and the second static features.

Systems and methods for dynamically classifying point cloud data points
11455789 · 2022-09-27 · ·

Disclosed is a system for dynamically classifying different data point sets within a point cloud with different classifications that may alter how data point sets with different classifications are processed, edited, and/or rendered. The system may generate a model based on a first set of relationships between a first set of data point elements that result in the first classification, and a second set of relationships between a second set of data point elements that result in the second classification. The system may compare the data point elements from unclassified data point sets against the first set of relationships and the second set of relationships in the model, and may assign the first classification to a particular unclassified data point set in response to the data point elements of the particular data point set having a threshold amount of the first set of relationships.

Model training using partially-annotated images

Methods and systems for training a model labeling two or more organic structures in an image. One method includes receiving a set of training images including a first plurality of images and a second plurality of images. Each of the first plurality of images including a label for a first subset of the two or more organic structures and each of the second plurality of images including a label for a second subset of the two or more organic structures, the second subset being different than the first subset. The method also includes training the model using the first plurality of images, the second plurality of images, and a label merging function mapping a label included in the first plurality of images to a label included in the second plurality of images.

MEDICAL IMAGE PROCESSING METHOD, PROCESSING APPARATUS, AND COMPUTER-READABLE STORAGE MEDIUM
20220262498 · 2022-08-18 ·

A medical image processing method and processing apparatus, and a computer readable storage medium. The method includes: obtaining a to-be-processed image; performing a feature extraction on the to-be-processed image to obtain a corresponding feature image; and re-determining a pixel value of each pixel in the to-be-processed image based on first information and second information of a corresponding pixel in the feature image, and processing the to-be-processed image; wherein the first information is information of a pixel adjacent to the corresponding pixel in the features image, and the second information is information of a pixel that is not adjacent to and is similar to the corresponding pixel in the features image.

Method for training a deep learning model to obtain histopathological information from images

A method and a system for training a deep learning model to obtain histopathological information from images.

MAP CONSTRUCTING METHOD, POSITIONING METHOD AND WIRELESS COMMUNICATION TERMINAL
20220114750 · 2022-04-14 ·

According to embodiments of the present disclosure, a map constructing method, a positioning method, and a wireless communication terminal are provided. The map constructing method includes: a series of environment images of a current; first image feature information of the environment image is obtained, where the first image feature information includes feature point information and descriptor information and based on the first image feature information, a feature point matching is performed on the environment images to select keyframe images; depth information of matched feature points in the keyframe image are acquired, based on the feature point information; and map data of the current environment are generated based on the keyframe images, where the map data includes the image feature information and the depth information of the keyframe image.

DEPTH DATA MODEL TRAINING WITH UPSAMPLING, LOSSES AND LOSS BALANCING

Techniques for training a machine learned (ML) model to determine depth data based on image data are discussed herein. Training can use stereo image data and depth data (e.g., lidar data). A first (e.g., left) image can be input to a ML model, which can output predicted disparity and/or depth data. The predicted disparity data can be used with second image data (e.g., a right image) to reconstruct the first image. Differences between the first and reconstructed images can be used to determine a loss. Losses may include pixel, smoothing, structural similarity, and/or consistency losses. Further, differences between the depth data and the predicted depth data and/or differences between the predicted disparity data and the predicted depth data can be determined, and the ML model can be trained based on the various losses. Thus, the techniques can use self-supervised training and supervised training to train a ML model.

LEARNING-BASED 3D MODEL CREATION APPARATUS AND METHOD

Disclosed herein are a learning-based three-dimensional (3D) model creation apparatus and method. A method for operating a learning-based 3D model creation apparatus includes generating multi-view feature images using supervised learning, creating a three-dimensional (3D) mesh model using a point cloud corresponding to the multi-view feature images and a feature image representing internal shape information, generating a texture map by projecting the 3D mesh model into three viewpoint images that are input, and creating a 3D model using the texture map.