G06V20/653

LOCALIZATION AND MAPPING METHOD
20180012105 · 2018-01-11 ·

A method comprising: obtaining a three-dimensional (3D) point cloud about an object; obtaining binary feature descriptors for feature points in a 2D image about the object; assigning a plurality of index values for each feature point as multiple bits of the corresponding binary feature descriptor; storing the binary feature descriptor in a table entry of a plurality of hash key tables of a database image; obtaining query binary feature descriptors for feature points in a query image; matching the query binary feature descriptors to the binary feature descriptors of the database image; reselecting one bit of the hash key of the matched database image; and re-indexing the feature points in the table entries of the hash key table of the database image.

AUTOMATIC ANALYSER
20180012375 · 2018-01-11 ·

A two-dimensional code is attached to a location of a reagent storage unit which is visually recognizable from the outside, and a coordinate position of the two-dimensional code in a coordinate system of the two-dimensional code and coordinate information of an installation position of a reagent bottle are held. After that, an image of the two-dimensional code is captured by a portable terminal so that a coordinate system of an image capture unit of the portable terminal is converted into the coordinate system of the two-dimensional code using AR technology. The coordinate information of the installation position of the reagent bottle in the coordinate system of the two-dimensional code is regarded as positional coordinates in the captured image on the basis of the conversion, thereby ascertaining the position of the reagent bottle on the captured image and displaying the ascertained position on a display unit.

Method for monitoring an orthodontic treatment

A method for monitoring the positioning of the teeth including production of a three-dimensional digital initial reference model of the arches of the patient and, for each tooth, definition, from the initial reference model, of a three-dimensional digital reference tooth model; acquisition of updated image of at least one two-dimensional image of the arches in actual acquisition conditions; analysis of each updated image and production, for each updated image, of an updated map; optionally, determination, for each updated image, of rough virtual acquisition conditions approximating the actual acquisition conditions; searching, for each updated image, for a final reference model corresponding to the positioning of the teeth during the acquisition of the updated image, for each tooth model, comparison of the positionings of the tooth model in the initial reference model and in the reference model obtained at the end of the preceding steps to determine the movement of the teeth.

Automated classification and taxonomy of 3D teeth data using deep learning methods

A computer-implemented method for automated classification of 3D image data of teeth includes a computer receiving one or more of 3D image data sets where a set defines an image volume of voxels representing 3D tooth structures within the image volume associated with a 3D coordinate system. The computer pre-processes each of the data sets and provides each of the pre-processed data sets to the input of a trained deep neural network. The neural network classifies each of the voxels within a 3D image data set on the basis of a plurality of candidate tooth labels of the dentition. Classifying a 3D image data set includes generating for at least part of the voxels of the data set a candidate tooth label activation value associated with a candidate tooth label defining the likelihood that the labelled data point represents a tooth type as indicated by the candidate tooth label.

Contour-based detection of closely spaced objects

A system includes a sensor and a client. The client receives a set of frames of top-view depth images generated by the sensor. The client identifies a frame of the received frames in which a first contour associated with a first object is merged with a second contour associated with a second object. The client determines, at a first depth in the identified frame, a merged-contour region which is associated with the merged contours. The client detects a third contour at a second depth that is less than the first depth and determines a first region associated with the third contour. The client detects a fourth contour at the second depth and determines a second region associated with the fourth contour. If criteria are satisfied, the client associates the first region with a position of the first object and associates the second region with a position of the second object.

MULTI-VIEW NEURAL HUMAN RENDERING
20230027234 · 2023-01-26 ·

An image-based method of modeling and rendering a three-dimensional model of an object is provided. The method comprises: obtaining a three-dimensional point cloud at each frame of a synchronized, multi-view video of an object, wherein the video comprises a plurality of frames; extracting a feature descriptor for each point in the point cloud for the plurality of frames without storing the feature descriptor for each frame; producing a two-dimensional feature map for a target camera; and using an anti-aliased convolutional neural network to decode the feature map into an image and a foreground mask.

Bending estimation device, bending estimation method, and program

Even when a missing portion occurs in a solid data set on a columnar structure, an estimator for a deflection value and an accuracy of the deflection value are correctly estimated according to an extent of the missing portion and the like. A measurement accuracy estimation unit (15) is included that: calculates a deflection of a columnar structure and an extent of a missing portion, from a solid data set on the columnar structure; calculates an accuracy assessment indicator for the deflection that is acquirable when a plurality of missing portion patterns occur on a virtual basis, based on a plurality of solid data sets in each of which the calculated extent of the missing portion is smaller than a preset threshold value, the accuracy assessment indicator being calculated for each missing portion pattern; and calculates an accuracy of the deflection calculated from the solid data set, based on the calculated accuracy assessment indicator for each missing portion pattern, and based on the calculated extent of the missing portion in the solid data set.

Visual dubbing using synthetic models
11562597 · 2023-01-24 · ·

A computer-implemented method of processing target footage of a target human face includes training an encoder-decoder network comprising an encoder network, a first decoder network, and a second decoder network. The training includes training a first path through the encoder-decoder network including the encoder network and the first decoder network to reconstruct the target footage of the target human face, and training a second path through the encoder-decoder network including the encoder network and the second decoder network to process renders of a synthetic face model exhibiting a range of poses and expressions to determine parameter values for the synthetic face model corresponding to the range of poses and expressions. The method includes processing, using a trained network path comprising or trained using the encoder network and comprising the first decoder network, source data representing the synthetic face model exhibiting a source sequence of expressions, to generate output video data.

Product onboarding machine
11704888 · 2023-07-18 · ·

A method for generating training examples for a product recognition model is disclosed. The method includes capturing images of a product using an array of cameras. A product identifier for the product is associated with each of the images. A bounding box for the product is identified in each of the images. The bounding boxes are smoothed temporally. A segmentation mask for the product is identified in each bounding box. The segmentation masks are optimized to generate an optimized set of segmentation masks. A machine learning model is trained using the optimized set of segmentation masks to recognize an outline of the product. The machine learning model is run to generate a set of further-optimized segmentation masks. The bounding box and further-optimized segmentation masks from each image are stored in a master training set with its product identifier as a training example to be used to train a product recognition model.

Biometric Authentication Using Head-Mounted Devices
20230222197 · 2023-07-13 ·

A head-mounted wearable device includes a frame mountable on a head of a user; an infrared imaging device arranged to image a face of the user when the frame is mounted on the head of the user; and a computing system configured to perform operations including causing the infrared imaging device to capture an image of the face of the user using infrared light received at the infrared camera and initiating a biometric authentication process based on the image. The head-mounted wearable device may include a visible-light imaging device to image the face of the user with the computing system configured to perform operations including causing the visible-light imaging device to capture a second image of the face of the user using visible light received at the visible-light imaging device, with the biometric authentication process being based in part on the second image.