G06V10/77

SYSTEM AND METHOD FOR UNSUPERVISED LEARNING OF SEGMENTATION TASKS
20230050573 · 2023-02-16 ·

Apparatuses and methods are provided for training a feature extraction model determining a loss function for use in unsupervised image segmentation. A method includes determining a clustering loss from an image; determining a weakly supervised contrastive loss of the image using cluster pseudo labels based on the clustering loss; and determining the loss function based on the clustering loss and the weakly supervised contrastive loss.

SIMULTANEOUS ORIENTATION AND SCALE ESTIMATOR (SOSE)

A method and hardware based system provide for descriptor-based feature mapping during terrain relative navigation (TRN). A first reference image/premade terrain map and a second image are acquired. Features in the first reference image and the second image are detected. A scale and an orientation of the one or more detected features are estimated based on an intensity centroid (IC), moments of the detected features, an orientation which is in turn based on an angle between a center of each of the detected features and the IC, and an orientation stability measure which is in turn based on a radius. Signatures are computed for each of the detected features using the estimated scale and orientation and then converted into feature descriptors. The descriptors are used to match features from the two images which are then used to perform TRN.

METHOD AND APPARATUS FOR VIDEO RECOGNITION

Broadly speaking, the present techniques generally relate to a method and apparatus for video recognition, and in particular relate to a computer-implemented method for performing video recognition using a transformer-based machine learning, ML, model. Put another way, the present techniques provide new methods of image processing in order to automatically extract feature information from a video.

SYSTEM AND METHOD FOR RARE OBJECT LOCALIZATION AND SEARCH IN OVERHEAD IMAGERY

A feature extractor and novel training objective are provided for content-based image retrieval. For example, a computer-implemented method includes applying a query image and a search image to a neural network of a feature extraction network of a computing device, the query image indicating an object to be searched for in the search image. The feature extraction network includes the neural network, a spatial feature neural network receiving a first output of the neural network pertaining to the search image, and an embedding network receiving a second output of the neural network pertaining to the query image. The method includes generating spatial search features from the spatial feature neural network, generating a query feature from the embedding network, applying the query feature to an artificial neural network (ANN) index, and determining an optimal matching result of an object in the search image based on an operation using the ANN index.

PATHOLOGICAL DIAGNOSIS ASSISTING METHOD USING AI, AND ASSISTING DEVICE
20230045882 · 2023-02-16 ·

Diagnosis is assisted by acquiring microscopical observation image data while specifying the position, classifying the image data into histological types with the use of AI, and reconstructing the classification result in a whole lesion. There is provided a pathological diagnosis assisting method that can provide an assistance technology which performs a pathological diagnosis efficiently with satisfactory accuracy by HE staining which is usually used by pathologists. Furthermore, there are provided a pathological diagnosis assisting system, a pathological diagnosis assisting program, and a pre-trained model.

METHOD FOR TRAINING STUDENT NETWORK AND METHOD FOR RECOGNIZING IMAGE

Disclosed are a method for training a Student Network and a method for recognizing an image. The method includes: acquiring first prediction feature information of a sample image on the first granularity and second prediction feature information of the sample image on the second granularity by inputting the sample image into a Student Network, and acquiring first feature information of the sample image on the first granularity and second feature information of the sample image on the second granularity by inputting the sample image into a Teacher Network, and acquiring a target Student Network.

INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND STORAGE MEDIUM

There is provided with an information processing apparatus. An approximate discrimination unit discriminates an approximate type of an object from a first captured image obtained by capturing the object to which identification information is added. A setting unit sets, based on the approximate type of the object, an image capturing condition for capturing an image to obtain the identification information. A detail discrimination unit identifies the identification information from a second captured image obtained by capturing the object under the image capturing condition and discriminates a detailed type of the object based on a result of the identification.

Deep learning based methods and systems for nucleic acid sequencing

Methods and systems for determining a plurality of sequences of nucleic acid (e.g., DNA) molecules in a sequencing-by-synthesis process are provided. In one embodiment, the method comprises obtaining images of fluorescent signals obtained in a plurality of synthesis cycles. The images of fluorescent signals are associated with a plurality of different fluorescence channels. The method further comprises preprocessing the images of fluorescent signals to obtain processed images. Based on a set of the processed images, the method further comprises detecting center positions of clusters of the fluorescent signals using a trained convolutional neural network (CNN) and extracting, based on the center positions of the clusters of fluorescent signals, features from the set of the processed images to generate feature embedding vectors. The method further comprises determining, in parallel, the plurality of sequences of DNA molecules using the extracted features based on a trained attention-based neural network.

Method for generating web code for UI based on a generative adversarial network and a convolutional neural network
11579850 · 2023-02-14 · ·

Provided is a method for generating web codes for a user interface (UI) based on a generative adversarial network (GAN) and a convolutional neural network (CNN). The method includes steps described below. A mapping relationship between display effects of a HyperText Markup Language (HTML) element and source codes of the HTML element is constructed. A location of an HTML element in an image I is recognized. Complete HTML codes of the image I are generated. The similarity between manually-written HTML codes and the generated complete HTML codes and the similarity between the image I and an image I.sub.1 generated by the generated complete HTML codes are obtained. After training, an image-to-HTML-code generation model M is obtained. A to-be-processed UI image is input into the model M so as to obtain corresponding HTML codes. According to the method of the present disclosure, an image-to-HTML-code generation model M can be obtained.

BEHAVIOR RECOGNITION METHOD AND SYSTEM, ELECTRONIC DEVICE AND COMPUTER-READABLE STORAGE MEDIUM
20230042187 · 2023-02-09 ·

A behavior recognition method and system, including: dividing video data into a plurality of video clips, performing frame extraction processing on each video clip to obtain frame images, and performing optical flow extraction on the frame images to obtain optical flow images; performing feature extraction on the frame images and the optical flow images to obtain feature maps of the frame images and the optical flow images; performing spatio-temporal convolution processing on the feature maps of the frame images and the optical flow images, and determining a spatial prediction result and a temporal prediction result; fusing the spatial prediction results of all the video clips to obtain a spatial fusion result, and fusing the temporal prediction results of all the video clips to obtain a temporal fusion result; and performing two-stream fusion on the spatial fusion result and the temporal fusion result to obtain a behavior recognition result.