Patent classifications
G06T5/00
Generative adversarial neural network assisted video reconstruction
A latent code defined in an input space is processed by the mapping neural network to produce an intermediate latent code defined in an intermediate latent space. The intermediate latent code may be used as appearance vector that is processed by the synthesis neural network to generate an image. The appearance vector is a compressed encoding of data, such as video frames including a person's face, audio, and other data. Captured images may be converted into appearance vectors at a local device and transmitted to a remote device using much less bandwidth compared with transmitting the captured images. A synthesis neural network at the remote device reconstructs the images for display.
Platform and methods for dynamic thin film measurements using hyperspectral imaging
Dynamic thin film interferometry is a technique used to non-invasively characterize the thickness of thin liquid films that are evolving in both space and time. Recovering the underlying thickness from the captured interferograms, unconditionally and automatically is still an open problem. A compact setup is provided employing a snapshot hyperspectral camera and the related algorithms for the automated determination of thickness profiles of dynamic thin liquid films. The technique is shown to recover film thickness profiles to within 100 nm of accuracy as compared to those profiles reconstructed through the manual color matching process. Characteristics and advantages of hyperspectral interferometry are discussed including the increased robustness against imaging noise as well as the ability to perform thickness reconstruction without considering the absolute light intensity information.
Image positioning system and image positioning method based on upsampling
An image positioning system based on upsampling and a method thereof are provided. The image positioning method based on upsampling is to fetch a region image covering a target from a wide region image, determine a rough position of the target, execute an upsampling process on the region image based on neural network data model for obtaining a super-resolution region image, map the rough position onto the super-resolution region image, and analyze the super-resolution region image for determining a precise position of the target. The present disclosed example can significantly improve the efficiency of positioning and effectively reduce the required cost of hardware.
License plate detection and recognition system
A license plate detection and recognition system receives training data comprising images of license plates. The system prepares ground truth data from the training data based predefined parameters. The system trains a first machine learning algorithm based on the ground truth data to generate a license plate detection model. The license plate detection model is configured to detect one or more regions in the images. The one or more regions contains a candidate for a license plate. The LPDR system generates a bounding box for each region. The LPDR system trains a second machine learning algorithm based on the ground truth data and the license plate detection model to generate a license plate recognition model. The license plate recognition model generates a sequence of alphanumeric characters with a level of recognition confidence for the sequence.
Method of image processing based on plurality of frames of images, electronic device, and storage medium
A method of image processing based on a plurality of frames of images, an electronic device, and a storage medium are provided. The method includes: capturing a plurality of frames of original images; obtaining a high dynamic range (HDR) image by performing image synthesis on the plurality of frames of original images; performing artificial intelligent-based denoising on the HDR image to obtain a target denoised image.
Image denoising model training method, imaging denoising method, devices and storage medium
A training method for an image denoising model that can include collecting multiple sample image groups through a shooting device, each sample image group including multiple frames of sample images with a same photographic sensitivity and sample images in different sample image groups having different photographic sensitivities. The method can further include acquiring a photographic sensitivity of each sample image group, determining a noise characterization image corresponding to each sample image group based on the photographic sensitivity, determining a training input image group and a target image associated with each sample image group, each training input image group including all or part of sample images in a corresponding sample image group and a corresponding noise characterization image, constructing multiple training pairs each including a training input image group and a target image, and training the image denoising model based on the multiple training pairs until the image denoising model converges.
Method for processing image, electronic device, and storage medium
An image processing method for identifying text on production line components obtains an image to be recognized and a standard image for reference and extracts a first text area of the image to be recognized. A second text area of the standard image is obtained, and a text window is extracted based on the second text area. The method further obtains a target text area of the image to be recognized based on the first text area and the text window, and obtains a first set of first text sub-areas, and obtains a second set of second text sub-areas, by dividing the second text area into sub-windows of the text window. The method further marks the image to be recognized as a qualifying image when each first text sub-area of the first set is the same as a corresponding second text sub-area of the second set.
Method and apparatus for detecting abnormal objects in video
Disclosed are a method and an apparatus for detecting abnormal objects in a video. The method for detecting abnormal objects in a video reconstructs a restored batch by applying each input batch to which an inpainting pattern is applied to a trained auto-encoder model, and fuses a time domain reconstruction error using time domain restored frames output by extracting and restoring a time domain feature point by applying a spatial domain reconstruction error and a plurality of successive frames using a restored frame output by combining the reconstructed restoring batch to a trained LSTM auto-encoder model to estimate an area where an abnormal object is positioned.
Image processing device
An image processing device includes a rotation processor and an image processor. The rotation processor receives an input image and generates a temporary image according to the input image. The image processor is coupled to the rotation processor and outputs a processed image according to the temporary image, wherein the image processor has a predetermined image processing width, a width of the input image is larger than the predetermined image processing width, and a width of the temporary image is less than the predetermined image processing width.
Image processing system, image processing apparatus, and non-transitory computer readable medium
An image processing apparatus includes a processor configured to extract a component related to luminance of each of a sample image and a processing target image that is to undergo image processing to match an impression of the processing target image to the sample image, extract feature values of the processing target image and the sample image by attaching to a pixel value of each pixel forming the processing target image and the sample image a weight responsive to the component related to the luminance, and make adjustment to match the feature value of the processing target image to the feature value of the sample image.