Patent classifications
G06V10/77
META-OPTIC ACCELERATORS FOR OBJECT CLASSIFIERS
A system for identifying objects in images is provided. The system may include an optical front end and a digital back end. The optical front end includes a metalens that duplicates a received image into multiple images, and a metasurface that receives the duplicate images and outputs a feature map based on the received images. The feature map may be equivalent to the computationally expensive convolution operations previously performed by a neural network. The feature map is provided to the digital back end, which uses a neural network to classify the object. Because the feature map included the convolution operations, the digital back end can classify the object more quickly and using fewer computing resources than previous systems.
Deep Learning Based Multi-Sensor Detection System for Executing a Method to Process Images from a Visual Sensor and from a Thermal Sensor for Detection of Objects in Said Images
A Deep Learning based Multi-sensor Detection System for executing a method to process images from a visual sensor and from a thermal sensor for detection of objects in said images, wherein a first deep learning network for processing images from the visual sensor and a second deep learning network for pro-cessing images from the thermal sensor are jointly used and collaboratively trained for improving both networks ability to accurately detect said objects in said images.
SELF-POSITION ESTIMATION DEVICE, MOVING BODY, SELF-POSITION ESTIMATION METHOD, AND SELF-POSITION ESTIMATION PROGRAM
An own-position estimating device for estimating an own-position of a moving body by matching a feature extracted from an acquired image with a database in which position information and the feature are associated with each other in advance, includes an evaluation result acquiring unit acquiring an evaluation result obtained by evaluating matching eligibility of the feature in the database, and a processing unit processing the database on the basis of the evaluation result acquired by the evaluation result acquiring unit.
SYSTEM AND METHOD FOR PERFORMING FACE RECOGNITION
A system and a method of performing face recognition may include: receiving a first facial image, depicting a first face, and a second facial image depicting a second face; applying an ML model on the first image, to produce a first representation vector, and applying the ML model on the second image to produce a second representation vector; comparing the first representation vector and the second representation vector; and associating the first face with the second face based on the comparison, where the ML model is trained to produce the representation vectors from the facial images, based on regions in the facial images that correspond to distinctiveness scores that are beneath a distinctiveness threshold.
ANALYSIS DEVICE AND ANALYSIS METHOD
An analysis device for visualizing an accuracy of a trained determination device includes an acquisition unit acquiring an image pair of a non-defective product image and a defective product image, an extraction unit extracting an image region of a defective part of the defective product, a generation unit generating a plurality of image regions of pseudo-defective parts, a compositing unit synthesizing each of the image regions of the plurality of pseudo-defective parts with the non-defective product image to generate a plurality of composite images having different feature quantities, an unit outputting the plurality of composite images to the determination device and acquiring a label corresponding to each of the plurality of composite images from the determination device, and a display control unit displaying an object indicating the label corresponding to each of the plurality of composite images in an array based on the feature quantities.
AUTOMATIC IMAGE CLASSIFICATION AND PROCESSING METHOD BASED ON CONTINUOUS PROCESSING STRUCTURE OF MULTIPLE ARTIFICIAL INTELLIGENCE MODEL, AND COMPUTER PROGRAM STORED IN COMPUTER-READABLE RECORDING MEDIUM TO EXECUTE THE SAME
Disclosed is an automatic image classification and processing method based on the continuous processing structure of multiple artificial intelligence models. An automatic image classification and processing method based on a continuous processing structure of multiple artificial intelligence models includes receiving image data, generating a first feature extraction value by inputting the image data into a first feature extraction model among feature extraction models, generating a second feature extraction value by inputting the image data into a second feature extraction model among the feature extraction models, and determining a classification value of the image data by inputting the first and second feature extraction values into a classification model.
DISMANTLING PROCEDURE SELECTING APPARATUS, DISMANTLING PROCEDURE SELECTING METHOD, AND DISMANTLING APPARATUS
A dismantling procedure selecting apparatus includes a dismantling information storage unit that stores a plurality of pieces of dismantling information respectively including a plurality of predetermined dismantling procedures for a plurality of dismantled objects each used as a reference for dismantling an object to be dismantled, a whole detector that captures a whole image of the object to be dismantled, a detail detector that captures an image of at least a portion of the object to be dismantled, and a dismantling procedure deriving unit that extracts a first feature as a feature of the object to be dismantled based on data obtained by capturing an image by at least one of the whole detector and the detail detector, obtains a degree of matching between the first feature and each of a plurality of second features that are respectively features of a plurality of dismantled objects, and selects a dismantling procedure associated with one of a plurality of the dismantled objects having a second feature having a highest degree of matching among a plurality of second features.
LABEL IDENTIFICATION METHOD AND APPARATUS, DEVICE, AND MEDIUM
Provided are a label identification method and apparatus, a device, and a medium. The method includes: obtaining a target feature of a first image, in which the target feature characterizes a visual feature of the first image and a word feature of at least one label; and identifying a label of the first image from the at least one label based on the target feature. By characterizing the visual feature of the first image and the target feature of the word feature of the at least one label, the label of the first image is identified from the at least one label. Thus, identification accuracy of the label can be improved.
IMAGE PROCESSING DEVICE, IMAGE PROCESSING METHOD, IMAGE PROCESSING PROGRAM, AND DIAGNOSIS SUPPORT SYSTEM
An image processing device 100 includes, in a case where designation of a plurality of partial regions corresponding to a cell morphology is received, the plurality of partial regions being extracted from a pathological image, a generation unit 154 that generates auxiliary information indicating information about a feature amount effective when a plurality of partial regions is classified or extracted with respect to a plurality of feature amounts calculated from the image; and in a case where setting information about an adjustment item according to the auxiliary information is received, an image processing unit 155 that performs an image process on the image using the setting information.
DEVICE FOR PROCESSING IMAGE AND METHOD FOR OPERATING SAME
Provided are a device and operating method thereof for obtaining compression ratio information for recognizing a target object in an image using a deep neural network model, and compressing an image using the compression ratio information and encoding the compressed image. According to an embodiment of the present disclosure, there is provided a device that receives an image via at least one camera or a communication interface, obtains a feature map for detecting a target object in the received image, outputs a compression ratio for correctly recognizing the target object in the image by inputting the image and the feature map to a deep neural network model composed of pre-trained model parameters, and generates a bitstream by compressing the image using the output compression ratio and encoding the compressed image.