Patent classifications
G06V30/2504
IMAGING SYSTEM AND METHOD USING A MULTI-LAYER MODEL APPROACH TO PROVIDE ROBUST OBJECT DETECTION
A system and method of detecting an image of a template object in a captured image may include comparing, by a processor, an image model of an imaged template object to multiple locations, rotations, and scales in the captured image. The image model may be defined by multiple model base point sets derived from contours of the imaged template object, where each model base point set inclusive of a plurality of model base points that are positioned at corresponding locations associated with distinctive features of the imaged template object. Each corresponding model base point of the model base point sets may (i) be associated with respective layers and (ii) have an associated gradient vector. A determination may be made as to whether and where the image of the object described by the image model is located in the captured image.
Object detection in vehicles using cross-modality sensors
A system includes first and second sensors and a controller. The first sensor is of a first type and is configured to sense objects around a vehicle and to capture first data about the objects in a frame. The second sensor is of a second type and is configured to sense the objects around the vehicle and to capture second data about the objects in the frame. The controller is configured to down-sample the first and second data to generate down-sampled first and second data having a lower resolution than the first and second data. The controller is configured to identify a first set of the objects by processing the down-sampled first and second data having the lower resolution. The controller is configured to identify a second set of the objects by selectively processing the first and second data from the frame.
Apparatus and method for identifying obstacle around vehicle
In an apparatus for identifying an obstacle around a vehicle, an acquirer is configured to acquire an image captured by a camera mounted to the vehicle. An extractor is configured to extract feature points of the image. A generator is configured to generate an optical flow that is a movement vector from each of the feature points of the image acquired before the current time to a corresponding feature point of the image acquired at the current time. A classifier configured to classify the optical flows into groups each corresponding to an object in the image based on pixel positions of the feature points. An identifier is configured to, for each of the groups that the optical flows are classified by the classifier into, identify whether an object corresponding to the group in the image is a stationary object or a moving object based on a degree of variability in lengths of the optical flows of the group.
Image-capturing unit and component-mounting device
The image-capturing unit includes an imaging section; a holding section configured to hold a subject to be imaged by the imaging section; a light irradiation section configured to select light of one or more light sources out of multiple light sources having different wavelengths, and to irradiate the subject held in the holding section with the light; a storage section configured to store a correspondence among a color of the light emitted for irradiating the subject by the light irradiation section, a material of an irradiation surface irradiated with the light, and a resolution representing the number of pixels per unit length; and an image processing section configured to obtain the resolution from the correspondence, based on the color of the light emitted for irradiating the subject and the material of the irradiation surface of the subject, and to process a subject image by using the resolution.
Method, System, and Computer-Readable Medium for Obtaining Aggregated Multi-Scale First and Second Dimensional Receptive Field Information
A method includes: aggregating information from a corresponding combination of all of multi-scale first dimensional receptive fields and each of multi-scale second dimensional receptive fields, so that information from multi-scale first and second dimensional receptive fields corresponding to the multi-scale second dimensional receptive fields is obtained; wherein the multi-scale first dimensional receptive fields being first dimensional is being one of spatial and temporal, and the multi-scale second dimensional receptive fields being second dimensional is being the other of spatial and temporal; wherein a corresponding first convolutional neural network operation set provides each of the multi-scale second dimensional receptive fields and each operation of the first convolutional neural network operation set has a corresponding first dimensional local to local constraint; and aggregating the information from the multi-scale first and second dimensional receptive fields to obtain aggregated multi-scale first and second dimensional receptive field information.
PREPROCESSOR TRAINING FOR OPTICAL CHARACTER RECOGNITION
A method includes executing a Optical Character Recognition (OCR) preprocessor on training images to obtain OCR preprocessor output, executing an OCR engine on the OCR preprocessor output to obtain OCR engine output, and executing an approximator on the OCR preprocessor output to obtain approximator output. The method further includes iteratively adjusting the approximator to simulate the OCR engine using the OCR engine output and the approximator output, and generating OCR preprocessor losses using the approximator output and target labels. The method further includes iteratively adjusting the OCR preprocessor using the OCR preprocessor losses to obtain a customized OCR preprocessor.
GRAIN-BASED MINEROLOGY SEGMENTATION SYSTEM AND METHOD
A method of enhancing a resolution of an EDS image of a sample includes generating an EDS image of the sample, generating a non-EDS image of the sample generating, using a machine learning algorithm, an enhanced resolution EDS image of the sample based on the generated feature map and based on the first EDS image, where a resolution of the enhanced resolution EDS image is higher than a resolution of the first EDS image.
Artificial intelligence for robust drug dilution detection
Techniques are provided detecting diluted drugs using machine learning. Measurements and images corresponding to a product are obtained, wherein the product is formulated as a liquid, and wherein the measurements and images capture physical, spectral, optical, and/or chemical properties of the product. The measurements and images are provided to a machine learning model, wherein the machine learning model is trained using data generated from interactive learning modules (e.g., a generative adversarial network). The machine learning model detects whether the product or chemical is a real or counterfeit product. In addition, these techniques may be used by practitioners (e.g., medical personnel dispensing a prescribed dosage of a drug with a specific dilution level) to detect prescription errors at the point of administration.
Vehicular vision system with enhanced range for pedestrian detection
A vision system for a vehicle includes a camera and an electronic control unit (ECU) with an image processor. The ECU generates a reduced resolution frame of captured image data and the ECU determines a reduced resolution detection result based on pedestrian detection using the reduced resolution frame of captured image data. The ECU, responsive to processing by the image processor of image data, generates a cropped frame of captured image data and the ECU determines a cropped detection result based on pedestrian detection using the cropped frame of captured image data. Responsive to determining the reduced resolution detection result and determining the cropped detection result, the ECU merges the reduced resolution detection result and the cropped detection result into a final pedestrian detection result. The final pedestrian detection result is indicative of presence of a pedestrian within the field of view of the camera.
Multi-Task Multi-Sensor Fusion for Three-Dimensional Object Detection
Provided are systems and methods that perform multi-task and/or multi-sensor fusion for three-dimensional object detection in furtherance of, for example, autonomous vehicle perception and control. In particular, according to one aspect of the present disclosure, example systems and methods described herein exploit simultaneous training of a machine-learned model ensemble relative to multiple related tasks to learn to perform more accurate multi-sensor 3D object detection. For example, the present disclosure provides an end-to-end learnable architecture with multiple machine-learned models that interoperate to reason about 2D and/or 3D object detection as well as one or more auxiliary tasks. According to another aspect of the present disclosure, example systems and methods described herein can perform multi-sensor fusion (e.g., fusing features derived from image data, light detection and ranging (LIDAR) data, and/or other sensor modalities) at both the point-wise and region of interest (ROI)-wise level, resulting in fully fused feature representations.