Patent classifications
G06V10/759
AIRCRAFT DOOR CAMERA SYSTEM FOR DOCKING ALIGNMENT MONITORING
A camera with a field of view toward an external environment of an aircraft is disposed within an aircraft door such that a ground surface is within the field of view of the camera during taxiing of the aircraft. A display device is disposed within an interior of the aircraft. A processor is operatively coupled to the camera and to the display device. The processor analyzes image data captured by the camera for docking guidance by identifying, within the captured image data, a region on the ground surface corresponding to an alignment fiducial indicating a parking location for the aircraft, determining, based on the region of the captured image data corresponding to the alignment fiducial indicating the parking location, a relative location of the aircraft with respect to the alignment fiducial, and outputting an indication of the relative location of the aircraft to the alignment fiducial.
Information processing apparatus, control method, and program
The information processing apparatus (2000) includes a feature point detection unit (2020), a determination unit (2040), an extraction unit (2060), and a comparison unit (2080). A feature point detection unit (2020) detects a plurality of feature points from the query image. The determination unit (2040) determines, for each feature point, one or more object images estimated to include the feature point. The extraction unit (2060) extracts an object region estimated to include the object in the query image in association with the object image of the object estimated to be included in the object region, on the basis of the result of the determination. The comparison unit (2080) cross-checks the object region with the object image associated with the object region and determines an object included in the object region.
QUERY OPTIMIZATION FOR DEEP CONVOLUTIONAL NEURAL NETWORK INFERENCES
A method may include generating views materializing tensors generated by a convolutional neural network operating on an image. Determining the outputs of the convolutional neural network operating on the image with a patch occluding various portions of the image. The outputs being determined by generating queries on the views that performs, based at least on the changes associated with occluding different portions of the image, partial re-computations of the views. A heatmap may be generated based on the outputs of the convolutional neural network. The heatmap may indicate the quantities to which the different portions of the image contribute to the output of the convolutional neural network operating on the image. Related systems and articles of manufacture, including computer program products, are also provided.
Detecting changes in forest composition
A method of producing a model to detect changes in forest cover is disclosed. The method includes obtaining forest-cover classification data of a land area. The land area includes one or more subregions having unchanged forest-cover classifications between a first time and a second time. The method further includes obtaining image data of the subregions at multiple times. For at least one forest-cover classification, the method includes applying a statistical analysis to the image data to determine one or more threshold values representing measurement variations. The method further includes comparing subsequently obtained image data to the one or more threshold values and classifying the one or more subregions as changed or unchanged based on the comparison of subsequently obtained image data to the one or more threshold values.
IMAGE PROCESSING DEVICE OF PERSON DETECTION SYSTEM
An image processing device of a person detection system mounted on a moving body is configured to: detect, in image data obtained from a camera, an area in which an obstacle appears; determine whether the area meets an upper body detection process condition that the obstacle in the area is distanced from a road surface within a predetermined range from the camera; perform an upper body detection process in which the area of the image data is compared with upper body comparison data to determine whether the obstacle in the area is a person, for the area that meets the upper body detection process condition; and perform a whole-body detection process in which the area of the image data is compared with whole-body comparison data to determine whether the obstacle in the area is a person, for the area that does not meet the upper body detection process condition.
FORGERY DETECTION OF FACE IMAGE
In implementations of the subject matter as described herein, there is provided a method for forgery detection of a face image. Subsequent to inputting a face image, it is detected whether a blending boundary due to the blend of different images exists in the face image, and then a corresponding grayscale image is generated based on a result of the detection, where the generated grayscale image can reveal whether the input face image is formed by blending different images. If a visible boundary corresponding to the blending boundary exists in the generated grayscale image, it indicates that the face image is a forged image; on the contrary, if the visible boundary does not exist in the generated grayscale image, it indicates that the face image is a real image.
AUTOMATICALLY GENERATING CONTEXT-BASED ALTERNATIVE TEXT USING ARTIFICIAL INTELLIGENCE TECHNIQUES
Methods, apparatus, and processor-readable storage media for automatically generating context-based alternative text using artificial intelligence techniques are provided herein. An example computer-implemented method includes generating text captions for an image derived from a web page by processing the image using an artificial intelligence-based image captioning model; determining context information pertaining to the image by processing the image using an artificial intelligence-based context and emotion recognition library; generating context-based alternative text for at least a portion of the image by processing, using at least one artificial intelligence-based alternative text generation model, at least a portion of one or more of the generated text caption(s) for the image and the determined context information pertaining to at least a portion of the image; and performing one or more automated actions based on the generated context-based alternative text.
DYNAMIC SEARCH INPUT SELECTION
Described is a system and method for enabling dynamic selection of a search input. For example, rather than having a static search input box, the search input may be dynamically positioned such that it encompasses a portion of displayed information. An image segment that includes a representation of the encompassed portion of the displayed information is generated and processed to determine an object represented in the portion of the displayed information. Additional images with visually similar representations of objects are then determined and presented to the user.
HOLOGRAM DETECTION SERVICE PROVIDING SERVER AND HOLOGRAM DETECTION METHOD
A hologram detection method according to an aspect of the disclosure, includes: inputting a first image, obtained by capturing a detection object on the basis of first flash intensity, to a neural network model to obtain a first detection result value representing the detection or not of a hologram for each of predetermined at least one detection unit regions; and comparing a threshold value with the first detection result value obtained for each detection unit region to determine the detection or not of a hologram in the first image and a first detection unit region where a hologram is detected.
Automatic field of view detection
Implementations are described herein for analyzing a sequence of digital images captured by a mobile vision sensor (e.g., integral with a robot), in conjunction with information (e.g., ground truth) known about movement of the vision sensor, to determine spatial dimensions of object(s) and/or an area captured in a field of view of the mobile vision sensor. Techniques avoid the use of visual indicia of known dimensions and/or other conventional tools for determining spatial dimensions, such as checkerboards. Instead, techniques described herein allow spatial dimensions to be determined using less resources, and are more scalable than conventional techniques.