Patent classifications
G06V10/80
METHOD OF PROCESSING IMAGE, ELECTRONIC DEVICE, AND STORAGE MEDIUM
A method of processing an image, an electronic device, and a storage medium, which relate to the artificial intelligence field, in particular to fields of computer vision and intelligent transportation technologies. The method includes: determining at least one key frame image in a scene image sequence captured by a target camera; determining a camera pose parameter associated with each key frame image in the at least one key frame image, according to a geographic feature associated with the key frame image; and projecting each scene image in the scene image sequence to obtain a target projection image according to the camera pose parameter associated with the key frame image, so as to generate a scene map based on the target projection image. The geographic feature associated with any key frame image indicates localization information of the target camera at a time instant of capturing the corresponding key frame image.
Domain adaptation and fusion using weakly supervised target-irrelevant data
Aspects include receiving a request to perform an image classification task in a target domain. The image classification task includes identifying a feature in images in the target domain. Classification information related to the feature is transferred from a source domain to the target domain. The transferring includes receiving a plurality of pairs of task-irrelevant images that each includes a task-irrelevant image in the source domain and in the target domain. The task-irrelevant image in the source domain has a fixed correspondence to the task-irrelevant image in the target domain. A target neural network is trained to perform the image classification task in the target domain. The training is based on the plurality of pairs of task-irrelevant images. The image classification task is performed in the target domain and includes applying the target neural network to an image in the target domain and outputting an identified feature.
METHOD AND SYSTEM OF MULTI-ATTRIBUTE NETWORK BASED FAKE IMAGERY DETECTION (MANFID)
A method for detecting fake images includes: obtaining an image for authentication, and hand-crafting a multi-attribute classifier to determine whether the image is authentic. Hand-crafting the multi-attribute classifier includes fusing at least an image classifier, an image spectrum classifier, a co-occurrence matrix classifier, and a one-dimensional (1D) power spectrum density (PSD) classifier. The multi-attribute classifier is trained by pre-processing training images to generate an attribute-specific training dataset to train each of the image classifier, the image spectrum classifier, the co-occurrence matrix classifier, and the 1D PSD classifier.
Sensor fusion for precipitation detection and control of vehicles
An apparatus includes a processor configured to be disposed with a vehicle and a memory coupled to the processor. The memory stores instructions to cause the processor to receive, at least two of: radar data, camera data, lidar data, or sonar data. The sensor data is associated with a predefined region of a vicinity of the vehicle while the vehicle is traveling during a first time period. At least a portion of the vehicle is positioned within the predefined region during the first time period. The method also includes detecting that no other vehicle is present within the predefined region. An environment of the vehicle during the first time period is classified as one state from a set of states that includes at least one of dry, light rain, heavy rain, light snow, or heavy snow, based on at least two of the sensor data to produce an environment classification. An operational parameter of the vehicle based on the environment classification is modified.
Apparatus for Q-learning for continuous actions with cross-entropy guided policies and method thereof
An apparatus for performing continuous actions includes a memory storing instructions, and a processor configured to execute the instructions to obtain a first action of an agent, based on a current state of the agent, using a cross-entropy guided policy (CGP) neural network, and control to perform the obtained first action. The CGP neural network is trained using a cross-entropy method (CEM) policy neural network for obtaining a second action of the agent based on an input state of the agent, and the CEM policy neural network is trained using a CEM and trained separately from the training of the CGP neural network.
Urban remote sensing image scene classification method in consideration of spatial relationships
An urban remote sensing image scene classification method in consideration of spatial relationships is provided and includes following steps of: cutting a remote sensing image into sub-images in an even and non-overlapping manner; performing a visual information coding on each of the sub-images to obtain a feature image Fv; inputting the feature image Fv into a crossing transfer unit to obtain hierarchical spatial characteristics; performing convolution of dimensionality reduction on the hierarchical spatial characteristics to obtain dimensionality-reduced hierarchical spatial characteristics; and performing a softmax model based classification on the dimensionality-reduced hierarchical spatial characteristics to obtain a classification result. The method comprehensively considers the role of two kinds of spatial relationships being regional spatial relationship and long-range spatial relationship in classification, and designs three paths in a crossing transfer unit for relationships fusion, thereby obtaining a better urban remote sensing image scene classification result.
Image augmentation and object detection
Computing systems may support image classification and image detection services, and these services may utilize object detection/image classification machine learning models. The described techniques provide for normalization of confidence scores corresponding to manipulated target images and for non-max suppression within the range of confidence scores for manipulated images. In one example, the techniques provide for generating different scales of a test image, and the system performs normalization of confidence scores corresponding to each scaled image and non-max suppression per scaled image These techniques may be used to provide more accurate image detection (e.g., object detection and/or image classification) and may be used with models that are not trained on modified image sets. The model may be trained on a standard (e.g. non-manipulated) image set but used with manipulated target images and the described techniques to provide accurate object detection.
Image-based techniques for stabilizing positioning estimates
A device implementing a system for estimating device location includes at least one processor configured to receive a first estimated position of the device at a first time. The at least one processor is further configured to capture, using an image sensor of the device, images during a time period defined by the first time and a second time, and determine, based on the images, a second estimated position of the device, the second estimated position being relative to the first estimated position. The at least one processor is further configured to receive a third estimated position of the device at the second time, and estimate a location of the device based on the second estimated position and the third estimated position.
Method and apparatus for mammographic multi-view mass identification
A method, applied to an apparatus for mammographic multi-view mass identification, includes receiving a main image, a first auxiliary image, and a second auxiliary image. The main image and the first auxiliary image are images of a breast of a person, and the second auxiliary image is an image of another breast of the person. The method further includes detecting the nipple location based on the main image and the first auxiliary image; generating a first probability map of the main image based on the main image, the first auxiliary image, and the nipple location; generating a second probability map of the main image based on the main image, the second auxiliary image, and the nipple location; and generating and outputting a fused probability map based on the first probability map and the second probability map.
Human body attribute recognition method and apparatus, electronic device, and storage medium
The present disclosure describes human body attribute recognition methods and apparatus, electronic devices, and a storage medium. The method includes acquiring a sample image containing a plurality of to-be-detected areas being labeled with true values of human body attributes; generating, through a recognition model, a heat map of the sample image and heat maps of the to-be-detected areas to obtain a global heat map and local heat maps; fusing the global and the local heat maps to obtain a fused image, and performing human body attribute recognition on the fused image to obtain predicted values; determining a focus area of each type of human body attribute according to the global and the local heat maps; correcting the recognition model by using the focus area, the true values, and the predicted values; and performing, based on the corrected recognition model, human body attribute recognition on a to-be-recognized image.