Patent classifications
G06V10/806
RADAR DATA DETERMINATION CIRCUITRY AND RADAR DATA DETERMINATION METHOD
The present disclosure generally pertains to radar data determination circuitry, configured to: obtain image data; and determine predictive radar data from the image data based on learned features being represented in the image data, for determining actual radar data based on an association of the predictive radar data with radar measurement data.
HUMAN ACTIVITY RECOGNITION FUSION METHOD AND SYSTEM FOR ECOLOGICAL CONSERVATION REDLINE
A human activity recognition fusion method and system for ecological protection red line is disclosed. The method includes: obtaining a pre-stage remote sensing image and a post-stage remote sensing image of a target ecological protection red line region, and performing a data pre-processing; inputting the pre-processed pre-stage remote sensing image and the post-stage remote sensing image into a human activity recognition model after a pre-training; identifying a human activity pattern of the target ecological protection red line region as a first detection result; segmenting, calculating and analyzing the latest image data corresponding to the target ecological protection red line region based on a geographical country situation data to obtain a change pattern as a second detection result; and fusing the first detection result and the second detection result to obtain a change detection pattern of the target ecological protection red line region.
Image generation using surface-based neural synthesis
Aspects of the present disclosure involve a system and a method for performing operations comprising: receiving a two-dimensional continuous surface representation of a three-dimensional object, the continuous surface comprising a plurality of landmark locations; determining a first set of soft membership functions based on a relative location of points in the two-dimensional continuous surface representation and the landmark locations; receiving a two-dimensional input image, the input image comprising an image of the object; extracting a plurality of features from the input image using a feature recognition model; generating an encoded feature representation of the extracted features using the first set of soft membership functions; generating a dense feature representation of the extracted features from the encoded representation using a second set of soft membership functions; and processing the second set of soft membership functions and dense feature representation using a neural image decoder model to generate an output image.
Systems and methods for quantitative phenotyping of fibrosis
Systems and methods are provided for computer aided phenotyping of fibrosis-related conditions. A digital image indicates presence of collagens in a biological tissue sample. The image is processed to quantify parameters, each parameter describing a feature of the collagens that is expected to be different for different phenotypes of fibrosis. At least some features are tissue level features that describe macroscopic characteristics of the collagens, morphometric level features that describe morphometric characteristics of the collagens, and texture level features that describe an organization of the collagens. At least some of the plurality of parameters are statistics associated with histograms corresponding to distributions of the associated parameters across at least some of the digital image. At least some of the plurality of parameters are combined to obtain one or more composite scores that quantify a phenotype of fibrosis for the biological tissue sample.
VIDEO CLASSIFICATION METHOD AND APPARATUS
A video classification method and apparatus are provided in embodiments of the present invention. The method includes: establishing a neural network classification model according to a relationship between features of video samples and a semantic relationship of the video samples; obtaining a feature combination of a to-be-classified video file; and classifying the to-be-classified video file by using the neural network classification model and the feature combination of the to-be-classified video file The neural network classification model is established according to the relationship between the features of the video samples and the semantic relationship of the video samples, and the relationship between the features and the semantic relationship are fully considered. Therefore, video classification accuracy are improved.
EMBEDDED SEMANTIC DIVISION NETWORK APPARATUS OPTIMIZED FOR MMA THAT CLASSIFIES PIXELS IN VEHICLE IMAGES
Provided is an embedded semantic division network including a communication module configured to receive an image captured by a camera, a memory configured to store a semantic division network (MMANet)-based program for extracting a context of the captured image, and a processor extracts the context of the captured image by selecting a convolutional neural network (CNN) processing module or a depth-wise separable convolution (DSC) processing module according to a size of a activation map in each layer of the semantic division network that includes an encoder unit and a decoder unit including at least one of the CNN processing module and the DSC processing module that are connected from an upper layer to a lower layer and reduce features of an input image.
Systems and Methods for End-to-End Trajectory Prediction Using Radar, Lidar, and Maps
Systems and methods for trajectory prediction are provided. A method can include obtaining LIDAR data, radar data, and map data; inputting the LIDAR data, the radar data, and the map data into a network model; transforming, by the network model, the radar data into a coordinate frame associated with a most recent radar sweep in the radar data; generating, by the network model, one or more features for each of the LIDAR data, the transformed radar data, and the map data; combining, by the network model, the one or more generated features to generate fused feature data; generating, by the network model, prediction data based at least in part on the fused feature data; and receiving, as an output of the network model, the prediction data. The prediction data can include a respective predicted trajectory for a future time period for one or more detected objects.
VERIFYING AND CORRECTING TEXT PRESENTED IN COMPUTER BASED AUDIOVISUAL PRESENTATIONS
Technology for taking presentation data (for example, video images from a movie, audio from a podcast), determining that the content includes an untrue assertion (for example, “the United States only has 48 states”) and automatically correcting the presentation so that the untrue assertion is corrected (for example, replacing an incorrect video caption with “the United States has 50 states as of early 2021”).
IMPROVEMENTS IN AND RELATING TO TARGETING
A targeting method comprising the steps of determining a bearing to a target from an observer using first and second independent techniques; comparing the bearings as determined by the first and second independent techniques and determining whether the bearings are accurate; and if the bearing is deemed to be accurate; measuring a range from the observer to the target; and calculating the position the target based on the verified bearing and range from the observer's position. The bearing can be measured by using a magnetometer, and cross-checked or verified using calculations based on three-dimensional satellite cartography data. The range to the target can be cross-checked, as can the position and viewpoint of the observer.
Ground environment detection method and apparatus
A ground environment detection method and apparatus are disclosed, where the method includes: scanning a ground environment by using laser sounding signals having different operating wavelengths, receiving a reflected signal that is reflected back by the ground environment, determining scanning spot information of each scanning spot of the ground environment based on the reflected signal, determining space coordinate information and a laser reflection feature of each scanning spot based on each piece of scanning spot information, partitioning the ground environment into sub-regions having different laser reflection features, and determining a ground environment type of each sub-region. Lasers having different operating wavelengths are used to scan the ground, and the ground environment type is determined based on the reflection intensity of the ground environment under different wavelengths of lasers, thereby improving a perception effect of a complex ground environment, and better determining a passable road surface.