Patent classifications
G06V10/811
BOATERS SAFE DISTANCE WARNING DEVICE
A device for measuring shoreline distance and distance to objects on a boat, including a measurement mechanism for measuring shoreline distance and object distance including a global positioning system (GPS) in combination with a camera and sensor. A method of determining shoreline distance and distance to objects on a boat, by actuating the device, measuring the shoreline distance and object distance from the boat, and displaying the shoreline distance and object distance on the graphical interface display.
Embedding human labeler influences in machine learning interfaces in computing environments
A mechanism is described for facilitating embedding of human labeler influences in machine learning interfaces in computing environments, according to one embodiment. A method of embodiments, as described herein, includes detecting sensor data via one or more sensors of a computing device, and accessing human labeler data at one or more databases coupled to the computing device. The method may further include evaluating relevance between the sensor data and the human labeler data, where the relevance identifies meaning of the sensor data based on human behavior corresponding to the human labeler data, and associating, based on the relevance, human labeler data with the sensor data to classify the sensor data as labeled data. The method may further include training, based on the labeled data, a machine learning model to extract human influences from the labeled data, and embed one or more of the human influences in one or more environments representing one or more physical scenarios involving one or more humans.
Electronic apparatus and control method thereof
A method for controlling an electronic apparatus includes storing a plurality of artificial intelligence models in a first memory, based on receiving a control signal for loading a first artificial intelligence model among the plurality of stored artificial intelligence models into a second memory, identifying an available memory size of the second memory, and based on a size of the first artificial intelligence model being larger than the available memory size of the second memory, obtaining a first compression artificial intelligence model by compressing the first artificial intelligence model based om the available memory size of the second memory, and loading the first compression artificial intelligence model into the second memory.
Method and device for situation awareness
A method for situation awareness is provided. The method comprises: preparing a neural network trained by a learning set, wherein the learning set includes a plurality of maritime images and maritime information including object type information which includes a first type index for a vessel, a second type index for a water surface and a third type index for a ground surface, and distance level information which includes a first level index indicating that a distance is undefined, a second level index indicating a first distance range and a third level index indicating a second distance range greater than the first distance range; obtaining a target maritime image generated from a camera; and determining a distance of a target vessel based on the distance level index of the maritime information being outputted from the neural network which receives the target maritime image and having the first type index.
Vehicle debris strike mitigation
A vehicle computer comprises a processor and a memory. The memory stores instructions executable by the processor to detect debris flying above a roadway, to input vehicle sensor data to a first classifier that outputs a source of the debris, and based on the source of the debris, to compare sensor data representing the debris to stored reference data to determine a type of physical material included in the debris. The memory stores instruction to input the type of physical material and an environmental condition to a second classifier that outputs a risk assessment, and to actuate the vehicle based on the risk assessment.
Single-camera stereoaerophotogrammetry using UAV sensors
The disclosure presents novel methods to conduct aerial surveying, inspection and measurements with higher accuracy in a fast and easy way, comprising: (1) flying a drone with an accelerometer, gyro, and camera sensors over a target object; (2) capturing a first aerial image at a first position; (3) capturing a second aerial image at a second position, wherein the second position has a horizontal and vertical displacement from the first position; (4) calculating the displacements between the first and second location using a sensor fusion estimation algorithm from the position sensors' data; (5) solving for the pixel depth information of the aerial images by using a single-camera stereophotogrammetry algorithm with the relative altitude and horizontal distance; (6) deriving the ground sample distance (GSD) of each pixel from the calculated depth information; (7) using the image pixel GSD values to compute any geometric properties of the target objects.
Generating modified digital images utilizing a dispersed multimodal selection model
The present disclosure relates to systems, methods, and non-transitory computer readable media for generating modified digital images based on verbal and/or gesture input by utilizing a natural language processing neural network and one or more computer vision neural networks. The disclosed systems can receive verbal input together with gesture input. The disclosed systems can further utilize a natural language processing neural network to generate a verbal command based on verbal input. The disclosed systems can select a particular computer vision neural network based on the verbal input and/or the gesture input. The disclosed systems can apply the selected computer vision neural network to identify pixels within a digital image that correspond to an object indicated by the verbal input and/or gesture input. Utilizing the identified pixels, the disclosed systems can generate a modified digital image by performing one or more editing actions indicated by the verbal input and/or gesture input.
Flexible multi-channel fusion perception
A method may include obtaining first sensor data from a first sensor system and second sensor data from a second sensor system. The first and the second sensor systems may capture sensor data from a total measurable world. The method may include identifying a first object included in the first sensor data and a second object included in the second sensor data and determining first parameters corresponding to the first object and second parameters corresponding to the second object. The first parameters may be compared with the second parameters and whether the first object and the second object are a same object may be determined based on the comparing the first parameters and the second parameters. Responsive to determining that the first object and the second object are the same object, a set of objects representative of objects in the total measurable world including the same object may be generated.
METHODS, SYSTEMS, AND MEDIA FOR GENERATING VIDEO CLASSIFICATIONS USING MULTIMODAL VIDEO ANALYSIS
Methods, systems, and media for generating video classifications using multimodal video analysis are provided. In some embodiments, a method for classifying videos comprising: receiving, from a computing device, a video identifier; parsing a video associated with the video identifier into an audio portion and a plurality of image frames; analyzing the plurality of images frames associated with the video using (i) an optical character recognition technique to obtain first textual information corresponding to text appearing in at least one of the plurality of image frames and (ii) an image classifier to obtain, for each of a plurality of objects appearing in at least one of the plurality of frames of the video, a probability that an object appearing in at least one of the plurality of images falls within an image class; concurrently with analyzing the plurality of image frames associated with the video, analyzing the audio portion of the video using an automated speech recognition technique to obtain second textual information corresponding to words spoken in the video; combining the first textual information, the probability of each of the plurality of objects appearing in the at least one of the plurality of frames of the video, and the second textual information to obtain a combined analysis output for the video; determining, using a neural network, a safety score for each of a plurality of categories that the video contains content belonging to a category of the plurality of categories, wherein the combined analysis output is input into the neural network; and, in response to receiving the video identifier, transmitting a plurality of safety scores corresponding to the plurality of categories to the computing device for the video associated with the video identifier.
Gating model for video analysis
Implementations described herein relate to methods, devices, and computer-readable media to perform gating for video analysis. In some implementations, a computer-implemented method includes obtaining a video comprising a plurality of frames and corresponding audio. The method further includes performing sampling to select a subset of the plurality of frames based on a target frame rate and extracting a respective audio spectrogram for each frame in the subset of the plurality of frames. The method further includes reducing resolution of the subset of the plurality of frames. The method further includes applying a machine-learning based gating model to the subset of the plurality of frames and corresponding audio spectrograms and obtaining, as output of the gating model, an indication of whether to analyze the video to add one or more video annotations.