Patent classifications
G06V30/19167
SYSTEMS AND APPLICATIONS FOR AUTOMATICALLY IDENTIFYING AND VERIFYING VEHICLE LICENSE PLATE DATA
The present disclosure relates to systems and methods for automatically identifying and verifying vehicle license plate data. Specifically, the inventive system utilizes multiple automated data points to correlate a more accurate read of a license plate taken on roadways than systems that rely solely on Optical Character Recognition. The inventive system utilizes machine learning to automatically determine the make and model of the vehicle, which is then matched against motor vehicle records to provide for automation in the 80-90% range. The inventive system also utilizes analytics to determine if there are issues with the cameras and provides for near real time alerts to maintenance personnel.
LEARNING APPARATUS, METHOD AND NON-TRANSITORY COMPUTER-READABLE STORAGE MEDIUM
According to one embodiment, a learning apparatus includes processing circuitry. The processing circuitry acquires a plurality of learning samples to be learned and a plurality of target labels associated with the respective learning samples, iteratively learns a learning model so that a learning error between output data corresponding to the learning sample and the target label is small with respect to the learning model to which the output data is output by inputting the learning sample, and displays a layout image in which at least some of the learning samples are arranged based on a learning progress regarding the iterative learning of the learning model and a plurality of the learning errors.
TECHNIQUES FOR LABELING, REVIEWING AND CORRECTING LABEL PREDICTIONS FOR P&IDS
In example embodiments, techniques are provided for efficiently labeling, reviewing and correcting predictions for P&IDs in image-only formats. To label text boxes in the P&ID, the labeling application executes an OCR algorithm to predict a bounding box around, and machine-readable text within, each text box, and displays these predictions in its user interface. The labeling application provides functionality to receive a user confirmation or correction for each predicted bounding box and predicted machine-readable text. To label symbols in the P&ID, the labeling application receives user input to draw bounding boxes around symbols and assign symbols to classes of equipment. Where there are multiple occurrences of specific symbols, the labeling application provides functionality to duplicate and automatically detect and assign bounding boxes and classes. To label connections in the P&ID, the labeling application receives user input to define connection points at corresponding symbols and creates connections between the connection points.
DEEP LEARNING BASED ADAPTIVE ARITHMETIC CODING AND CODELENGTH REGULARIZATION
A deep learning based compression (DLBC) system applies trained models to compress binary code of an input image to a target codelength. For a set of binary codes representing the quantized coefficents of an input image, the DLBC system applies a first model that is trained to predict feature probabilities based on the context of each bit of the binary codes. The DLBC system compresses the binary code via adaptive arithmetic coding based on the determined probability of each bit. The compressed binary code represents a balance between a reconstruction quality of a reconstruction of the input image and a target compression ratio of the compressed binary code.
Edge-based adaptive machine learning for object recognition
Examples of techniques for adaptive object recognition for a target visual domain given a generic machine learning model are provided. According to one or more embodiments of the present invention, a computer-implemented method for adaptive object recognition for a target visual domain given a generic machine learning model includes creating, by a processing device, an adapted model and identifying classes of the target visual domain using the generic machine learning model. The method further includes creating, by the processing device, a domain-constrained machine learning model based at least in part on the generic machine learning model such that the domain-constrained machine learning model is restricted to recognize only the identified classes of the target visual domain. The method further includes computing, by the processing device, a recognition result based at least in part on combining predictions of the domain-constrained machine learning model and the adapted model.
Deep learning based adaptive arithmetic coding and codelength regularization
A deep learning based compression (DLBC) system applies trained models to compress binary code of an input image to a target codelength. For a set of binary codes representing the quantized coefficents of an input image, the DLBC system applies a first model that is trained to predict feature probabilities based on the context of each bit of the binary codes. The DLBC system compresses the binary code via adaptive arithmetic coding based on the determined probability of each bit. The compressed binary code represents a balance between a reconstruction quality of a reconstruction of the input image and a target compression ratio of the compressed binary code.
System and method for improving recognition of characters
System and method for improving recognition of characters. A system for improving recognition of characters is disclosed. The system comprises at least one processor (10), configured to receive an image (1004) of an article (102) comprising characters to be recognized. The system (100) displays characters as recognized on a display screen (1006). Further, the system (100) is configured to receive user feedback comprising correction of an error made by the system (100) in recognizing at least one character and provide a system feedback comprising display of images or textual descriptions of one or more variants (1012, 1014, 1016, 1018, 1020, 1022) of a character, which is incorrectly recognized by the system, which enables the natural person to adapt writing style to enable better quality inputs to the recognition module. The article (102) is a handwritten paper form (102), filled and captured by the natural person.
MACHINE LEARNING-BASED TEXT RECOGNITION SYSTEM WITH FINE-TUNING MODEL
A non-transitory processor-readable medium stores instructions to be executed by a processor. The instructions cause the processor to receive a first trained machine learning model that generates a transcription based on a document. The instructions cause the processor to execute the first trained machine learning model and a second trained machine learning model to generate a refined transcription based on the transcription. The instructions cause the processor to execute a quality assurance program to generate a transcription score based on the document and the transcription. The instructions cause the processor to execute the quality assurance program to generate a refined transcription score based on the refined transcription and at least one of the document or the transcription. The at least one refined transcription score indicates an automation performance better than an automation performance for the at least one transcription score.
Polarity semantics engine analytics platform
Embodiments of the systems and methods disclosed herein provide a prescriptive analytics platform, a polarity analysis engine, and a semantic analysis engine in which a user can identify a target objective and use the system to find out whether the user's objectives are being met, what predictive factors are positively or negatively affecting the targeted objectives, as well as what recommended changes the user can make to better meet the objectives. The systems and methods may include a polarity analysis engine configured to determine the polarity of terms in free-text input in view of the target objective and the predictive factors and use the polarity to generate the recommended changes. The systems and methods may also include a semantic analysis engine to extend the results of the polarity analysis engine for improved determination of predictive factors and improved recommendations.
Manual Curation Tool for Map Data Using Aggregated Overhead Views
Examples disclosed herein may involve (i) obtaining a first layer of map data associated with sensor data capturing a geographical area, the first layer of map data comprising an aggregated overhead-view image of the geographical area, where the aggregated overhead-view image is generated from aggregated pixel values from a plurality of images associated with the geographical area, (ii) obtaining a second layer of map data, the second layer of map data comprising label data for the geographical area derived from the aggregated overhead-view image of the geographical area, and (iii) causing the first layer of map data and the second layer of map data to be presented to a user for curation of the label data.