Patent classifications
G06V30/148
License Plate Reading System with Enhancements
System and methods are disclosed for capturing license plate (LP) information of a vehicle in relative motion to a camera device. In one example, the camera system detects the LP in multiple frames, then aligns and geometrically rectifies the image of the LP by scaling, warping, rotating, and/or performing other functions on the images. The camera system may optimize capturing of the LP information by executing a temporal noise filter on the aligned, geometrically rectified images to generate a composite image of the LP for optical character recognition. In some examples, the camera device may include an image sensor, such as a high dynamic range (HDR) sensor, modified to set long and short exposures of the HDR sensor to capture frames of a vehicle's LP, but without consolidating the images into a composite image. The camera system may set optimal exposure settings based on detected relative speed of the vehicle.
Image processing system, image processing method, and storage medium
According to the present disclosure, a handwriting image and a background image are combined, thereby generating a combined image, a correct answer label image is generated based on the handwriting image, and the generated combined image and the generated correct answer label image are used as learning data for training a neural network.
METHOD OF DETECTING PRINTING DEFECTS, COMPUTER DEVICE, AND STORAGE MEDIUM
This application provides a method of detecting printing defects. The method includes obtaining a first image of each character in a reference image. A third image of each character is obtained based on the first image of each character, a fourth image of each character is obtained based on a second image of each character obtained from an image to be detected. Once a fifth image of each character is obtained based on the third image of each character, a sixth image of each character is obtained according to the fourth image and the fifth image of each character, a detection result of each character in the image to be detected is determined according to the fifth image and the sixth image of the each character.
UTILIZING MACHINE-LEARNING BASED OBJECT DETECTION TO IMPROVE OPTICAL CHARACTER RECOGNITION
The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately enhancing optical character recognition with a machine learning approach for determining words from reverse text, vertical text, and atypically-sized text. For example, the disclosed systems segment a digital image into text regions and non-text regions utilizing an object detection machine learning model. Within the text regions, the disclosed systems can determine reverse text glyphs, vertical text glyphs, and/or atypically-sized text glyphs utilizing an edge based adaptive binarization model. Additionally, the disclosed systems can utilize respective modification techniques to manipulate reverse text glyphs, vertical text glyphs, and/or atypically-sized glyphs for analysis by an optical character recognition model. The disclosed systems can further utilize an optical character recognition model to determine words from the modified versions of the reverse text glyphs, the vertical text glyphs, and/or the atypically-sized text glyphs.
HANDWRITTEN CONTENT REMOVING METHOD AND DEVICE AND STORAGE MEDIUM
A handwritten content removing method and device and a storage medium. The handwritten content removing method comprises: acquiring an input image of a text page to be processed, the input image comprising a handwritten region, which comprises a handwritten content (S10); identifying the input image so as to determine the handwritten content in the handwritten region (S11); and removing the handwritten content in the input image so as to obtain an output image (S12).
EXTRACTING KEY INFORMATION FROM DOCUMENT USING TRAINED MACHINE-LEARNING MODELS
Techniques for extracting key information from a document using machine-learning models in a chatbot system is disclosed herein. In one particular aspect, a method is provided that includes receiving a set of data, which includes key fields, within a document at a data processing system that includes a table detection module, a key information extraction module, and a table extraction module. Text information and corresponding location data are extracted via optical character recognition. The table detection module detects whether one or more tables are present in the document and, if applicable, a location of each of the tables. The key information extraction module extracts text from the key fields. The table extraction module extracts each of the tables based on input from the optical character recognition and the table detection module. Extraction results include the text from the key fields and each of the tables can be output.
SYSTEM AND METHOD FOR TEXT LINE AND TEXT BLOCK EXTRACTION
The invention concerns a method implemented by a device for displaying strokes of digital ink in a display area and for performing text line extraction to extract text lines from the strokes. In particular, the text line extraction may involve slicing the display area into strips, ordering for each strip the strokes into ordered lists which form collectively a first set of ordered lists, forming for each strip a second set of ordered lists by filtering out from the ordered lists of the first set strokes which are below a given size threshold, and performing a neural net analysis based on said first and second sets to determine for each stroke a respective text line to which it belongs.
SYSTEM AND METHOD FOR TEXT LINE AND TEXT BLOCK EXTRACTION
The invention concerns a method implemented by a device for displaying strokes of digital ink in a display area and for performing text line extraction to extract text lines from the strokes. In particular, the text line extraction may involve slicing the display area into strips, ordering for each strip the strokes into ordered lists which form collectively a first set of ordered lists, forming for each strip a second set of ordered lists by filtering out from the ordered lists of the first set strokes which are below a given size threshold, and performing a neural net analysis based on said first and second sets to determine for each stroke a respective text line to which it belongs.
Text Line Detection
Implementations of the present disclosure provide a solution for text line detection. In this solution, a first text region comprising a first portion of at least a first text element and a second text region comprising a second portion of at least a second text element are determined from an image. A first feature representation is extracted from the first text region and a second feature representation is extracted from the second text region. The first and second feature representations comprise at least one of an image eature representation or a semantic feature representation of the image. A link relationship between the first and second text regions can then be determined based at least in part on the first and second feature representations. The link relationship can indicate whether the first and second portions of the first and second text elements are located in a same text line. In this way, by detecting text regions and determining the link relationship thereof based on their feature representations, the accuracy and efficiency for detecting text lines in various images can be improved
SCREEN RESPONSE VALIDATION OF ROBOT EXECUTION FOR ROBOTIC PROCESS AUTOMATION
Screen response validation of robot execution for robotic process automation (RPA) is disclosed. Whether text, screen changes, images, and/or other expected visual actions occur in an application executing on a computing system that an RPA robot is interacting with may be recognized. Where the robot has been typing may be determined and the physical position on the screen based on the current resolution of where one or more characters, images, windows, etc. appeared may be provided. The physical position of these elements, or the lack thereof, may allow determination of which field(s) the robot is typing in and what the associated application is for the purpose of validation that the application and computing system are responding as intended. When the expected screen changes do not occur, the robot can stop and throw an exception, go back and attempt the intended interaction again, restart the workflow, or take another suitable action.