G06V30/15

LOW POWER MACHINE LEARNING USING REAL-TIME CAPTURED REGIONS OF INTEREST

Systems and methods are described for generating image content. The systems and methods may include, in response to receiving a request to cause a sensor of a computing device to identify image content associated with optical data captured by the sensor, detecting a first sensor data stream having a first image resolution, and detecting a second sensor data stream having a second image resolution. The systems and method may also include identifying, by processing circuitry of the computing device, at least one region of interest in the first sensor data stream, determining cropping coordinates that define a first plurality of pixels in the at least one region of interest in the first sensor data stream, and generating a cropped image representing the at least one region of interest.

DYNAMIC CAPTURE PARAMETER PROCESSING FOR LOW POWER

In one general aspect, a method can include capturing, using an image sensor, a first raw image at a first resolution, converting the first raw image to a digitally processed image using an image signal processor, and analyzing at least a portion of the digitally processed image based on a processing condition. The method can include determining that the first resolution does not satisfy the processing condition; and triggering capture of a second raw image at the image sensor at a second resolution greater than the first resolution.

License plate detection and recognition system

A license plate detection and recognition system receives training data comprising images of license plates. The system prepares ground truth data from the training data based predefined parameters. The system trains a first machine learning algorithm based on the ground truth data to generate a license plate detection model. The license plate detection model is configured to detect one or more regions in the images. The one or more regions contains a candidate for a license plate. The LPDR system generates a bounding box for each region. The LPDR system trains a second machine learning algorithm based on the ground truth data and the license plate detection model to generate a license plate recognition model. The license plate recognition model generates a sequence of alphanumeric characters with a level of recognition confidence for the sequence.

METHOD OF IDENTIFYING CHARACTERS IN IMAGES, ELECTRONIC DEVICE, AND STORAGE MEDIUM
20230023611 · 2023-01-26 ·

A method of identifying characters in images extracts features of a detection image including characters. Enhancement processing is performed on the detection image according to the features to obtain an enhanced image. Closed edges of the characters are detected in the enhanced image. First rectangular outlines of the characters are determined according to the closed edges. The first rectangular outlines are corrected to obtain second rectangular outlines. The characters are cropped from the detection image according to the second rectangular outlines. The method identifies characters in images accurately and rapidly.

METHODS, SYSTEMS, ARTICLES OF MANUFACTURE, AND APPARATUS FOR DECODING PURCHASE DATA USING AN IMAGE

Methods, apparatus, systems, and articles of manufacture are disclosed that decode purchase data using an image. An example apparatus includes processor circuitry to execute machine readable instructions to at least crop an image of a receipt based on detected regions of interest, apply a first mask to a first cropped image to generate first bounding boxes corresponding to rows of the receipt, apply a second mask to a second cropped image to generate second bounding boxes corresponding to columns of the receipt, generate a structure of the receipt by mapping words detected by an optical character recognition engine to corresponding first bounding boxes and second bounding boxes based on a mapping criterion, classify the second bounding boxes by identifying an expression of interest in ones of the second bounding boxes, and generate purchase information by extracting text of interest from the structured receipt based on the classifications.

METHODS AND APPARATUSES FOR RECOGNIZING TEXT, RECOGNITION DEVICES AND STORAGE MEDIA

Methods and an apparatuses for recognizing a text, recognition devices and storage media are provided, which belong to the field of text detections. A method includes: extracting, by the recognition device, a feature map of a to-be-recognized image, then determining segmentation information of a text region of the to-be-recognized image based on a preset segmentation network and the feature map, and then determining boundary key points in the text region based on the segmentation information, and then converting a text in the text region into a text with a target arrangement sequence based on the boundary key points and then inputting the text obtained by conversion into a preset recognition model for recognition processing.

Systems and methods of instant-messaging bot for robotic process automation and robotic textual-content extraction from images

Systems and methods of instant-messaging bot for robotic process automation (RPA) and robotic textual-content extraction from digital images include a chatbot application, a software RPA manager, and an instant-messaging (IM) platform, all built for an enterprise. The enterprise IM platform is connected to one or more public IM platforms over the Internet. The RPA manager contains multiple modules of enterprise workflows and receives instructions from the enterprise chatbot for executing individual workflows. The system allows enterprise users connected to the enterprise IM platform, and external users connected to the public IM platforms, to use instant messaging to initiate enterprise workflows that are automated with the help of the enterprise chatbot and delivered via instant messaging. Furthermore, textual-content extraction from digital images is incorporated in the RPA manager as an enterprise workflow, and provides improved convolutional neural network (CNN) methods for textual-content extraction.

UTILIZING MACHINE-LEARNING BASED OBJECT DETECTION TO IMPROVE OPTICAL CHARACTER RECOGNITION

The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately enhancing optical character recognition with a machine learning approach for determining words from reverse text, vertical text, and atypically-sized text. For example, the disclosed systems segment a digital image into text regions and non-text regions utilizing an object detection machine learning model. Within the text regions, the disclosed systems can determine reverse text glyphs, vertical text glyphs, and/or atypically-sized text glyphs utilizing an edge based adaptive binarization model. Additionally, the disclosed systems can utilize respective modification techniques to manipulate reverse text glyphs, vertical text glyphs, and/or atypically-sized glyphs for analysis by an optical character recognition model. The disclosed systems can further utilize an optical character recognition model to determine words from the modified versions of the reverse text glyphs, the vertical text glyphs, and/or the atypically-sized text glyphs.

HANDWRITTEN CONTENT REMOVING METHOD AND DEVICE AND STORAGE MEDIUM
20230037272 · 2023-02-02 · ·

A handwritten content removing method and device and a storage medium. The handwritten content removing method comprises: acquiring an input image of a text page to be processed, the input image comprising a handwritten region, which comprises a handwritten content (S10); identifying the input image so as to determine the handwritten content in the handwritten region (S11); and removing the handwritten content in the input image so as to obtain an output image (S12).

Method and apparatus for detecting and interpreting price label text

A method of price text detection by an imaging controller comprises obtaining, by the imaging controller, an image of a shelf supporting labels bearing price text, generating, by the imaging controller, a plurality of text regions containing candidate text elements from the image, assigning, by the imaging controller, a classification to each of the text regions, selected from a price text classification and a non-price text classification. The imaging controller, within each of a subset of the text regions having the price text classification: detects a price text sub-region and generates a price text string by applying character recognition to the price text sub-region. The method further includes presenting, by the imaging controller, the locations of the subset of text regions, in association with the corresponding price text strings.