Patent classifications
G06V30/19013
Method, apparatus and electronic device for annotating information of structured document
Disclosed are a method, apparatus and electronic device for annotating information of a structured document. A specific implementation is: obtaining a template image of a structured document and at least one piece of annotation information of a field to be filled in the template image, where the annotation information includes attribute value and historical content of the field to be filled, and historical position of the field to be filled in the template image; generating, according to the attribute value of the field to be filled, the historical content of the field to be filled and the historical position of the field to be filled in the template image, target filling information of the field to be filled; obtaining, according to the target filling information of the field to be filled, an image of an annotated structured document.
METHODS, SYSTEMS, APPARATUS AND ARTICLES OF MANUFACTURE FOR RECEIPT DECODING
Methods, apparatus, systems and articles of manufacture are disclosed for receipt decoding. An example apparatus includes processor circuitry to execute instructions to extract text from the receipt image, the text including bounding boxes; associate ones of the bounding boxes to link horizontally related fields of a the receipt image by selecting a first bounding box; identifying first horizontally aligned bounding boxes, the first horizontally aligned bounding boxes to include at least one bounding box of the bounding boxes that is horizontally aligned relative to the first bounding box; adding the first horizontally aligned bounding boxes to a word sync list; and connecting ones of the first horizontally aligned bounding boxes and the first bounding box based on at least one of an amount of the first horizontally aligned bounding boxes in the word sync list and a relationship among the first horizontally aligned bounding boxes and the first bounding box.
Marking inspection device, marking inspection method and article inspection apparatus
A marking region image is obtained by cutting out the part corresponding to a marking region from an article image obtained by imaging an article to be inspected. Then, whether or not the marking is properly provided is determined by performing a character recognition of a marking part for a marking region image. Further, an image of an article having no marking and no defect is stored as a reference image, whereas a marking periphery image obtained by removing the image of the marking part from the marking region image is compared to the reference image. By that comparison, whether or not any defect is included in the marking peripheral part of the marking region except the marking part is determined.
CHARACTER INFORMATION RECOGNITION METHOD BASED ON IMAGE PROCESSING
The present invention relates to a character information recognition method based on image processing. The method comprises: collecting images to obtain a target character image; then sequentially comparing the target character image with character template images in a character template library to find a maximum of a coincidence area of the character in the target character image with the character templates in the character template images; and when the coincidence area meets a preset condition, determining the target character to be recognized as the character in the corresponding character template image. The character templates are designed to include not only a coincidence-permitted region but also a coincidence-restricted region. The coincidence-restricted region is set, so that the direct comparing and matching of the character templates can be more accurately carried out, thereby improving the recognition speed.
Image processing device and operating method thereof
An image processing device includes: an image sensor for acquiring a pixel value of each of a plurality pixels; and a controller for acquiring a pattern image including the pixel value of each of the plurality of pixels and an exposure value representing an exposure time, generating a plurality of super resolution images based on pixels having the same exposure value among the plurality of pixels included in the pattern image, generating a motion map, which represents a motion of an object based on a ratio of exposure values of pixels at a selected position among a plurality of pixels included in the plurality of super resolution images and a ratio of pixel values of the pixels at the selected position, and generating a target image according to a weighted sum of the plurality of super resolution images and the motion map.
Method and apparatus for enabling text editing in a scanned document while maintaining fidelity of the appearance of the text
A computer implemented method and apparatus for enabling text editing in a scanned document while maintaining fidelity of appearance of the text. The method comprises creating a synthesized font comprising a plurality of characters using characters present in a scanned document; replacing the plurality of characters in the scanned document with characters from the plurality of characters from the synthesized font; and enabling editing of the scanned document wherein enabling editing comprises adding at least some characters from the plurality of characters of the synthesized font to the document for at least some characters added during editing.
DOCUMENT INFORMATION EXTRACTION FOR COMPUTER MANIPULATION
Systems and apparatuses are disclosed for extracting information from document images. An example method includes segmenting a document image into multiple segments and determining formatting information for each segment. Determining formatting information for a segment includes determining one or more features of the segment and comparing the one or more features of the segment to one or more clusters of features associated with different document types. The formatting information for the segment is based on the comparison. The method also includes, for each segment, storing the formatting information in a data structure associated with the segment. The method further includes, for each segment including text to be identified during information extraction, applying OCR to the segment to generate machine-encoded text and storing the machine-encoded text in the associated data structure.
Image forming apparatus and image forming system
An image forming apparatus forms only a required entry field on an application form, and, upon the application form in which this location has been filled in being scanned, further forms an additional entry field and prompts the applicant to fill in this entry field. An application form that has been completely filled in by the applicant is created by repeating this processing.
METHOD, APPARATUS AND ELECTRONIC DEVICE FOR ANNOTATING INFORMATION OF STRUCTURED DOCUMENT
Disclosed are a method, apparatus and electronic device for annotating information of a structured document. A specific implementation is: obtaining a template image of a structured document and at least one piece of annotation information of a field to be filled in the template image, where the annotation information includes attribute value and historical content of the field to be filled, and historical position of the field to be filled in the template image; generating, according to the attribute value of the field to be filled, the historical content of the field to be filled and the historical position of the field to be filled in the template image, target filling information of the field to be filled; obtaining, according to the target filling information of the field to be filled, an image of an annotated structured document.
Text detection, caret tracking, and active element detection
Detection of typed and/or pasted text, caret tracking, and active element detection for a computing system are disclosed. The location on the screen associated with a computing system where the user has been typing or pasting text, potentially including hot keys or other keys that do not cause visible characters to appear, can be identified and the physical position on the screen where typing or pasting occurred can be provided based on the current resolution of where one or more characters appeared, where the cursor was blinking, or both. This can be done by identifying locations on the screen where changes occurred and performing text recognition and/or caret detection on these locations. The physical position of the typing or pasting activity allows determination of an active or focused element in an application displayed on the screen.