Patent classifications
G06V30/244
Preserving document design using font synthesis
Automatic font synthesis for modifying a local font to have an appearance that is visually similar to a source font is described. A font modification system receives an electronic document including the source font together with an indication of a font descriptor for the source font. The font descriptor includes information describing various font attributes for the source font, which define a visual appearance of the source font. Using the source font descriptor, the font modification system identifies a local font that is visually similar in appearance to the source font by comparing local font descriptors to the source font descriptor. A visually similar font is then synthesized by modifying glyph outlines of the local font to achieve the visual appearance defined by the source font descriptor. The synthesized font is then used to replace the source font and output in the electronic document at the computing device.
Automatic generation of training data for hand-printed text recognition
A method for generating training data for hand-printed text recognition includes obtaining a structured document, obtaining a set of hand-printed character images and database metadata from a database, generating a modified document page image, and outputting a training file. The structured document includes a document page image that includes text characters and document metadata that associates each of the text characters to a document character label. The database metadata associates each of the set of hand-printed character images to a database character label. The modified document page image is generated by iteratively processing each of the text characters. The iterative processing includes determining whether an individual text character should be replaced, selecting a replacement hand-printed character image from the set of hand-printed character images, scaling the replacement hand-printed character image, and inserting the replacement hand-printed character image into the modified document page image.
HANDWRITTEN POSTAGE
The technology described herein provides a handwritten postage that comprises handwriting on a postal item that forms a unique identifier for the postal item (e.g., envelope, postcard, sticker) when analyzed by a computer vision application. The unique identifier is computer derived from the handwritten postage and allows one instance of handwritten postage to be differentiated from all other instances of handwritten postage. The unique identifier may be derived from an image of an envelope that includes an instance of handwritten postage when the handwritten postage is activated. The unique identifier may be formed from a combination of handwriting content (e.g., to and from address), metadata (e.g., date activated), pre-printed content on the postal item (e.g., fiducial marks), post-printed content (e.g., to or from address) and the visual image created by all or a portion of the handwriting. Postage value is added to the handwritten postage through an activation process.
APPLICATION-SPECIFIC OPTICAL CHARACTER RECOGNITION CUSTOMIZATION
A method for customizing an optical character recognition system is disclosed. The optical character recognition system includes a general-purpose decoder configured to convert character images, recognized in a digital image, into text based on a general-purpose text structure. An application-specific customization is received. The application-specific customization includes an application-specific text structure that differs from the general-purpose text structure. A customized model is generated based on the application-specific customization. An enhanced application-specific decoder is generated by modifying the general-purpose decoder to, during run-time execution of the optical character recognition system, leverage the customized model to convert character images demonstrating the application-specific text structure into text.
Optical character recognition systems and methods
The present disclosure is generally directed to systems and methods for executing optical character recognition faster than at least some traditional OCR systems, without sacrificing recognition accuracy. Towards this end, various exemplary embodiments involve the use of a bounding box and a grid-based template to identify certain unique aspects of each of various characters and/or numerals. For example, in one embodiment, the grid-based template can be used to recognize a numeral and/or a character based on a difference in centerline height between the numeral and the character when a monospaced font is used. In another exemplary embodiment, the grid-based template can be used to recognize an individual digit among a plurality of digits based on certain parts of the individual digit being uniquely located in specific portions of the grid-based template.
Optical character recognition systems and methods
The present disclosure is generally directed to systems and methods for executing optical character recognition faster than at least some traditional OCR systems, without sacrificing recognition accuracy. Towards this end, various exemplary embodiments involve the use of a bounding box and a grid-based template to identify certain unique aspects of each of various characters and/or numerals. For example, in one embodiment, the grid-based template can be used to recognize a numeral and/or a character based on a difference in centerline height between the numeral and the character when a monospaced font is used. In another exemplary embodiment, the grid-based template can be used to recognize an individual digit among a plurality of digits based on certain parts of the individual digit being uniquely located in specific portions of the grid-based template.
Method and system for converting font of Chinese character in image, computer device and medium
A method and a system for converting a font of a Chinese character in an image, a computer device and a medium are disclosed. A specific implementation of the method includes: acquiring a stroke of a to-be-converted Chinese character in the image and spatial distribution information of the stroke; and generating a Chinese character in a target font that corresponds to the to-be-converted Chinese character in the image according to the stroke of the to-be-converted Chinese character, the spatial distribution information of the stroke and standard stroke information of the target font, to replace the to-be-converted Chinese character.
HANDWRITTEN CONTENT REMOVING METHOD AND DEVICE AND STORAGE MEDIUM
A handwritten content removing method and device and a storage medium. The handwritten content removing method comprises: acquiring an input image of a text page to be processed, the input image comprising a handwritten region, which comprises a handwritten content (S10); identifying the input image so as to determine the handwritten content in the handwritten region (S11); and removing the handwritten content in the input image so as to obtain an output image (S12).
Machine learning-based inference of granular font properties
A textual properties model is used to infer values for certain font properties of interest given certain text-related data, such as rendered text images. The model may be used for numerous purposes, such as aiding with document layout, identifying font families that are similar to a given font families, and generating new font families with specific desired properties. In some embodiments, the model is trained from a combination of synthetic data that is labeled with values for the font properties of interest, and partially-labeled data from existing “real-world” documents.
INSPECTION APPARATUS, CONTROL METHOD, AND INSPECTION METHOD
An inspection apparatus selects at least one character area, in a first preview image obtained by reading and previewing a print product, sets a direction, for a character in the selected character area, registers the set direction and the character in the selected character area in association with each other, selects at least one character inspection area, in a second preview image obtained by reading and previewing a print product as an inspection target, sets a direction, for a character in the selected character inspection area, rotates the character inspection area to match the set direction, with the direction set for the character in the selected character area, performs character recognition, for the character in the rotated character inspection area, and inspects the character inspection area, based on a result of the character recognition and a result of recognizing the character in the selected character area.