Patent classifications
G06V30/15
Image Enhancement in a Genealogy System
Methods, systems, and computer-program products for image enhancement include receiving an image and optionally a user request, classify the image, crop image components of the image, restore cropped image components of the image, colorized restored image components, and reconstruct the image from the colorized, restored image components and other components. The other components may include text components that are restored in a separate treatment pipeline.
DATA GENERATION APPARATUS, DATA GENERATION METHOD, AND COMPUTER-READABLE RECORDING MEDIUM
A data generation apparatus includes: a separation unit that separates a serial number region and a background region from an original image of a paper currency that includes a serial number; a character image acquisition unit that identifies each of characters included in the separated serial number region, and acquires a character image of each of the identified characters; a background image acquisition unit that acquires a background image by complementing the serial number region in the separated background region; a pre-processing unit that generates a serial number image by combining the character images; an incorporation unit that incorporates the serial number image at a position corresponding to the serial number image in the background image; and an output unit that outputs image data in which the serial number that is combined by the pre-processing unit is associated with an incorporated image generated by the incorporation unit.
METHOD AND APPARATUS FOR CHARACTER SELECTION BASED ON CHARACTER RECOGNITION, AND TERMINAL DEVICE
Embodiments of this application are applicable to the field of artificial intelligence technologies, and provide a method and an apparatus for character selection based on character recognition, and a terminal device. The method includes: obtaining a connectionist temporal classification sequence corresponding to text content in an original picture; calculating character coordinates of each character in the connectionist temporal classification sequence; mapping the character coordinates of each character to the original picture, to obtain target coordinates of each character in the original picture; and generating a character selection control in the original picture based on the target coordinates. The character selection control is used to indicate a user to select a character in the original picture. By using the foregoing method, when the user manually selects the character, precision of positioning the character can be improved, and efficiency and accuracy of manually selecting the character can be improved.
Image processing apparatus, method, and storage medium
A binary image of an input image is generated, and a character region within the binary image and a region surrounding each character are acquired as character segmentation rectangle information. A thinning process is executed on a region within the binary image which is identified based on the character segmentation rectangle information to acquire a thinned image. An edge detected image of the region identified based on the character segmentation rectangle information is acquired. Whether each character identified based on the character segmentation rectangle information is a character to be separated from a background by the binarization process or not is determined based on a result of a logical AND of the thinned image and the edge detected image.
IMAGE PROCESSING DEVICE, IMAGE READING APPARATUS AND NON-TRANSITORY COMPUTER READABLE MEDIUM STORING PROGRAM
An image processing device includes: an obtaining unit that obtains image information of a second region to detect an erecting direction of an image formed on a document, the second region being defined in the image in advance according to a criterion different from a criterion for defining a first region in the image, in which character recognition is performed; and an output unit that outputs character information of the first region, the character information being recognized in accordance with the erecting direction of the image obtained from the image information.
INTELLIGENT SCORING METHOD AND SYSTEM FOR TEXT OBJECTIVE QUESTION
An intelligent scoring method and system for a text objective question, the method comprising: acquiring an answer image of a text objective question (101); segmenting the answer image to obtain one or more segmentation results of an answer string to be identified (102); determining whether any of the segmentation results has the same number of characters as the standard answer (103); if no, the answer is determined to be wrong (106); otherwise, calculating identification confidence of the segmentation result having the same number of words as the standard answer, and/or calculating the identification confidence of respective characters in the segmentation result having the same number of words as the standard answer (104); determining whether the answer is correct according to the calculated identification confidence (105). The method can automatically score text objective questions, thus reducing consumption of human resource, and improving scoring efficiency and accuracy.
IMAGE PROCESSING APPARATUS, IMAGE PROCESSING METHOD, AND NON-TRANSITORY STORAGE MEDIUM
An image processing apparatus that generates an image for character recognition from a read image includes at least one memory that stores instructions, and at least one processor that executes the instructions to perform extracting of an area of handwritten character information and an area of printed character information from the read image, clipping of a partial image of the area of handwritten character information and a partial image of the area of printed character information out of the read image, and generating of the image for character recognition by combining the partial image of the area of handwritten character information and the partial image of the area of printed character information being associated with each other.
IMAGE, PATTERN AND CHARACTER RECOGNITION
Some aspects of the disclosure provide a method for image processing. The method includes receiving one or more first images corresponding to first portions in a section of characters for recognition, splicing the one or more first images into a first intermediate spliced image and performing a first intermediate character recognition on the first intermediate spliced image based on a first optical character recognition model. The first intermediate character recognition generates a first intermediate recognition result for display. The method further includes performing a final character recognition on a final spliced image corresponding to the section of characters for recognition based on a second optical character recognition model that is different from the first optical character recognition model. The final character recognition generates a final recognition result of the section of characters. Apparatus and non-transitory computer-readable storage medium counterpart embodiments are also contemplated.
Neural Network-based Optical Character Recognition
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for neural network-based optical character recognition. An embodiment of the system may generate a set of bounding boxes based on reshaped image portions that correspond to image data of a source image. The system may merge any intersecting bounding boxes into a merged bounding box to generate a set of merged bounding boxes indicative of image data portions that likely portray one or more words. Each merged bounding box may be fed by the system into a neural network to identify one or more words of the source image represented in the respective merged bounding box. The one or more identified words may be displayed by the system according to a standardized font and a confidence score.
Handwriting detector, extractor, and language classifier
Disclosed are methods for handwriting recognition. In some aspects, an image representing a page of a sample document is analyzed to identify a region having indications of handwriting. The region is analyzed to determine frequencies of a plurality of geometric features within the region. The frequencies may be compared to profiles or histograms of known language types, to determine if there are similarities between the frequencies in the sample document relative to those of the known language types. In some aspects, machine learning may be used to characterize the document as a particular language type based on the frequencies of the geometric features.