Patent classifications
G06V30/36
ELECTRONIC DEVICE AND HANDWRITING RECOGNITION METHOD
According to certain embodiments, an electronic device may include a display, a memory, and a processor operatively connected to the display and the memory. The processor may be configured to, while receiving user's touch input in a handwriting area of the display, the user's touch input comprising successive input stokes: output the successive input strokes in the handwriting area on the display; determine a first stroke group including some of the successive input strokes, to determine a first character corresponding to the first stroke group, to output the first stroke group in an output area adjacent to the handwriting area on the display, to determine a second stroke group including at least another input stroke received after the some of the successive input strokes, to determine a second character corresponding to the second stroke group, and to output the second stroke group in the output area, move the first stroke group to on one side of the second stroke group on the display.
METHOD, APPARATUS, AND COMPUTER-READABLE STORAGE MEDIUM FOR RECOGNIZING CHARACTERS IN A DIGITAL DOCUMENT
Method, computer readable medium, and apparatus of recognizing character zone in a digital document. In an embodiment, the method includes classifying a segment of the digital document as including text, calculating at least one parameter value associated with the classified segment of the digital document, determining, based on the calculated at least one parameter value, a zonal parameter value, classifying the segment of the digital document as a handwritten text zone or as a printed text zone based on the determined zonal parameter value and a threshold value, the threshold value being based on a selection of an intersection of a handwritten text distribution profile and a printed text distribution profile, each of the handwritten text distribution profile and the printed text distribution profile being associated with a zonal parameter corresponding to the determined zonal parameter value, and generating, based on the classifying, a modified version of the digital document.
GESTURE STROKE RECOGNITION IN TOUCH-BASED USER INTERFACE INPUT
A method for recognizing gesture strokes in user input, comprising: receiving data generated based on the user input, the data representing a stroke and comprising a plurality of ink points in a rectangular coordinate space and a plurality of timestamps associated respectively with the plurality of ink points; segmenting the plurality of ink points into a plurality of segments each corresponding to a respective sub-stroke of the stroke and comprising a respective subset of the plurality of ink points; generating a plurality of feature vectors based respectively on the plurality of segments; and applying the plurality of feature vectors as an input sequence representing the stroke to a trained stroke classifier to generate a vector of probabilities including a probability that the stroke is a non-gesture stroke and a probability that the stroke is a given gesture stroke of a set of gesture strokes.
System and method of character recognition using fully convolutional neural networks with attention
Embodiments of the present disclosure include a method that obtains a digital image. The method includes extracting a word block from the digital image. The method includes processing the word block by evaluating a value of the word block against a dictionary. The method includes outputting a prediction equal to a common word in the dictionary when a confidence factor is greater than a predetermined threshold. The method includes processing the word block and assigning a descriptor to the word block corresponding to a property of the word block. The method includes processing the word block using the descriptor to prioritize evaluation of the word block. The method includes concatenating a first output and a second output. The method includes predicting a value of the word block.
RECOGNIZING HANDWRITTEN TEXT BY COMBINING NEURAL NETWORKS
A method for recognizing handwritten text is disclosed. The method comprises receiving data comprising a sequence of ink points; applying the received data to a neural network-based sequence classifier trained with a Connectionist Temporal Classification (CTC) output layer using forced alignment to generate an output; generating a character hypothesis as a portion of the sequence of ink points; applying the character hypothesis to a character classifier to obtain a first probability corresponding to the probability that the character hypothesis includes the given character; processing the output of the CTC output layer to determine a second probability corresponding to the probability that the given character is observed within the character hypothesis; and combining the first probability and the second probability to obtain a combined probability corresponding to the probability that the character hypothesis includes the given character.
Method, apparatus, and computer-readable storage medium for recognizing characters in a digital document
Method, computer readable medium, and apparatus of recognizing character zone in a digital document. In an embodiment, the method includes classifying a segment of the digital document as including text, calculating at least one parameter value associated with the classified segment of the digital document, determining, based on the calculated at least one parameter value, a zonal parameter value, classifying the segment of the digital document as a handwritten text zone or as a printed text zone based on the determined zonal parameter value and a threshold value, the threshold value being based on a selection of an intersection of a handwritten text distribution profile and a printed text distribution profile, each of the handwritten text distribution profile and the printed text distribution profile being associated with a zonal parameter corresponding to the determined zonal parameter value, and generating, based on the classifying, a modified version of the digital document.
Ink experience for images
Techniques for an ink experience with images are discussed herein. In various implementations, an image is displayed via an image management application for viewing and/or editing images. In conjunction with interaction scenarios provided via the application, an inking mode for adding inked annotations to the image is enabled. Input to apply one or more inked annotations to the image is obtained, such as via finger touches on a touchscreen, drawing with a stylus, camera-based gestures, or other natural input mechanisms. Responsive to obtaining the input, data blocks corresponding to the one or more inked annotations are appended to an image file as additional data blocks for the image.
HANDWRITING RECOGNITION METHOD AND APPARATUS EMPLOYING CONTENT AWARE AND STYLE AWARE DATA AUGMENTATION
A content aware and style aware neural network based data augmentation model generates augmented data sets to train neural network based handwriting recognition models to recognize individuals' handwriting. In embodiments, the augmented data sets are generated so as to be artificial, and to lack personal or confidential information. In embodiments, the data augmentation model generates content reference sets of individual characters generated in different fonts, and style reference sets of pluralities of characters of a particular style, for example, an individual's handwriting.
Utilizing machine learning models, position based extraction, and automated data labeling to process image-based documents
A device may receive image data that includes an image of a document and lexicon data identifying a lexicon, and may perform an extraction technique on the image data to identify at least one field in the document. The device may utilize form segmentation to automatically generate label data identifying labels for the image data, and may process the image data, the label data, and data identifying the at least one field, with a first model, to identify visual features. The device may process the image data and the visual features, with a second model, to identify sequences of characters, and may process the image data and the sequences of characters, with a third model, to identify strings of characters. The device may compare the lexicon data and the strings of characters to generate verified strings of characters that may be utilized to generate a digitized document.
STRUCTURAL DECOMPOSITION IN HANDWRITING
A method for processing lists in handwriting, comprising: initially classifying each of a plurality of text lines as a distinct text item which is not part of a list; and a classification process comprising a pattern detection in each text line for classifying each text line starting with a predetermined list symbol as a distinct list item which is part of a list; determining an item indentation of each text item with respect to a reference position and determining for each list item a text indentation representing the indentation of text comprised in said list item; and a merging step for merging, as part of a same text item, or as part of a same list item, if predefined conditions are met. A text structure data model may then be generated based on a result of the merging process.