Patent classifications
G06V30/1478
Method, system, digital camera and asic for geometric image transformation based on text line searching
A method, system and/or a digital camera provide a geometrical transformation of deformed images of documents comprising text, by text line tracking, resulting in an image comprising parallel text lines. The transformed image is provided as an input to an OCR program either running in a computer system or in a processing element comprised in said digital camera.
Payment card OCR with relaxed alignment
Extracting financial card information with relaxed alignment comprises a method to receive an image of a card, determine one or more edge finder zones in locations of the image, and identify lines in the one or more edge finder zones. The method further identifies one or more quadrilaterals formed by intersections of extrapolations of the identified lines, determines an aspect ratio of the one or more quadrilateral, and compares the determined aspect ratios of the quadrilateral to an expected aspect ratio. The method then identifies a quadrilateral that matches the expected aspect ratio and performs an optical character recognition algorithm on the rectified model. A similar method is performed on multiple cards in an image. The results of the analysis of each of the cards are compared to improve accuracy of the data.
APPARATUS AND METHOD FOR DOCUMENT IMAGE ORIENTATION DETECTION
An apparatus and method for document image orientation detection. When a ratio of a difference between similarities between a current text line and reference samples in two selected candidate orientations is greater than or equal to a first threshold value, 1 is added to a voting value of a candidate orientation corresponding to the largest similarity in the orientations, and when the ratio of the difference is less than the first threshold value, a product of the ratio of the difference and a parameter related to the first threshold value is added to the voting value of the candidate orientation corresponding to the largest similarity in the orientations. Hence, setting a voting value can efficiently lower influence of noise text lines, low-quality text lines and unsupported text lines on the orientation detection, thereby achieving accurate document image orientation detection.
METHOD FOR PROCESSING INFORMATION FROM A HAND-HELD SCANNING DEVICE
A method implemented on a mobile device for selecting information on a paper document using a hand-held scanning device. The method provides an interface to the user, receives information from the hand-held pen scanning device connected to the mobile device and determines if the received information is valid information for the type of information to be selected. The invention further relates to using a remote computer for performing image processing and data extraction when more powerful resources are needed.
Apparatus and method for recommending learning using optical character recognition
There are provided a learning recommendation apparatus and method for detecting a problem from an image through character recognition and providing at least one sub-topic learning among a plurality of sub-topic learnings related to the detected problem. The provided learning recommendation apparatus recommends, as a recommendation target, a plurality of learning topics including the concept of a formula which has been read through the character recognition for an image, wherein a priority order is set to the plurality of learning topics based on the concept distance between the learning topic and the learning history, and the learning topics are recommended so that the learning topic having a higher priority order is located at a higher position.
Text recognition system with feature recognition and method of operation thereof
A text recognition system and method of operation thereof including: a storage unit for storing a text unit; and a processing unit, connected to the storage unit, the processing unit including: a communication interface for receiving the text unit, a feature detection module for determining an isolated feature of the text unit, an angle detection module for determining angle features of the text unit, a feature vector module for generating a feature vector for the text unit based on the isolated feature and the angle features, and a text recognition module for determining recognized text using the feature vector for display on a display interface.
EXTRACTING CARD DATA FROM MULTIPLE CARDS
Extracting financial card information with relaxed alignment comprises a method to receive an image of a card, determine one or more edge finder zones in locations of the image, and identify lines in the one or more edge finder zones. The method further identifies one or more quadrilaterals formed by intersections of extrapolations of the identified lines, determines an aspect ratio of the one or more quadrilateral, and compares the determined aspect ratios of the quadrilateral to an expected aspect ratio. The method then identifies a quadrilateral that matches the expected aspect ratio and performs an optical character recognition algorithm on the rectified model. A similar method is performed on multiple cards in an image. The results of the analysis of each of the cards are compared to improve accuracy of the data.
IMAGE PROCESSING METHOD
An image processing method for a picture of a participant, photographed in an event, such as a marathon race, increases the accuracy of recognition of a race bib number by performing image processing on a detected race bib area, and associates the recognized race bib number with a person included in the picture. This image processing method detects a person from an input image, estimates an area in which a race bib exists based on a face position of the detected person, detects an area including a race bib number from the estimated area, performs image processing on the detected area to thereby perform character recognition of the race bib number from an image subjected to image processing, and associates the result of character recognition with the input image.
Systems and methods for processing images
A apparatus and method are provided for processing images. In one embodiment, the apparatus includes an image sensor configured to capture real time images from an environment of a user. The apparatus also includes a mobile power source, and at least one processor device configured to process, at an initial resolution, images to determine existence of a trigger, and access rules associating image context with image capture resolution to enable images of a first context to be processed at a lower capture resolution than images of a second context. The processor device analyzes at least one first image, selects a first image capture resolution rule, and applies the first image capture resolution rule to a subsequent captured image. The processor device analyzes at least one second image, selects a second image capture resolution rule, and applies the second image capture resolution rule to a second subsequent captured image.
Deep-learning-based system and process for image recognition
Disclosed are methods and systems for using artificial intelligence (AI) for image recognition by using predefined coordinates to extract a portion of a received image, the extracted portion comprising a word to be identified having at least a first letter and a second letter; executing an image recognition protocol to identify the first letter; when the server is unable to identify the second letter, the server executes an AI model having a nodal data structure to identify the second letter based upon the identified first letter, the nodal data structure comprising a set of nodes where each node represents a letter, each node connected to at least one other node, wherein connection of a first node to a second node corresponds to a probability that a letter corresponding to the second node is used in a word subsequent to a letter corresponding to the first node.