G06V30/18171

Document retrieval through assertion analysis on entities and document fragments

Document retrieval through assertion analysis on entities and document fragments is disclosed. A document is received. Logical structures and entities are extracted from the document by parsing the document. For an entity in the extracted entities, an object representing the entity is created, an assertion made in the document associated with the entity is determined, and the assertion is linked to the object representing the entity. A logical structure from the extracted logical structures and content of the logical structure containing the assertion are identified and linked to the object representing the entity.

METHODS, SYSTEMS, AND COMPUTER PROGRAM PRODUCTS FOR AUTOMATICALLY PROCESSING A CLINICAL RECORD FOR A PATIENT TO DETECT PROTECTED HEALTH INFORMATION (PHI) VIOLATIONS
20230096820 · 2023-03-30 ·

A method includes receiving a record including at least one page and containing clinical information associated with a first patient; receiving respective first patient identification values for one or more patient identification parameters corresponding to the first patient; automatically processing the record to identify first example instances referencing the patient identification parameters including values therefor; automatically processing the record to identify second example instances of the first patient identification values; automatically processing the record to determine whether any of the at least one page contained therein cannot be semantically linked to another one of the at least one page in the record; and assigning a grade to each of the at least one page indicating a degree of confidence that the record does not include clinical information associated with a second patient based on the first example instances, the second example instances, and the determination whether any of the at least one page contained therein cannot be semantically linked to another one of the at least one page in the record.

PRINTED CHARACTER RECOGNITION
20220058416 · 2022-02-24 ·

A computer-implemented method for recognising a printed character string is provided. The method includes receiving an image comprising the character string, the character string comprising a plurality of characters, determining a readability quality for each character in the character string and selecting at least one anchor character based at least in part on the readability quality of the characters in the character string. The identity of the at least one anchor character is determined using a character recognition algorithm and the identity of the character string recognised based on the at least one identified anchor character.

Translation Method and Apparatus and Electronic Device
20210209428 · 2021-07-08 ·

A translation method includes acquiring an image, where the image includes a text to be translated; splitting the text to be translated in the image and acquiring a plurality of target objects, where each of the plurality of target objects includes a word or a phrase of the text to be translated; receiving an input operation for the plurality of target objects, acquiring an object to be translated among the plurality of target objects, and translating the object to be translated.

LAYOUT-AWARE, SCALABLE RECOGNITION SYSTEM
20200285878 · 2020-09-10 ·

Described herein is a mechanism for visual recognition of items or visual search using Optical Character Recognition (OCR) of text in images. Recognized OCR blocks in an image comprise position information and recognized text. The embodiments utilize a location-aware feature vector created using the position and recognized information in each recognized block. The location-aware features of the feature vector utilize position information associated with the block to calculate a weight for the block. The recognized text is used to construct a tri-character gram frequency, inverse document frequency (TGF-IDP) metric using tri-character grams extracted from the recognized text. Features in location-aware feature vector for the block are computed by multiplying the weight and the corresponding TGF-IDF metric. The location-aware feature vector for the image is the sum of the location-aware feature vectors for the individual blocks.

DOCUMENT RETRIEVAL THROUGH ASSERTION ANALYSIS ON ENTITIES AND DOCUMENT FRAGMENTS

Document retrieval through assertion analysis on entities and document fragments is disclosed. A document is received. Logical structures and entities are extracted from the document by parsing the document. For an entity in the extracted entities, an object representing the entity is created, an assertion made in the document associated with the entity is determined, and the assertion is linked to the object representing the entity. A logical structure from the extracted logical structures and content of the logical structure containing the assertion are identified and linked to the object representing the entity.

Layout reconstruction using spatial and grammatical constraints

During an image-analysis technique, the system calculates features by performing image analysis (such as optical character recognition) on a received image of a document. Using these features, as well as spatial and grammatical constraints, the system determines a layout of the document. For example, the layout may be determined using constraint-based optimization based on the spatial and the grammatical constraints. Note that the layout specifies locations of content in the document, and may be used to subsequently extract the content from the image and/or to allow a user to provide feedback on the extracted content by presenting the extracted content to the user in a context (i.e., the determined layout) that is familiar to the user.

Layout-aware, scalable recognition system

Described herein is a mechanism for visual recognition of items or visual search using Optical Character Recognition (OCR) of text in images. Recognized OCR blocks in an image comprise position information and recognized text. The embodiments utilize a location-aware feature vector created using the position and recognized information in each recognized block. The location-aware features of the feature vector utilize position information associated with the block to calculate a weight for the block. The recognized text is used to construct a tri-character gram frequency, inverse document frequency (TGF-IDP) metric using tri-character grams extracted from the recognized text. Features in location-aware feature vector for the block are computed by multiplying the weight and the corresponding TGF-IDF metric. The location-aware feature vector for the image is the sum of the location-aware feature vectors for the individual blocks.

LAYOUT-AWARE, SCALABLE RECOGNITION SYSTEM
20240169751 · 2024-05-23 ·

Described herein is a mechanism for visual recognition of items or visual search using Optical Character Recognition (OCR) of text in images. Recognized OCR blocks in an image comprise position information and recognized text. The embodiments utilize a location-aware feature vector created using the position and recognized information in each recognized block. The location-aware features of the feature vector utilize position information associated with the block to calculate a weight for the block. The recognized text is used to construct a tri-character gram frequency, inverse document frequency (TGF-IDF) metric using tri-character grams extracted from the recognized text. Features in location-aware feature vector for the block are computed by multiplying the weight and the corresponding TGF-IDF metric. The location-aware feature vector for the image is the sum of the location-aware feature vectors for the individual blocks.

METHOD AND SYSTEM FOR CONVERTING AN IMAGE TO TEXT

In a method of converting an input image patch to a text output, a convolutional neural network (CNN) is applied to the input image patch to estimate an n-gram frequency profile of the input image patch. A computer-readable database containing a lexicon of textual entries and associated n-gram frequency profiles is accessed and searched for an entry matching the estimated frequency profile. A text output is generated responsively to the matched entries.