Patent classifications
G06V30/40
Optical character recognition of documents having non-coplanar regions
Systems and methods for performing OCR of an image depicting text symbols and imaging a document having a plurality of planar regions are disclosed. An example method comprises: receiving a first image of a document having a plurality of planar regions and one or more second images of the document; identifying a plurality of coordinate transformations corresponding to each of the planar regions of the first image of the document; identifying, using the plurality of coordinate transformations, a cluster of symbol sequences of the text in the first image and in the one or more second images; and producing a resulting OCR text comprising a median symbol sequence for the cluster of symbol sequences.
Information processing apparatus and non-transitory computer readable medium
An information processing apparatus includes a processor configured to: acquire an image corresponding to a key character string from a target image in response to the key character string that serves as a character string specified beforehand as a key and is acquired from results of character recognition performed on the target image including character strings; extract, by using results of acquiring the image corresponding to the key character string, from the results of the character recognition a value character string that serves as a character string indicating a value corresponding to the key character string; and output the key character string and the value character string corresponding to the key character string.
Information processing apparatus and non-transitory computer readable medium
An information processing apparatus includes a processor configured to: acquire an image corresponding to a key character string from a target image in response to the key character string that serves as a character string specified beforehand as a key and is acquired from results of character recognition performed on the target image including character strings; extract, by using results of acquiring the image corresponding to the key character string, from the results of the character recognition a value character string that serves as a character string indicating a value corresponding to the key character string; and output the key character string and the value character string corresponding to the key character string.
Adversarial network for transforming handwritten text
Described herein are systems, methods, and other techniques for training a generative adversarial network (GAN) to perform an image-to-image transformation for recognizing text. A pair of training images are provided to the GAN. The pair of training images include a training image containing a set of characters in handwritten form and a reference training image containing the set of characters in machine-recognizable form. The GAN includes a generator and a discriminator. The generated image is generated using the generator based on the training image. Update data is generated using the discriminator based on the generated image and the reference training image. The GAN is trained by modifying one or both of the generator and the discriminator using the update data.
Automated non-native table representation annotation for machine-learning models
One embodiment provides a method, including: receiving two documents, one of the two documents having at least one table that includes the same information as a corresponding table in the other of the two documents, wherein (i) one of the two documents comprises the at least one table in an unstructured table representation and (ii) the other of the two documents comprises the at least one table in a structured table representation; identifying text elements within the at least one table in the unstructured table representation; matching the identified text elements with table elements within the at least one table in the structured table representation; and annotating the at least one table in the structured table representation based upon the matches between the table elements and text elements.
Method and system for detecting drift in image streams
Methods and systems disclosed herein may quantify a representation of a type of input an image analysis system should expect. The image analysis system may be trained on the type of input the image analysis system should expect using a first image stream. A first model of the type of input that the image analysis system should expect may be built from the first image stream. After the first model is built, a second image, or a second image stream, may be compared to the first model to determine a difference between the second image, or second image stream, and the first image stream. When the difference is greater than or equal to a threshold, a drift may be detected and steps may be taken to determine the cause of the drift.
Method and system for detecting drift in image streams
Methods and systems disclosed herein may quantify a representation of a type of input an image analysis system should expect. The image analysis system may be trained on the type of input the image analysis system should expect using a first image stream. A first model of the type of input that the image analysis system should expect may be built from the first image stream. After the first model is built, a second image, or a second image stream, may be compared to the first model to determine a difference between the second image, or second image stream, and the first image stream. When the difference is greater than or equal to a threshold, a drift may be detected and steps may be taken to determine the cause of the drift.
Classifying and grouping sentences using machine learning
A device that includes an enterprise data indexing engine (EDIE) configured to receive a set of sentences and to compare the words in the sentences to a set of predefined keywords. The EDIE is further configured to identify one or more sentences that do not contain any of the keywords and to associate the identified sentences with a first classification type. The EDIE is further configured to identify a sentence that contains one or more keywords and to associate the sentence with a second classification type. The EDIE is further configured to link together the sentence that is associated with the second classification type and the sentences that are associated with the first classification type.
Classifying and grouping sentences using machine learning
A device that includes an enterprise data indexing engine (EDIE) configured to receive a set of sentences and to compare the words in the sentences to a set of predefined keywords. The EDIE is further configured to identify one or more sentences that do not contain any of the keywords and to associate the identified sentences with a first classification type. The EDIE is further configured to identify a sentence that contains one or more keywords and to associate the sentence with a second classification type. The EDIE is further configured to link together the sentence that is associated with the second classification type and the sentences that are associated with the first classification type.
IDENTIFYING TROUBLED CONTRACTS
In an approach for identifying troubled contracts using a health score, a processor receives a contract. A processor identifies a list of requirements of the contract using a first Natural Language Processing technique. A processor trains a model to recognize the list of requirements of the contract. A processor receives at least one deliverable document associated with the contract. A processor applies a plurality of health metrics to the at least one deliverable document associated with the contract to identify evidence of completion of each requirement of the list of requirements of the contract. A processor outputs the health score for each requirement of the list of requirements of the contract.