Patent classifications
G06V30/146
System and Computer-Implemented Method for Character Recognition in Payment Card
The present disclosure relates to a system and computer-implemented method for character recognition in a payment card. The method includes receiving an image of a payment card and one or more details associated with the payment card. Further, a derivative of the image is determined based on the one or more details and a horizontal sum of pixel values is determined for a plurality of rows in the image. Furthermore, one or more Regions of Interest (ROIs) are identified in the image by comparing the horizontal sum of pixel values with a predefined first threshold. Subsequently, one or more characters in the one or more ROIs are extracted using one or more peak values in a histogram of the one or more ROIs. Finally, each of the one or more characters extracted from the one or more ROIs is recognized using a trained Artificial Intelligence technique.
System and method of character recognition using fully convolutional neural networks with attention
Embodiments of the present disclosure include a method that obtains a digital image. The method includes extracting a word block from the digital image. The method includes processing the word block by evaluating a value of the word block against a dictionary. The method includes outputting a prediction equal to a common word in the dictionary when a confidence factor is greater than a predetermined threshold. The method includes processing the word block and assigning a descriptor to the word block corresponding to a property of the word block. The method includes processing the word block using the descriptor to prioritize evaluation of the word block. The method includes concatenating a first output and a second output. The method includes predicting a value of the word block.
METHOD AND APPARATUS FOR RECOGNIZING SUBTITLE REGION, DEVICE, AND STORAGE MEDIUM
A method and an apparatus for recognizing a subtitle region, a device, and a storage medium are provided, relating to the field of computer vision technologies of artificial intelligence. The method includes: recognizing a video to obtain n candidate subtitle regions, the candidate subtitle regions being regions in which text contents are displayed in the video, and n being a positive integer; and screening the n candidate subtitle regions according to a subtitle region screening policy to obtain the subtitle region, the subtitle region screening policy being used for determining a candidate subtitle region in which text contents have a repetition rate being lower than a repetition rate threshold and have a longest total display duration as the subtitle region. By using the method and apparatus, device, and system, labor resources required for subtitle region recognition can be saved.
SYSTEM AND METHOD FOR CREATING AND USING IMMERSIVE MEDIA
A device includes a processing system including a processor; and a memory that stores executable instructions that, when executed by the processing system, facilitate performance of operations of loading a user profile for a user consuming undigitized, static media content; identifying an area of interest in the undigitized, static media content; analyzing the area of interest; responsive to the user profile, creating immersive content to enhance the area of interest; and providing the immersive content for presentation to the user.
Systems and methods for obtaining insurance offers using mobile image capture
Systems and methods for using a mobile device to submit an application for an insurance policy using images of documents captured by the mobile device are provided herein. The information is then used by an insurance company to generate a quote which is then displayed to the user on the mobile device. A user captures images of one or more documents containing information needed to complete an insurance application, after which the information on the documents is extracted and sent to the insurance company where a quote for the insurance policy can be developed. The quote can then be transmitted back to the user. Applications on the mobile device are configured to capture images of the documents needed for an insurance application, such as a driver's license, insurance information card or a vehicle identification number (VIN). The images are then processed to extract the information needed for the insurance application.
System for reading contents from a document
Disclosed is a system for reading contents from a document. The system includes a memory unit for storing a set of instructions, and a processing unit is coupled to the memory unit for processing the set of instruction. The set of instructions initiates with a step of rendering the document to retrieve the image of each page and retrieving the digital text and location of each word to create digital text blocks for each word. Running OCR on the images retrieved on rendering the document to an OCR location and creating OCR text blocks from the obtained OCR location of each word. Creating an empty array of text blocks, and adding the digital text blocks to the empty array. Adding one OCR text block to the corresponding empty array on confirming the location and performing the looping step on loop till the document is added with the OCR text block into each empty array.
System for reading contents from a document
Disclosed is a system for reading contents from a document. The system includes a memory unit for storing a set of instructions, and a processing unit is coupled to the memory unit for processing the set of instruction. The set of instructions initiates with a step of rendering the document to retrieve the image of each page and retrieving the digital text and location of each word to create digital text blocks for each word. Running OCR on the images retrieved on rendering the document to an OCR location and creating OCR text blocks from the obtained OCR location of each word. Creating an empty array of text blocks, and adding the digital text blocks to the empty array. Adding one OCR text block to the corresponding empty array on confirming the location and performing the looping step on loop till the document is added with the OCR text block into each empty array.
SYSTEMS AND METHODS FOR GENERATING TEXTUAL INSTRUCTIONS FOR MANUFACTURERS FROM HYBRID TEXTUAL AND IMAGE DATA
A system for generating textual instructions for manufacturers from hybrid textual and image data includes a manufacturing instruction generator that may generate a language processing module from a first training set including at least a training annotated file describing at least a first product to manufacture, the at least an annotated file containing one or more textual data, and at least an instruction set containing one or more manufacturing instructions to manufacture the at least a first product. Manufacturing instruction generator may use the language processing to generate textual instructions for manufacturers from at least an annotated file and may initiate manufacture using the generated manufacturing instructions.
Object detection and image cropping using a multi-detector approach
Systems, methods and computer program products for detecting objects using a multi-detector are disclosed, according to various embodiments. In one aspect, a computer-implemented method includes defining an analysis profile comprising an initial number of analysis cycles dedicated to each of a plurality of detectors, where each detector is independently configured to detect objects according to a unique set of analysis parameters and/or a unique detector algorithm. The method also includes: receiving digital video data that depicts at least one object; analyzing the digital video data using some or all of the detectors in accordance with the analysis profile, where the analyzing produces an analysis result for each detector used in the analysis. Further, the method includes updating the analysis profile by adjusting the number of analysis cycles dedicated to at least one of the detectors based on the analysis results.
METHODS, SYSTEMS, ARTICLES OF MANUFACTURE AND APPARATUS TO DECODE RECEIPTS BASED ON NEURAL GRAPH ARCHITECTURE
Methods, apparatus, systems, and articles of manufacture are disclosed to decode receipts based on neural graph architecture. An example apparatus for decoding receipts includes, vertex feature representation circuitry to extract features from optical-character-recognition (OCR) words, polar coordinate circuitry to: calculate polar coordinates of the OCR words based on respective ones of the extracted features, graph neural network circuitry to generate an adjacency matrix based on the extracted features, post-processing circuitry to traverse the adjacency matrix to generate cliques of OCR processed words, and output circuitry to generate lines of text based on the cliques of OCR processed words.