G06V30/18057

RESERVOIR COMPUTING NEURAL NETWORKS BASED ON SYNAPTIC CONNECTIVITY GRAPHS

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for implementing a reservoir computing neural network. In one aspect there is provided a reservoir computing neural network comprising: (i) a brain emulation sub-network, and (ii) a prediction sub-network. The brain emulation sub-network is configured to process the network input in accordance with values of a plurality of brain emulation sub-network parameters to generate an alternative representation of the network input. The prediction sub-network is configured to process the alternative representation of the network input in accordance with values of a plurality of prediction sub-network parameters to generate the network output. The values of the brain emulation sub-network parameters are determined before the reservoir computing neural network is trained and are not adjusting during training of the reservoir computing neural network.

Computer vision systems and methods for information extraction from text images using evidence grounding techniques

Computer vision systems and methods for text classification are provided. The system detects a plurality of text regions in an image and generates a bounding box for each detected text region. The system utilizes a neural network to recognize text present within each bounding box and classifies the recognized text, based on at least one extracted feature of each bounding box and the recognized text present within each bounding box, according to a plurality of predefined tags. The system can associate a key with a value and return a key-value pair for each predefined tag.

System and method for fashion attributes extraction

A system and a method for training an inference model using a computing device. The method includes: providing a text-to-vector converter; providing the inference model and pre-training the inference model using labeled fashion entries; providing non-labeled fashion entries; separating each of the non-labeled fashion entries into a target image and target text; converting the target text into a category vector and an attribute vector using the text-to-vector converter; processing the target image using the inference model to obtain processed target image and target image label; comparing the category vector to the target image label; when the category vector matches the target image label, updating the target image label based on the category vector and the attribute vector to obtain updated label; and retraining the inference model using the processed target image and the updated label.

Method and device for vertebra localization and identification

A vertebra localization and identification method includes: receiving one or more images of vertebrae of a spine; applying a machine learning model on the one or more images to generate three-dimensional (3-D) vertebra activation maps of detected vertebra centers; performing a spine rectification process on the 3-D vertebra activation maps to convert each 3-D vertebra activation map into a corresponding one-dimensional (1-D) vertebra activation signal; performing an anatomically-constrained optimization process on each 1-D vertebra activation signal to localize and identify each vertebra center in the one or more images; and outputting the one or more images, wherein on each of the one or more outputted images, a location and an identification of each vertebra center are specified.

MULTI-LAYER NEURAL NETWORK AND CONVOLUTIONAL NEURAL NETWORK FOR CONTEXT SENSITIVE OPTICAL CHARACTER RECOGNITION
20230019919 · 2023-01-19 ·

Aspects of the disclosure relate to OCR. A computing platform may train, using historical images, a CNN and a RNN to perform OCR/identify characters in context. The computing platform may receive an image of a document, and may input the image into the CNN, which may cause the CNN to output OCR information for the image and a confidence score. Based on identifying that the confidence score exceeds a confidence threshold, the computing platform may store the OCR information to enable subsequent access of a digital version of the document. Based on identifying that the confidence score does not exceed the confidence threshold, the computing platform may: 1) input the OCR information into the first RNN, which may cause the first RNN to output contextual OCR information for the image, and 2) store the contextual OCR information to enable subsequent access of the digital version of the document.

HANDWRITING RECOGNITION PIPELINES FOR GENEALOGICAL RECORDS

Disclosed herein relates to example embodiments for recognizing handwritten information in a genealogical record. A computing server may receive a genealogical record. The genealogical record may take the form of an image of a physical form having a structured layout, fields, and handwritten information. The computing server may divide the genealogical record into a plurality of areas based on the structured layout. The computing server may identify, for a particular area, a type of field that is included within the particular area. The computing server may select a handwriting recognition model for identifying the handwritten information in the particular area. The handwriting recognition model may be selected based on the type of the field. The computing server may input an image of the particular area to the handwriting recognition model to generate text of the handwritten information. The computing server may store the text of the handwritten information.

IDENTIFY CARD NUMBER
20230215201 · 2023-07-06 ·

A card number recognition method and apparatus, a storage medium, and an electronic device are disclosed. The method includes: obtaining distribution format information of character bits of a card number sequence, where the distribution format information includes character bit spacing information of the card number sequence; recognizing a character sequence in a target image through a neural network model trained in advance, and obtaining character bit spacing information of the recognized character sequence; determining whether the character bit spacing information of the recognized character sequence is consistent with the character bit spacing information in the obtained distribution format information; and if the character bit spacing information of the character sequence is consistent with the character bit spacing information in the obtained distribution format information, determining that the recognized character sequence is target card numbers.

Apparatus and method for detecting elements of an assembly
20230215154 · 2023-07-06 ·

The disclosure relates to apparatuses and methods for detecting elements of an assembly, such as electrical components in a printed circuit board. First and second artificially intelligent classifiers are provided for detecting elements in a high-resolution image of the assembly, wherein the first artificially intelligent classifier is pre-trained to detect first elements and the second artificially intelligent classifier is pre-trained to detect second elements, each of the first elements having a size within a first size range, and each of the second elements having a size within a second size range, in which the first size range includes elements having a size that is greater than the size of elements included within the second size range. The second artificially intelligent classifier can be prevented from subsequently searching for elements within bounding boxes previously obtained by the first artificially intelligent classifier.

Recurrent deep neural network system for detecting overlays in images
11551435 · 2023-01-10 · ·

In one aspect, an example method includes a processor (1) applying a feature map network to an image to create a feature map comprising a grid of vectors characterizing at least one feature in the image and (2) applying a probability map network to the feature map to create a probability map assigning a probability to the at least one feature in the image, where the assigned probability corresponds to a likelihood that the at least one feature is an overlay. The method further includes the processor determining that the probability exceeds a threshold, and responsive to the processor determining that the probability exceeds the threshold, performing a processing action associated with the at least one feature.

CHARACTER RECOGNITION MODEL TRAINING METHOD AND APPARATUS, CHARACTER RECOGNITION METHOD AND APPARATUS, DEVICE AND STORAGE MEDIUM

The present disclosure provides a character recognition model training method and apparatus, a character recognition method and apparatus, a device and a medium, relating to the technical field of artificial intelligence, and specifically to the technical fields of deep learning, image processing and computer vision, which can be applied to scenarios such as character detection and recognition technology. The specific implementing solution is: partitioning an untagged training sample into at least two sub-sample images; dividing the at least two sub-sample images into a first training set and a second training set; where the first training set includes a first sub-sample image with a visible attribute, and the second training set includes a second sub-sample image with an invisible attribute; performing self-supervised training on a to-be-trained encoder by taking the second training set as a tag of the first training set, to obtain a target encoder.