Patent classifications
G06V30/19173
CHARACTER RECOGNITION MODEL TRAINING METHOD AND APPARATUS, CHARACTER RECOGNITION METHOD AND APPARATUS, DEVICE AND STORAGE MEDIUM
The present disclosure provides a character recognition model training method and apparatus, a character recognition method and apparatus, a device and a medium, relating to the technical field of artificial intelligence, and specifically to the technical fields of deep learning, image processing and computer vision, which can be applied to scenarios such as character detection and recognition technology. The specific implementing solution is: partitioning an untagged training sample into at least two sub-sample images; dividing the at least two sub-sample images into a first training set and a second training set; where the first training set includes a first sub-sample image with a visible attribute, and the second training set includes a second sub-sample image with an invisible attribute; performing self-supervised training on a to-be-trained encoder by taking the second training set as a tag of the first training set, to obtain a target encoder.
INTELLIGENT SORTING OF TIME SERIES DATA FOR IMPROVED CONTEXTUAL MESSAGING
Systems for intelligent sorting of time series data for improved contextual messaging are included herein. An intelligent sorting server may receive time series data comprising a plurality of chat messages. The intelligent sorting server may determine a first order of the plurality of chat messages based on a chronologic order. The intelligent sorting server may use one or more machine learning classifiers to identify candidates for reordering the chat messages. The intelligent sorting server may generate a second order of the chat messages based on the identified candidates for reordering. Accordingly, the intelligent sorting server may present, to a client device, a transcript of the chat messages associated with the second order and an indication that at least one chat message has been repositioned.
Context-based object location via augmented reality device
Devices, computer-readable media, and methods for providing an enhanced indication of an object that is located via a visual feed in accordance with a user context are disclosed. For instance, in one example, a processing system including at least one processor may detect a user context from a visual feed, locate an object via the visual feed in accordance with the user context, and provide an enhanced indication of the object via an augmented reality display.
Identifying image aesthetics using region composition graphs
The disclosed computer-implemented method may include generating a three-dimensional (3D) feature map for a digital image using a fully convolutional network (FCN). The 3D feature map may be configured to identify features of the digital image and identify an image region for each identified feature. The method may also include generating a region composition graph that includes the identified features and image regions. The region composition graph may be configured to model mutual dependencies between features of the 3D feature map. The method may further include performing a graph convolution on the region composition graph to determine a feature aesthetic value for each node according to the weightings in the node's weighted connecting segments, and calculating a weighted average for each node's feature aesthetic value to provide a combined level of aesthetic appeal for the digital image. Various other methods, systems, and computer-readable media are also disclosed.
SYSTEMS AND METHODS FOR CLASSIFYING DOCUMENTS
A system may iteratively scan a portion of a document, extract first data from the portion of the document, and determine, using a trained model, whether the first data corresponds to one or more document types based on one or more confidence thresholds. The system may repeat this process, increasing the portion of the document scanned by a predetermined amount each iteration, until the first data corresponds to the one or more document types based on the one or more confidence thresholds. Responsive to determining the first data corresponds to the one or more document types based on the one or more confidence thresholds, the system may cause a graphical user interface (GUI) of a user device to display a notification indicating a document type match.
SYSTEMS AND METHODS FOR GENERATING TEXTUAL INSTRUCTIONS FOR MANUFACTURERS FROM HYBRID TEXTUAL AND IMAGE DATA
A system for generating textual instructions for manufacturers from hybrid textual and image data includes a manufacturing instruction generator that may generate a language processing module from a first training set including at least a training annotated file describing at least a first product to manufacture, the at least an annotated file containing one or more textual data, and at least an instruction set containing one or more manufacturing instructions to manufacture the at least a first product. Manufacturing instruction generator may use the language processing to generate textual instructions for manufacturers from at least an annotated file and may initiate manufacture using the generated manufacturing instructions.
SEARCH DEVICE, SEARCH SYSTEM, SEARCH METHOD, AND STORAGE MEDIUM
According to one embodiment, a search device generates a character string image of a first character string by using the first character string. The search device inputs the character string image to a classifier. The classifier outputs a classification of a character string according to an input of an image. The search device outputs an other character string based on a classification result of the classifier. The other character string is different from the first character string.
Text classification
A text classifying apparatus (100), an optical character recognition unit (1), a text classifying method (S220) and a program are provided for performing the classification of text. A segmentation unit (110) segments an image into a plurality of lines of text (401-412; 451-457; 501-504; 701-705) (S221). A selection unit (120) selects a line of text from the plurality of lines of text (S222-S223). An identification unit (130) identifies a sequence of classes corresponding to the selected line of text (S224). A recording unit (140) records, for the selected line of text, a global class corresponding to a class of the sequence of classes (S225-S226). A classification unit (150) classifies the image according to the global class, based on a confidence level of the global class (S227-S228).
Machine learning technique for automatic modeling of multiple-valued outputs
A method and system are disclosed for training a model that implements a machine-learning algorithm. The technique utilizes latent descriptor vectors to change a multiple-valued output problem into a single-valued output problem and includes the steps of receiving a set of training data, processing, by a model, the set of training data to generate a set of output vectors, and adjusting a set of model parameters and component values for at least one latent descriptor vector in the plurality of latent descriptor vectors based on the set of output vectors. The set of training data includes a plurality of input vectors and a plurality of desired output vectors, and each input vector in the plurality of input vectors is associated with a particular latent descriptor vector in a plurality of latent descriptor vectors. Each latent descriptor vector comprises a plurality of scalar values that are initialized prior to training the model.
Object detection and image cropping using a multi-detector approach
Systems, methods and computer program products for detecting objects using a multi-detector are disclosed, according to various embodiments. In one aspect, a computer-implemented method includes defining an analysis profile comprising an initial number of analysis cycles dedicated to each of a plurality of detectors, where each detector is independently configured to detect objects according to a unique set of analysis parameters and/or a unique detector algorithm. The method also includes: receiving digital video data that depicts at least one object; analyzing the digital video data using some or all of the detectors in accordance with the analysis profile, where the analyzing produces an analysis result for each detector used in the analysis. Further, the method includes updating the analysis profile by adjusting the number of analysis cycles dedicated to at least one of the detectors based on the analysis results.