G06V30/293

Multi-modal detection engine of sentiment and demographic characteristics for social media videos

A system and method for determining a sentiment, a gender and an age group of a subject in a video while the video is being played back. The video is separated into visual data and audio data, the video data is passed to a video processing pipeline and the audio data is passed to both an acoustic processing pipeline and a textual processing pipeline. The system and method performs, in parallel, a video feature extraction process in the video processing pipeline, an acoustic feature extraction process in the acoustic processing pipeline, and a textual feature extraction process in the textual processing pipeline. The system and method combines a resulting visual feature vector, acoustic feature vector, and a textual feature vector into a single feature vector, and determines the sentiment, the gender and the age group of the subject by applying the single feature vector to a machine learning model.

TEXT RECOGNITION METHOD AND DEVICE, AND ELECTRONIC DEVICE

A text recognition method includes: acquiring an image including text information, the text information including M characters, M being a positive integer greater than 1; performing text recognition on the image to acquire character information about the M characters; recognizing reading direction information about each character in accordance with the character information about the M characters, the reading direction information being used to indicate a next character corresponding to a current character in a semantic reading order; and ranking the M characters in accordance with the reading direction information about the M characters to acquire a text recognition result of the text information.

Utilizing glyph-based machine learning models to generate matching fonts
11216658 · 2022-01-04 · ·

The present disclosure relates to systems, methods, and non-transitory computer readable media for generating and providing matching fonts by utilizing a glyph-based machine learning model. For example, the disclosed systems can generate a glyph image by arranging glyphs from a digital document according to an ordering rule. The disclosed systems can further identify target fonts as fonts that include the glyphs within the glyph image. The disclosed systems can further generate target glyph images by arranging glyphs of the target fonts according to the ordering rule. Based on the glyph image and the target glyph images, the disclosed systems can utilize a glyph-based machine learning model to generate and compare glyph image feature vectors. By comparing a glyph image feature vector with a target glyph image feature vector, the font matching system can identify one or more matching glyphs.

Text independent writer verification method and system

A device, method, and non-transitory computer readable medium are described. The method includes receiving a dataset including hand written Arabic words and hand written Arabic alphabets from one or more users. The method further includes removing whitespace around alphabets in the hand written Arabic words and the hand written Arabic alphabets in the dataset. The method further includes splitting the dataset into a training set, a validation set, and a test set. The method further includes classifying one or more user datasets from the training set, the validation set, and the test set. The method further includes identifying the target user from the one or more user datasets. The identification of the target user includes a verification accuracy of the hand written Arabic words being larger than a verification accuracy threshold value.

Text recognition in image

According to implementations of the subject matter described herein, there is provided a solution for text recognition in an image. In this solution, a target text line area, which is expected to include a text to be recognized, is determined from an image. Probability distribution information of a character model element(s) present in the target text line area is determined using a single character model. The single character model is trained based on training text line areas and respective ground-truth texts in the training text line areas. Texts in the training text line areas are arranged in different orientations, and/or the ground-truth texts comprise texts are related to various languages (e.g., texts related to a Latin and an Eastern languages). The text in the target text line area can be determined based on the determined probability distribution information. The single character model enables more efficient and convenient text recognition.

Method for table extraction from journal literature based on text state characteristics

A method for table extraction from journal literature based on text state characteristics is disclosed. The method includes: constructing a table model according to characteristics of tables in journal literature, where the table model includes two parts: a table caption and table content, building a text line set, table detection, table data positioning, table reconstruction, building a cell data set, restoring data of merged cells, checking the cell data set, and outputting table data. The method is particularly designed based on characteristics of tables such as three-line tables widespread in PDF journal literature, which can realize accurate and correct extraction of specific tables in the PDF journal literature, and especially can ensure a logic relationship of a three-line table. The whole process neither requires manual intervention or interaction nor requires table selection, so that the whole extraction process is automatic.

Method, apparatus, device and computer readable storage medium for recognizing aerial handwriting

A method, an apparatus, a device and a computer-readable storage medium for recognizing aerial handwriting are provided. The method may include detecting a palm region of a user in a two-dimensional grayscale image; detecting a fingertip in the two-dimensional gray-scale image based on the palm area; determining a spatial trajectory of the fingertip based on a set of two-dimensional gray-scale images, the set of two-dimensional gray-scale images including the two-dimensional gray-scale image; and recognizing handwritten content of the user based on the spatial trajectory. A two-dimensional gray-scale image is used to recognize and track the spatial trajectory of the fingertip, which may speed up aerial handwriting recognition, and has low processing performance requirements for the device, while also ensuring high accuracy.

METHOD AND SYSTEM FOR IDEOGRAM CHARACTER ANALYSIS
20220222292 · 2022-07-14 ·

Ideogram character analysis includes partitioning an original ideogram character into strokes and mapping each stroke to a corresponding stroke identifier (id) to create an original stroke id sequence that includes stroke identifiers. A candidate ideogram character that has a candidate stroke id sequence within a threshold distance to the original stroke id sequence is selected. One or more embodiments may create a new phrase by replacing the original ideogram character with the candidate ideogram character in a search phrase. One or more embodiments perform a search using the search phrase and the new phrase to obtain a result and present the result. One or more embodiments may replace an original ideogram character in a character recognized document with the candidate ideogram character and store the character recognized document.

IDENTIFYING MATCHING FONTS UTILIZING DEEP LEARNING
20220083772 · 2022-03-17 ·

The present disclosure relates to systems, methods, and non-transitory computer readable media for generating and providing matching fonts by utilizing a glyph-based machine learning model. For example, the disclosed systems can generate a glyph image by arranging glyphs from a digital document according to an ordering rule. The disclosed systems can further identify target fonts as fonts that include the glyphs within the glyph image. The disclosed systems can further generate target glyph images by arranging glyphs of the target fonts according to the ordering rule. Based on the glyph image and the target glyph images, the disclosed systems can utilize a glyph-based machine learning model to generate and compare glyph image feature vectors. By comparing a glyph image feature vector with a target glyph image feature vector, the font matching system can identify one or more matching glyphs.

Visually-aware encodings for characters
11275969 · 2022-03-15 · ·

In some embodiments, a method inputs a set of images into a network and trains the network based on a classification of the set of images to one or more characters in a set of characters. The method obtains a set of encodings for the one or more characters based on a layer of the network that restricts the output of the layer to a number of values. Then, the method stores the set of encodings for the one or more characters, wherein an encoding in the set of encodings is retrievable when a corresponding character is determined.