Patent classifications
G06V30/19127
DEVICE ANTI-SURVEILLANCE SYSTEM
A method comprises receiving one or more inputs captured by a camera of a device, and determining, using one or more machine learning models, whether the one or more inputs depict at least one object configured to capture a visual representation of a screen of the device. A recommendation is generated responsive to an affirmative determination, the recommendation comprising at least one action to prevent the capture of the visual representation of the screen of the device.
METHOD FOR TRAINING A FONT GENERATION MODEL, METHOD FOR ESTABLISHING A FONT LIBRARY, AND DEVICE
Provided are a method for training a font generation model, a method for establishing a font library, and a device. The method for training a font generation model includes the following steps. A source-domain sample character is input into the font generation model to obtain a first target-domain generated character. The first target-domain generated character is input into a font recognition model to obtain the target adversarial loss of the font generation model. The model parameter of the font generation model is updated according to the target adversarial loss.
DYNAMIC DETECTION AND RECOGNITION OF MEDIA SUBJECTS
A system for indexing animated content receives detections extracted from a media file, where each one of the detections includes an image extracted from a corresponding frame of the media file that corresponds to a detected instance of an animated character. The system determines, for each of the received detections, an embedding defining a set of characteristics for the detected instance. The embedding associated with each detection is provided to a grouping engine that is configured to dynamically configure at least one grouping parameter based on a total number of the detections received. The grouping engine is also configured to sort the detections into groups using the grouping parameter and the embedding for each detection. A character ID is assigned to each one of the groups of detections, and the system indexes the groups of detections in a database in association with the character ID assigned to each group.
TEXT EXTRACTION METHOD, TEXT EXTRACTION MODEL TRAINING METHOD, ELECTRONIC DEVICE AND STORAGE MEDIUM
A text extraction method and a text extraction model training method are provided. The present disclosure relates to the technical field of artificial intelligence, in particular to the technical field of computer vision. An implementation of the method comprises: obtaining a visual encoding feature of a to-be-detected image; extracting a plurality of sets of multimodal features from the to-be-detected image, wherein each set of multimodal features includes position information of one detection frame extracted from the to-be-detected image, a detection feature in the detection frame and first text information in the detection frame; and obtaining second text information matched with a to-be-extracted attribute based on the visual encoding feature, the to-be-extracted attribute and the plurality of sets of multimodal features, wherein the to-be-extracted attribute is an attribute of text information needing to be extracted.
ENGLISH CHAR IMAGE RECOGNITION METHOD
The invention provides an English char image recognition method mainly generating a rectangular coordinate frame from a loaded English character image, finding a gravity center in the rectangular coordinate frame, obtaining feature points on the rectangular coordinate frame and around the gravity center, performing one-dimensional convolutional operation and processing on the feature points to generate six layer feature maps, and generating a character recognition result from the six layer feature maps to solve the problem of generating a large amount of computation in conventional two-dimensional recognition operation, thereby achieving an efficacy of reducing recognition equipment costs and enabling fast and accurate recognition.
OBJECT RECOGNITION WITH REDUCED NEURAL NETWORK WEIGHT PRECISION
A client device configured with a neural network includes a processor, a memory, a user interface, a communications interface, a power supply and an input device, wherein the memory includes a trained neural network received from a server system that has trained and configured the neural network for the client device. A server system and a method of training a neural network are disclosed.
METHOD FOR RECOGNIZING TEXT, DEVICE, AND STORAGE MEDIUM
A method for recognizing text includes: obtaining a first feature map of an image; for each target feature unit, performing a feature enhancement process on a plurality of feature values of the target feature unit respectively based on the plurality of feature values of the target feature unit, in which the target feature unit is a feature unit in the first feature map along a feature enhancement direction; and performing a text recognition process on the image based on the first feature map after the feature enhancement process.
TRAINING METHOD FOR HANDWRITTEN TEXT IMAGE GENERATION MODE, ELECTRONIC DEVICE AND STORAGE MEDIUM
A training method for a handwritten text image generation model includes: obtaining training data including a sample content image, a first sample handwritten text image and a second sample handwritten text image, constructing an initial training model; obtaining a first predicted handwritten text image by inputting the sample content image and the second sample handwritten text image into an initial handwritten text image generation model of the initial training model; obtaining a second predicted handwritten text image by inputting the sample content image and the first sample handwritten text image into an initial handwritten text image reconstruction model of the initial training model; training the initial training model according to the first and second predicted handwritten text images and the first sample handwritten text image; and determining a handwritten text image generation model of the training model after training as a target handwritten text image generation model.
METHOD AND DEVICE FOR CONSTRUCTING LEGAL KNOWLEDGE GRAPH BASED ON JOINT ENTITY AND RELATION EXTRACTION
A method and device for constructing a legal knowledge graph based on joint entity and relation extraction. The construction method comprises the following steps: constructing a triple data set; design of a model architecture and training of a model, wherein the model architecture comprises an encoding layer, a head entity extraction layer and a relation-tail entity extraction layer; determination of the relation between the sentences of the text; triple combination and graph visualization. The design of the model framework of the present disclosure adopts a Chinese Bert pre-training model as an encoder. In the entity extraction part, two BiLSTM binary classifiers are used to identify the start position and end position of an entity. The head entity is first extracted, and then the tail entity corresponding to the entity relation is extracted from the extracted head entity.
Artificial aperture adjustment for synthetic depth of field rendering
This disclosure relates to various implementations that dynamically adjust one or more shallow depth of field (SDOF) parameters based on a designated, artificial aperture value. The implementations obtain a designated, artificial aperture value that modifies an initial aperture value for an image frame. The designated, artificial aperture value generates a determined amount of synthetically-produced blur within the image frame. The implementations determine an aperture adjustment factor based on the designated, artificial aperture value in relation to a default so-called “tuning aperture value” (for which the camera's operations may have been optimized). The implementations may then modify, based on the aperture adjustment factor, one or more SDOF parameters for an SDOF operation, which may, e.g., be configured to render a determined amount of synthetic bokeh within the image frame. In response the modified SDOF parameters, the implementations may render an updated image frame that corresponds to the designated, artificial aperture value.