G06V30/19007

AUTOMATIC LABELING OF OBJECTS IN SENSOR DATA

Aspects of the disclosure provide for automatically generating labels for sensor data. For instance, first sensor data, for a vehicle may be identified. This first sensor data may have been captured by a first sensor of the vehicle at a first location during a first point in time and may be associated with a first label for an object. Second sensor data for the vehicle may be identified. The second sensor data may have been captured by a second sensor of the vehicle at a second location at a second point in time outside of the first point in time. The second location is different from the first location. A determination may be made as to whether the object is a static object. Based on the determination that the object is a static object, the first label may be used to automatically generate a second label for the second sensor data.

WINE LABEL RECOGNITION METHOD, WINE INFORMATION MANAGEMENT METHOD AND APPARATUS, DEVICE, AND STORAGE MEDIUM
20230237824 · 2023-07-27 ·

A wine label recognition method, a wine information management method and apparatus, a computer device, and a computer-readable storage medium are provided. The method includes: obtaining a wine image, and performing optical character recognition (OCR) on the wine image in a preset OCR manner, to obtain text included in the wine image (S21); performing deep learning recognition on the wine image in a preset deep learning recognition manner, to obtain an image feature included in the wine image (S22); and sifting out a target wine label matching the text and the image feature from a preset wine label database according to the text and the image feature, and using the target wine label as a wine label corresponding to the wine image (S33). Advantages of deep learning and OCR are fully utilized thereby improving accuracy and efficiency of wine label recognition and improving automation efficiency of wine information management.

SYSTEMS AND METHODS FOR IMMEDIATE IMAGE QUALITY FEEDBACK

An apparatus (1) for providing image quality feedback during a medical imaging examination includes at least one electronic processor (20) programmed to: receive a live video feed (17) of a display (6) of an imaging device controller (4) of an imaging device (2) performing the medical imaging examination; extract a preview image (12) from the live video feed; perform an image analysis (38) on the extracted preview image to determine whether the extracted preview image satisfies an alert criterion; and output an alert (30) when the extracted preview image satisfies the alert criterion as determined by the image analysis.

IMAGE PATTERN MATCHING TO ROBOTIC PROCESS AUTOMATIONS
20230009201 · 2023-01-12 · ·

Disclosed herein is a computing system. The computing system includes a memory and a processor. The memory stores processor executable instructions for a workflow recommendation assistant engine. The processor is coupled to the memory. The processor executes the workflow recommendation assistant engine to cause the computing device to analyze images of a user interface corresponding to user activity, execute a pattern matching of the images with respect to existing automations, and provide a prompt indicating that an existing automation matches the user activity.

IMAGE FORMING APPARATUS THAT PERFORMS INSPECTION PROCESSING ON PRINT DATA AND METHOD OF CONTROLLING IMAGE FORMING APPARATUS
20230004334 · 2023-01-05 ·

An image forming apparatus capable of controlling execution of inspection without increasing a time period required to complete printing. On a registration screen, whether or not to execute inspection of data to be printed is set, and keywords indicative of confidentiality are registered. Text information is extracted from the data to be printed, and whether or not any keyword matching the text information has been registered is determined. Execution of print processing of the data to be printed is controlled based on a result of the determination. When non-execution of inspection is set, the print processing of the data to be printed is executed without executing the determination, whereas when execution of inspection is set, the print processing of the data to be printed is controlled based on a result of the determination.

TRAINING METHOD OF TEXT RECOGNITION MODEL, TEXT RECOGNITION METHOD, AND APPARATUS

The present disclosure provides a training method of a text recognition model, a text recognition method, and an apparatus, relating to the technical field of artificial intelligence, and specifically, to the technical field of deep learning and computer vision, which can be applied in scenarios such as optional character recognition, etc. The specific implementation solution is: performing mask prediction on visual features of an acquired sample image, to obtain a predicted visual feature; performing mask prediction on semantic features of acquired sample text, to obtain a predicted semantic feature, where the sample image includes text; determining a first loss value of the text of the sample image according to the predicted visual feature; determining a second loss value of the sample text according to the predicted semantic feature; training, according to the first loss value and the second loss value, to obtain the text recognition model.

Methods and apparatus to determine the dimensions of a region of interest of a target object from an image using target object landmarks
11538235 · 2022-12-27 · ·

Methods and apparatus to determine the dimensions of a region of interest of a target object and a class of the target object from an image using target object landmarks are disclosed herein. An example method includes identifying a landmark of a target object in an image based on a match between the landmark and a template landmark; classifying a target object based on the identified landmark; projecting dimensions of the template landmark based on a location of the landmark in the image; and determining a region of interest based on the projected dimensions, the region of interest corresponding to text printed on the target object.

METHOD FOR TRAINING TEXT POSITIONING MODEL AND METHOD FOR TEXT POSITIONING
20220392242 · 2022-12-08 ·

A method for training a text positioning model includes: obtaining a sample image, where the sample image contains a sample text to be positioned and a text marking box for the sample text; inputting the sample image into a text positioning model to be trained to position the sample text, and outputting a prediction text box for the sample image; obtaining a sample prior anchor box corresponding to the sample image; and adjusting model parameters of the text positioning model based on the sample prior anchor box, the text marking box and the prediction text box, and continuing training the adjusted text positioning model based on a next sample image until model training is completed, to generate a target text positioning model.

METHOD, SYSTEM, SERVER, AND STORAGE MEDIUM FOR LOGISTICS MANAGEMENT BASED ON QR CODE

A method of logistics management based on QR code, the method includes: acquiring a video stream of a logistics box taken by a camera of a logistics box passing through an identification area, the logistics box having a corresponding QR code affixed to its surface; performing target detection on the video stream based on a predetermined neural network model to obtain a QR code image and a logistics box image, respectively; identifying the QR code image to generate QR code information; detecting and tracking the logistics box images in sequence to generate logistics box information; matching the QR code information and logistics box information, generating matching information, and updating logistics inventory information based on the matching information.

Multi-Dimensional Table Reproduction From Image
20230057485 · 2023-02-23 ·

Embodiments facilitate selection and assignment of a known user model, based upon input comprising table images of original data. A table engine receives the image and performs pre-processing (e.g., rasterization, Optical Character Recognition, coordinate representation) thereupon to identify image entities. After filtering original numerical data, a similarity (e.g., a distance) is calculated between an image entity and a dimension member of the known user model. Based upon this similarity, the table engine selects and assigns the known user model to the incoming tables images, generating a file representing table columns and rows. This file is received at the UI of an analytics platform, which in turn populates the model with data of the user (rather than the original data) via an API. Embodiments may be particularly valuable in allowing a user to rapidly generate multi-dimensional tables comprising their own data, based upon raw table images received from an external party.