G06V10/7784

COMPUTER VISION TECHNOLOGIES FOR RAPID DETECTION

A computing system includes a processor; and a memory having stored thereon an adjustment application comprising computer-executable instructions that, when executed, cause the computing system to: display a graphical user interface including a digital medical image of a patient; superimpose a bounding box; receive an adjustment of an area of interest; and provide an adjusted digital medical image. A non-transitory computer-readable medium includes computer-executable instructions that, when executed via one or more processors, cause a computer to: display a graphical user interface including a digital medical image of a patient; superimpose a bounding box; receive an adjustment of an area of interest; and provide an adjusted digital medical image. A computer-implemented method includes: displaying a graphical user interface including a digital medical image of a patient; superimposing a bounding box; receiving an adjustment of an area of interest; and providing an adjusted digital medical image.

Adaptive cyber-physical system for efficient monitoring of unstructured environments

The present disclosure provides a system for monitoring unstructured environments. A predetermined path can be determined according to an assignment of geolocations to one or more agronomically anomalous target areas, where the one or more agronomically anomalous target areas are determined according to an analysis of a plurality of first images that automatically identifies a target area that deviates from a determination of an average of the plurality of first images that represents an anomalous place within a predetermined area, where the plurality of first images of the predetermined area are captured by a camera during a flight over the predetermined area. A camera of an unmanned vehicle can capture at least one second image of the one or more agronomically anomalous target areas as the unmanned vehicle travels along the predetermined path.

Machine learning system and method for determining or inferring user action and intent based on screen image analysis
11704898 · 2023-07-18 · ·

System(s) and method(s) that analyze image data associated with a computing screen operated by a user, and learns the image data (e.g., using pattern recognition, historical information analysis, user implicit and explicit training data, optical character recognition (OCR), video information, 360°/panoramic recordings, and so on) to concurrently glean information regarding multiple states of user interaction (e.g., analyzing data associated with multiple applications open on a desktop, mobile phone or tablet). A machine learning model is trained on analysis of graphical image data associated with screen display to determine or infer user intent. An input component receives image data regarding a screen display associated with user interaction with a computing device. An analysis component employs the model to determine or infer user intent based on the image data analysis; and an action component provisions services to the user as a function of the determined or inferred user intent. In an implementation, a gaming component gamifies interaction with the user in connection with explicitly training the model.

Display device and content recommendation method

This disclosure can provide a display device and a display method. The display device includes at least one camera configured to capture an environmental scenario image; a display configured to display a user interface; a controller in communicated with the display, configured to receive a command, input by a user, for obtaining a content recommendation resource associated with content currently displayed in the user interface; determine whether an application corresponding to the content currently displayed in the user interface is an application invoking the at least one camera, and if yes, display a first user interface, where the first user interface displays a first image captured by the at least one camera.

Auto labeler
11556744 · 2023-01-17 · ·

Aspects of the disclosure relate to training a labeling model to automatically generate labels for objects detected in a vehicle's environment. In this regard, one or more computing devices may receive sensor data corresponding to a series of frames perceived by the vehicle, each frame being captured at a different time point during a trip of the vehicle. The computing devices may also receive bounding boxes generated by a first labeling model for objects detected in the series of frames. The computing devices may receive user inputs including an adjustment to at least one of the bounding boxes, the adjustment corrects a displacement of the at least one of the bounding boxes caused by a sensing inaccuracy. The computing devices may train a second labeling model using the sensor data, the bounding boxes, and the adjustment to increase accuracy of the second labeling model when automatically generating bounding boxes.

Automatic image selection for online product catalogs

Disclosed are systems, methods, and non-transitory computer-readable media for automatic image selection for online product catalogs. An image selection system gathers feature data for images of an item included in listings posted to an online marketplace. The image selection system uses the feature data as input in a machine learning model to determine probability scores indicating an estimated probability that each image is suitable to represent the item. The machine learning model is trained based on a set of training images of the item that have been labeled to indicate whether they are suitable to represent the image. The image selection system compares the probability scores and selects an image to represent the item as a stock image based on the comparison.

MODEL GENERATING APPARATUS AND METHOD

A model generating apparatus and method are provided. The apparatus receives a plurality of sample images. The apparatus generates a plurality of adversarial samples corresponding to the sample images. The apparatus inputs the sample images and the adversarial samples respectively to a first encoder and a second encoder in a self-supervised neural network to generate a plurality of first feature extractions and a plurality of second feature extractions. The apparatus calculates a similarity of each of the first feature extractions and the second feature extractions to train the self-supervised neural network. The apparatus generates a task model based on the first encoder and a plurality of labeled data.

METHOD FOR TRAINING FEATURE EXTRACTION MODEL, METHOD FOR CLASSIFYING IMAGE, AND RELATED APPARATUSES

The present disclosure provides a method for training a feature extraction model, a method for classifying an image and related apparatuses, and relates to the field of artificial intelligence technology such as deep learning and image recognition. The scheme comprises: extracting an image feature of each sample image in a sample image set using a basic feature extraction module of an initial feature extraction model, to obtain an initial feature vector set; performing normalization processing on each initial feature vector in the initial feature vector set using a normalization processing module of the initial feature extraction model, to obtain each normalized feature vector; and guiding training for the initial feature extraction model through a preset high discriminative loss function, to obtain a target feature extraction model as a training result.

DWELL TIME RECORDING OF DIGITAL IMAGE REVIEW SESSIONS
20230214452 · 2023-07-06 ·

Systems and methods describe dwell time recording of digital image review sessions. The system displays, at a user interface (UI), a portion of an image on at least one monitor, where the image is segmented into a multitude of patches. The system then receives UI events involving a change in the currently displayed patches. For each of the UI events, the system records one or more dwell times representing durations for which the current patches of the image were displayed. The system also receives a report associated with the image review session, and processes the text of the report to determine a classification label for the image. Finally, the system trains a machine learning model, using at least the recorded dwell times and the classification label for the image.

METHOD AND SYSTEM FOR CLASSIFYING FACES OF BOUNDARY REPRESENTATION (B-REP) MODELS USING ARTIFICIAL INTELLIGENCE

The invention relates to method and system for classifying faces of a Boundary Representation (B-Rep) model using Artificial Intelligence (AI). The method includes extracting topological information corresponding to each of a plurality of data points of a B-Rep model of a product; determining a set of parameters based on the topological information corresponding to each of the plurality of data points; transforming the set of parameters corresponding to each of the plurality of data points of the B-Rep model into a tabular format to obtain a parametric data table; and assigning each of the plurality of faces of the B-Rep model a category from a plurality of categories based on the parametric data table using an AI model