G06V10/809

METHOD FOR NON-CONTACT TRIGGERING OF BUTTONS

Disclosed are techniques for elevator control. In an aspect, a sensor senses time series data, wherein the time series data includes at least one image, and range of the image covers a plurality of buttons. A system module configured to determine whether the image contains a target object; determine a tip coordinate of a tip of the target object when the image contains the target object, wherein the tip refers to a point of the target object with the closest distance to the operation panel; and determine button information corresponding to the tip coordinate among a plurality of button information, and transmit a control signal at least according to the button information, wherein the plurality of button information is associated with the plurality of buttons. A controller receives the control signal and perform control operation according to the control signal.

IMAGE PROCESSING APPARATUS AND IMAGE PROCESSING METHOD

An image processing apparatus comprises a first tracking circuit configured to apply machine learning (ML)-based first tracking processing to images and a second tracking circuit configured to apply non-ML-based second tracking processing to the images. The apparatus controls operations of the first and second tracking circuits so that an operation frequency of the first tracking circuit to be lower than an operation frequency of the second tracking circuit.

MULTI-MODAL IMAGE CLASSIFICATION SYSTEM AND METHOD USING ATTENTION-BASED MULTI-INTERACTION NETWORK

The present disclosure belongs to the technical field of image processing, and provides a multi-modal image classification system and method using an attention-based multi-interaction network. The present disclosure utilizes a U-net network structure to fuse low-level visual features and high-level semantic features. An attention network is introduced to solve the problem of weak feature discrimination, and high attention is given to discriminative features, so that the attention network plays an important role in the final classification process. A sufficient multi-modal interaction mechanism is introduced, so that more effective correlation information and discriminative information are obtained among a plurality of modalities, and sufficient interaction among the plurality of modalities is completed, thereby solving the problems of weak feature discrimination and insufficient interaction among modalities in a multi-modal image classification task.

Utilizing machine learning models and captured video of a vehicle to determine a valuation for the vehicle

A valuation platform may receive, from a user device, video data associated with a vehicle, and may receive a vehicle history report of the vehicle based on a vehicle identification number of the vehicle. The valuation platform may receive, from the user device, feature data associated with the vehicle, and may process the video data, the vehicle history report, and the feature data, with a machine learning model, to determine one or more values for the vehicle. The valuation platform may determine a valuation for the vehicle based on the determined one or more values for the vehicle. The valuation platform may create a vehicle profile for the vehicle based on the video data, the vehicle history report, the feature data, the determined one or more values for the vehicle, and the valuation for the vehicle, and may perform one or more actions based on the vehicle profile.

SYSTEMS, ROBOTS, AND METHODS FOR SELECTING CLASSIFIERS BASED ON CONTEXT
20230111067 · 2023-04-13 ·

The present disclosure describes systems, robots, and methods for organizing and selecting classifiers of a library of classifiers. The classifiers of the library of classifiers can be organized in a relational model, such as a hierarchy or probability model. Instead of storing, activating, or executing the entire library of classifiers at once by a robot system, computational resource demand is reduced by executing subset of classifiers to determine context, and the determined context is used as a basis for selection of another subset of classifiers. This process can be repeated, to iteratively refine context and select more specific subsets of classifiers. A selected subset of classifiers can eventually be specific to a task to be performed by the robot system, such that the robot system can take action based on output from executing such specific classifiers.

SYSTEMS, ROBOTS, AND METHODS FOR SELECTING CLASSIFIERS BASED ON CONTEXT
20230111284 · 2023-04-13 ·

The present disclosure describes systems, robots, and methods for organizing and selecting classifiers of a library of classifiers. The classifiers of the library of classifiers can be organized in a relational model, such as a hierarchy or probability model. Instead of storing, activating, or executing the entire library of classifiers at once by a robot system, computational resource demand is reduced by executing subset of classifiers to determine context, and the determined context is used as a basis for selection of another subset of classifiers. This process can be repeated, to iteratively refine context and select more specific subsets of classifiers. A selected subset of classifiers can eventually be specific to a task to be performed by the robot system, such that the robot system can take action based on output from executing such specific classifiers.

Image Analysis System for Testing in Manufacturing

A vision analytics and validation (VAV) system for providing an improved inspection of robotic assembly, the VAV system comprising a trained neural network three-way classifier, to classify each component as good, bad, or do not know, and an operator station configured to enable an operator to review an output of the trained neural network, and to determine whether a board including one or more “bad” or a “do not know” classified components passes review and is classified as good, or fails review and is classified as bad. In one embodiment, a retraining trigger to utilize the output of the operator station to train the trained neural network, based on the determination received from the operator station.

Cascaded Neural Network-Based Attention Detection Method, Computer Device, And Computer-Readable Storage Medium
20220277558 · 2022-09-01 ·

The present invention provides an attention detection method based on a cascade neural network, a computer apparatus, and a computer-readable storage medium. The method includes: obtaining video data, recognizing a plurality of image frames, and extracting a face region of the plurality of image frames; recognizing the face region by using a first convolutional neural network to judge whether a first situation of inattention occurs; and recognizing, if it is confirmed that no first situation of inattention occurs, the face region by using a second convolutional neural network to judge whether a second situation of inattention occurs, where computational complexity of the first convolutional neural network is less than computational complexity of the second convolutional neural network. The present invention further provides the computer apparatus for implementing the foregoing method and the computer-readable storage medium.

Gesture recognition method, apparatus, and device

This application provides a gesture recognition method, and relates to the field of man-machine interaction technologies. The method includes: extracting M images from a first video segment in a video stream; performing gesture recognition on the M images by using a deep learning algorithm, to obtain a gesture recognition result corresponding to the first video segment; and performing result combination on gesture recognition results of N consecutive video segments including the first video segment, to obtain a combined gesture recognition result. In the foregoing recognition process, a gesture in the video stream does not need to be segmented or tracked, but phase actions are recognized by using a deep learning algorithm with a relatively fast calculation speed, and then the phase actions are combined, so as to improve a gesture recognition speed, and reduce a gesture recognition delay.

OBJECT DETECTION DEVICE, LEARNING METHOD, AND RECORDING MEDIUM

In an object detection device, a plurality of object detection units output a score indicating probability that a predetermined object exists, for each partial region set to image data inputted. The weight computation unit computes weights for merging the scores outputted by the plurality of object detection units, using weight calculation parameters, based on the image data. The merging unit merges the scores outputted by the plurality of object detection units, for each partial region, with the weights computed by the weight computation unit. The target model object detection unit configured to output a score indicating probability that the predetermined object exists, for each partial region set to the image data. The first loss computation unit computes a first loss indicating a difference of the score of the target model object detection unit from a ground truth label of the image data and the score merged by the merging unit. The first parameter correction unit corrects parameters of the target model object detection unit to reduce the first loss.