G06V10/806

Apparatus and method for image classification and segmentation based on feature-guided network, device, and medium
11763542 · 2023-09-19 · ·

The present invention provides an apparatus and method for image classification and segmentation based on a feature-guided network, a device, and a medium, and belongs to the technical field of deep learning. A feature-guided classification network and feature-guided segmentation network of the present invention include basic unit blocks. A local feature is enhanced and a global feature is extracted among the basic unit blocks. This resolves a problem that features are not fully utilized in existing image classification and image segmentation network models. In this way, a trained feature-guided classification network and feature-guided segmentation network have better effects and are more robust. The present invention selects the feature-guided classification network or the feature-guided segmentation network based on a requirement of an input image and outputs a corresponding category or segmented image, to resolve a problem that the existing classification or segmentation network model has an unsatisfactory classification or segmentation effect.

Synthesizing apparatus, synthesizing method and program
11188733 · 2021-11-30 · ·

A synthesizing apparatus comprises: an input part that inputs a plurality of feature point sets that are respectively extracted by a plurality of methods from an input image having a curved stripes pattern formed by ridges; and a synthesizing part that synthesizes the plurality of feature point sets by executing a logical operation on the plurality of feature point sets. The synthesizing part can execute a logical OR operation on the plurality of feature point sets. The synthesizing part can also execute a logical AND operation on the plurality of feature point sets.

IMAGE CLASSIFICATION METHOD, ELECTRONIC DEVICE AND STORAGE MEDIUM

Provided are an image classification method and apparatus, an electronic device and a storage medium, relating to the field of artificial intelligence and, in particular, to computer vision and deep learning. The method includes inputting a to-be-classified document image into a pretrained neural network and obtaining a feature submap of each text box of the to-be-classified document image by use of the neural network; inputting the feature submap of each text box, a semantic feature corresponding to preobtained text information of each text box and a position feature corresponding to preobtained position information of each text box into a pretrained multimodal feature fusion model and fusing, by use of the multimodal feature fusion model, the three into a multimodal feature corresponding to each text box; and classifying the to-be-classified document image based on the multimodal feature corresponding to each text box.

OBJECT FUNCTIONALITY PREDICATION METHODS, COMPUTER DEVICE, AND STORAGE MEDIUM
20210365718 · 2021-11-25 ·

A method is disclosed. The method includes obtaining an object for prediction and a plurality of candidate scenes by a computer device; inputting the object for prediction and a current candidate scene to a distance measurement model, the distance measurement model calculates a feature vector corresponding to the current candidate scene based on a trained scene feature subnetwork, and outputs a distance from the object for prediction to the current candidate scene based on the object for prediction and the feature vector corresponding to the current candidate scene, model parameters of the distance measurement model including a parameter determined by a trained object feature subnetwork; obtaining distances from the object for prediction to the plurality of candidate scenes based on the distance measurement model; determining a target scene corresponding to the object for prediction based on the distances from the object for prediction to the plurality of candidate scenes.

METHOD AND APPARATUS FOR OBJECT DETECTION IN IMAGE, VEHICLE, AND ROBOT
20210365726 · 2021-11-25 ·

This application discloses a method and apparatus for object detection in an image, a vehicle, and a robot. The method for object detection in an image is performed by a computing device. The method includes determining an image feature of an image; determining a correlation of pixels in the image based on the image feature; updating the image feature of the image based on the correlation to obtain an updated image feature; and determining an object detection result in the image according to the updated image feature.

IMAGE CLASSIFICATION METHOD, COMPUTER-READABLE STORAGE MEDIUM, AND COMPUTER DEVICE
20210365741 · 2021-11-25 ·

A computer device obtains a plurality of medical images. The device generates a texture image based on image data of a region of interest in the medical images. The device extracts a local feature from the texture image using a first network model. The device extracts a global feature from the medical images using a second network model. The device fuses the extracted local feature and the extracted global feature to form a fused feature. The device performs image classification based on the fused feature.

Face identification method and terminal device using the same

The present disclosure provides a face identification method and a terminal device using the same. The method includes: obtaining a to-be-detected image; performing a brightness enhancement process on the to-be-detected image based on a preset second calculation method to generate a to-be-identified face image; obtaining a first channel value of each channel corresponding to each pixel in the to-be-identified face image; performing another brightness enhancement process on the to-be-identified face image based on each first channel value and a preset first calculation method to obtain a target to-be-identified face image; and performing a face identification process on the target to-be-identified face image to obtain an identification result. Through the above-mentioned scheme, an enhanced face identification manner for the images of low brightness is provided.

Gesture Recognition Using Multiple Antenna
20220019291 · 2022-01-20 · ·

Various embodiments wirelessly detect micro gestures using multiple antenna of a gesture sensor device. At times, the gesture sensor device transmits multiple outgoing radio frequency (RF) signals, each outgoing RF signal transmitted via a respective antenna of the gesture sensor device. The outgoing RF signals are configured to help capture information that can be used to identify micro-gestures performed by a hand. The gesture sensor device captures incoming RF signals generated by the outgoing RF signals reflecting off of the hand, and then analyzes the incoming RF signals to identify the micro-gesture.

SMART NODE FOR AUTONOMOUS VEHICLE PERCEPTION AUGMENTATION
20220019225 · 2022-01-20 ·

A system for navigation of a vehicle that includes a vehicle computer vision system to receive digital images of an environment along a path. The vision system has a vision range and includes a processor and programming instructions. The processor detects in the digital images a first set of objects of interest (OOIs) and determines motion of each OOI in the first set of OOIs. The system includes a communication device that receives augmented perception data associated with a node along a portion of the path. The received perception data identifies motion of each OOI of a second set of OOIs detected within a vision range of the node. The system includes a navigation controller that uses a fusion of the first set of OOIs and the second set of OOIs to control motion of the vehicle along the path.

Method and apparatus for pose planar constraining on the basis of planar feature extraction

The present application provides a method and apparatus for pose planar constraining on the basis of planar feature extraction, wherein the method includes: inputting the acquired RGB color image and point cloud image into spatial transformation network to obtain two-dimensional and three-dimensional affine transformation matrixes; extracting the planar features of the transformed two-dimensional affine transformation matrix and three-dimensional affine transformation matrix; inputting the acquired planar features into the decoder and obtain the pixel classification of the planar features; clustering the vectors corresponding to the planar pixels to obtain the segmentation result of the planar sample; using planar fitted by the segmentation result to make planar constraint to the pose calculated by vision algorithm. The application combines RGB-D information to perform plane extraction, and designs a new spatial transformation network to transform two-dimensional color image and three-dimensional point cloud image.