G06V10/46

CONTOUR SHAPE RECOGNITION METHOD
20230047131 · 2023-02-16 ·

Provided is a contour shape recognition method, including: sampling and extracting salient feature points of a contour of a shape sample; calculating a feature function of the shape sample at a semi-global scale by using three types of shape descriptors; dividing the scale with a single pixel as a spacing to acquire a shape feature function in a full-scale space; storing feature function values at various scales into a matrix to acquire three types of feature grayscale map representations of the shape sample in the full-scale space; synthesizing the three types of grayscale map representations of the shape sample, as three channels of RGB, into a color feature representation image; constructing a two-stream convolutional neural network by taking the shape sample and the feature representation image as inputs at the same time; and training the two-stream convolutional neural network, and inputting a test sample into a trained network model to achieve shape classification.

IMAGE GAZE CORRECTION METHOD, APPARATUS, ELECTRONIC DEVICE, COMPUTER-READABLE STORAGE MEDIUM, AND COMPUTER PROGRAM PRODUCT

An image gaze correction method, apparatus, electronic device, computer-readable storage medium, and computer program product related to the field of artificial intelligence technologies are provided. The image gaze correction method includes: acquiring an eye image from an image; performing feature extraction processing on the eye image to obtain feature information of the eye image; performing, based on the feature information and a target gaze direction, gaze correction processing on the eye image to obtain an initially corrected eye image and an eye contour mask; performing, by using the eye contour mask, adjustment processing on the initially corrected eye image to obtain a corrected eye image; and generating a gaze corrected image based on the corrected eye image.

IMAGE GAZE CORRECTION METHOD, APPARATUS, ELECTRONIC DEVICE, COMPUTER-READABLE STORAGE MEDIUM, AND COMPUTER PROGRAM PRODUCT

An image gaze correction method, apparatus, electronic device, computer-readable storage medium, and computer program product related to the field of artificial intelligence technologies are provided. The image gaze correction method includes: acquiring an eye image from an image; performing feature extraction processing on the eye image to obtain feature information of the eye image; performing, based on the feature information and a target gaze direction, gaze correction processing on the eye image to obtain an initially corrected eye image and an eye contour mask; performing, by using the eye contour mask, adjustment processing on the initially corrected eye image to obtain a corrected eye image; and generating a gaze corrected image based on the corrected eye image.

APPARATUS AND METHOD FOR IDENTIFYING CONDITION OF ANIMAL OBJECT BASED ON IMAGE
20230049090 · 2023-02-16 ·

An image-based animal object condition identification apparatus includes: a communication module that receives an image of an object; a memory that stores therein a program configured to extract animal condition information from the received image; and a processor that executes the program. The program extracts continuous animal detection information of each object by inputting the received image into an animal detection model that is trained based on learning data composed of animal images and determines predetermined animal condition information for each class of each animal object by inputting the continuous animal detection information of each object into an animal condition identification model.

APPARATUS AND METHOD FOR IDENTIFYING CONDITION OF ANIMAL OBJECT BASED ON IMAGE
20230049090 · 2023-02-16 ·

An image-based animal object condition identification apparatus includes: a communication module that receives an image of an object; a memory that stores therein a program configured to extract animal condition information from the received image; and a processor that executes the program. The program extracts continuous animal detection information of each object by inputting the received image into an animal detection model that is trained based on learning data composed of animal images and determines predetermined animal condition information for each class of each animal object by inputting the continuous animal detection information of each object into an animal condition identification model.

CLASSIFICATION AND SORTING WITH SINGLE-BOARD COMPUTERS

A material handling system sorts materials utilizing a vision system of multiple vision devices configured with single board computers that each implement an artificial intelligence system in order to identify or classify materials, which are then sorted into separate groups based on such an identification or classification by sorting devices that are each coupled to one of the vision devices.

CLASSIFICATION AND SORTING WITH SINGLE-BOARD COMPUTERS

A material handling system sorts materials utilizing a vision system of multiple vision devices configured with single board computers that each implement an artificial intelligence system in order to identify or classify materials, which are then sorted into separate groups based on such an identification or classification by sorting devices that are each coupled to one of the vision devices.

Detecting interactions with non-discretized items and associating interactions with actors using digital images

Commercial interactions with non-discretized items such as liquids in carafes or other dispensers are detected and associated with actors using images captured by one or more digital cameras including the carafes or dispensers within their fields of view. The images are processed to detect body parts of actors and other aspects therein, and to not only determine that a commercial interaction has occurred but also identify an actor that performed the commercial interaction. Based on information or data determined from such images, movements of body parts associated with raising, lowering or rotating one or more carafes or other dispensers may be detected, and a commercial interaction involving such carafes or dispensers may be detected and associated with a specific actor accordingly.

Global and local binary pattern image crack segmentation method based on robot vision

A global and local binary pattern image crack segmentation method based on robot vision comprises the following steps: enhancing a contrast of an acquired original image to obtain an enhanced map; using an improved local binary pattern detection algorithm to process the enhanced map and construct a saliency map; using the enhanced map and the saliency map to segment cracks and obtaining a global and local binary pattern automatic crack segmentation method; and evaluating performance of the obtained global and local binary pattern automatic crack segmentation method. The present application uses logarithmic transformation to enhance the contrast of a crack image, so that information of dark parts of the cracks is richer. Texture features of a rotation invariant local binary pattern are improved. Global information of four directions is integrated, and the law of universal gravitation and gray and roundness features are introduced to correct crack segmentation results, thereby improving segmentation accuracy. Crack regions can be segmented in the background of uneven illumination and complex textures. The method has good robustness and meets requirements of online detection.

Associating three-dimensional coordinates with two-dimensional feature points
11580662 · 2023-02-14 · ·

An example method includes causing a light projecting system of a distance sensor to project a three-dimensional pattern of light onto an object, wherein the three-dimensional pattern of light comprises a plurality of points of light that collectively forms the pattern, causing a light receiving system of the distance sensor to acquire an image of the three-dimensional pattern of light projected onto the object, causing the light receiving system to acquire a two-dimensional image of the object, detecting a feature point in the two-dimensional image of the object, identifying an interpolation area for the feature point, and computing three-dimensional coordinates for the feature point by interpolating using three-dimensional coordinates of two points of the plurality of points that are within the interpolation area.