G06V10/462

Edge data network for providing three-dimensional character image to user equipment and method for operating the same

An edge data network for providing a three-dimensional (3D) character image to a user equipment and an operating method thereof are provided. The edge data network obtains key points information including feature point coordinates related to the body parts of a first user, from a first user equipment via a network, and obtains view points information including virtual position coordinate value information of virtual view points from which a second user views a 3D character image from a second user equipment, measures a key points similarity and a view points similarity by respectively comparing the obtained key points information and view points information with key points information and view points information cached in a data cache, and reads out a 3D character image cached in the data cache based on the measured key points similarity and the measured view points similarity, and transmits the read out 3D character image to the second user equipment.

Image processing techniques to quickly find a desired object among other objects from a captured video scene
11694440 · 2023-07-04 · ·

Techniques are provided for identifying objects (such as products within a physical store) within a captured video scene and indicating which of object in the captured scene matches a desired object requested by a user. The matching object is then displayed in an accentuated manner to the user in real-time (via augmented reality). Object identification is carried out via a multimodal methodology. Objects within the captured video scene are identified using a neural network trained to identify different types of objects. The identified objects can then be compared against a database of pre-stored images of the desired product to determine if a close match is found. Additionally, text on the identified objects is analyzed and compared to the text of the desired object. Based on either or both identification methods, the desired object is indicated to the user on their display, via an augmented reality graphic.

HUMAN POSTURE DETERMINATION METHOD AND MOBILE MACHINE USING THE SAME
20230004739 · 2023-01-05 ·

Human posture determination is disclosed. Human posture is determined by obtaining range image(s) through a range camera, detecting key points of an estimated skeleton of a human in color data of the range image(s) and calculating positions of the detected key points based on depth data of the range image(s), choosing a feature map from a set of predefined feature maps based on the detected key points among a set of predefined key points, obtaining two features of a body of the human corresponding to the chosen feature map based on the positions of the detected key points, and determining a posture of the human according to the two features in the chosen feature map.

Automated gauge reading and related systems, methods, and devices

Computing devices and methods for reading gauges are disclosed. A gauge reading method includes capturing image data corresponding to a captured image of one or more gauges, detecting one or more gauges in the captured image, cropping a detected gauge in the captured image to provide a use image including the detected gauge, and classifying the detected gauge to correlate the detected gauge with a template image. The gauge reading method also includes attempting to perform feature detection rectification on the use image to produce a rectified image of the detected gauge, performing template matching rectification on the use image to produce the rectified image responsive to a failure to perform the feature detection rectification, and estimating a gauge reading responsive to the rectified image. A computing device may implement at least a portion of a gauge reading method.

Medical Image Registration Method Based on Progressive Images

A two-stage medical image registration method based on progressive images (PIs) to solve the technical problem of low registration accuracy of traditional image registration methods includes: merging a reference image with a floating image to generate multiple intermediate PIs; registering, by a speeded-up robust features (SURF) algorithm and an affine transformation, the floating image with the intermediate PIs to acquire coarse registration results; registering, by the SURF algorithm and the affine transformation, the reference image with the coarse registration results to acquire fine registration results; and comparing the fine registration results of the intermediate PIs, which are acquired by iteration, and selecting an optimal registration result as a final registration image. The method can achieve multimodal registration for brain imaging with MI, NCC, MSD, and NMI superior to those of the existing registration algorithms. The method effectively improves the registration accuracy through the progressive medical image registration strategy.

IMAGE PROCESSING SYSTEM

The present invention discloses a system and method for image processing and recognizing a scene of an image. The system utilizes a Multi-mode scalable network system and regrouping pipeline. The system is AI based system which uses neuro network. The system includes a pre-processing, processing and a post-processing unit. The system uses optical information recorded from the camera of a mobile device to extract and analyze the content in an image such as a photo or video clip. Based on the retrieved information, a label is given to best describe the scene of the image.

OBJECT IDENTIFICATION
20220413507 · 2022-12-29 ·

Object identification may be provided herein. A feature extractor may extract a first set of visual features, extract a second set of visual features, concatenate the first set of visual features, the second set of visual features, and a set of bounding box information, determine a number of object features and a global feature for a scene, and receive ego-vehicle feature information associated with an ego-vehicle. An object classifier may receive the number of object features, the global feature, and the ego-vehicle feature information, generate relational features with respect to relationships between each of the number of objects from the scene, and classify each of the number of objects from the scene based on the number of object features, the relational features, the global feature, the ego-vehicle feature information, and an intention of the ego-vehicle.

VIDEO ACTION RECOGNITION AND MODIFICATION

A system, method, and computer program product for implementing video action recognition is provided. The method includes receiving a video stream comprising user movement actions. Skeleton points associated with a video representation of a user executing the user movement actions are extracted and categorized with respect to multiple digital levels. Initial visual windows points are generated within video frames and an average movement distance for the group of skeleton points are determined with respect to the video frames. In response, sizes for the visual windows are adjusted and feature vectors are extracted from the group of skeleton points. Point coordinates of the skeleton points are extracted and linked with the feature vectors. A convolutional neural network associated with linking the feature vectors with the point coordinates is generated and the video stream is enabled with respect to video action recognition associated with accurate presentation of the video stream.

Fully convolutional interest point detection and description via homographic adaptation

Systems, devices, and methods for training a neural network and performing image interest point detection and description using the neural network. The neural network may include an interest point detector subnetwork and a descriptor subnetwork. An optical device may include at least one camera for capturing a first image and a second image. A first set of interest points and a first descriptor may be calculated using the neural network based on the first image, and a second set of interest points and a second descriptor may be calculated using the neural network based on the second image. A homography between the first image and the second image may be determined based on the first and second sets of interest points and the first and second descriptors. The optical device may adjust virtual image light being projected onto an eyepiece based on the homography.

Tools and methods for placing a medical appliance on a user

A method for operating an accessory device to guide the placement of an ostomy appliance on a user having a stoma. The method can include capturing an image or a sequence of images of the user applying the ostomy appliance to the user's body, processing the captured image or sequence of images, including: identifying a location of the stoma in one or more of the captured image or sequence of images, identifying a location of the ostomy appliance in one or more of the captured image or sequence of images, and generating location indicia representative of the location of the ostomy appliance with respect to the stoma in one or more of the captured image or sequence of images. A visual display including the location indicia associated with one or more of the captured image or sequence of images can be provided.