Patent classifications
G06T2207/20081
BLOOD FLOW FIELD ESTIMATION APPARATUS, LEARNING APPARATUS, BLOOD FLOW FIELD ESTIMATION METHOD, AND PROGRAM
A blood flow field estimation apparatus is provided, including an estimation unit that uses a learned model obtained in advance by performing machine learning to learn a relationship between organ tissue three-dimensional structure data including image data of a plurality of organ cross-sectional images serving as cross-sectional images of an organ and having each pixel provided with two or more bit depths and image position information serving as information indicating a position of an image reflected on each of the organ cross-sectional images in the organ, and a blood flow field in the organ, and estimates the blood flow field in the organ of an estimation target, based on the organ tissue three-dimensional structure data of the organ of the estimation target, and an output unit that outputs an estimation result of the estimation unit.
IMAGE PROCESSING METHOD, NETWORK TRAINING METHOD, AND RELATED DEVICE
This application provides an image processing method, a network training method, and a related device, and relates to image processing technologies in the artificial intelligence field. The method includes: inputting a first image including a first vehicle into an image processing network to obtain a first result output by the image processing network, where the first result includes location information of a two-dimensional 2D bounding frame of the first vehicle, coordinates of a wheel of the first vehicle, and a first angle of the first vehicle, and the first angle of the first vehicle indicates an included angle between a side line of the first vehicle and a first axis of the first image; and generating location information of a three-dimensional 3D outer bounding box of the first vehicle based on the first result.
DEEP LEARNING-BASED VIDEO EDITING METHOD, RELATED DEVICE, AND STORAGE MEDIUM
A deep learning-based video editing method can allow for automated editing of a video, reducing or eliminating user input, saving time and labor investments, and thereby improving video editing efficiency. Attribute recognition is performed on an object in a target video using a deep learning model. A target object is selected that satisfies an editing requirement of the target video. A plurality of groups of pictures associated with the target object from the target video are obtained using editing. An edited video corresponding to the target video is generated using the plurality of groups of pictures.
IMAGE REGISTRATION METHOD AND ELECTRONIC DEVICE
An image registration method includes: acquiring a target image comprising a target object; inputting the target image to a preset network model, and outputting position information and rotation angle information of the target object; obtaining a reference image comprising the target object by querying a preset image database according to the position information and the rotation angle information; and performing image registration on the target image and the reference image to obtain a corresponding position of the target object of the target image in the reference image.
METHOD FOR TRAINING IMAGE PROCESSING MODEL
This disclosure relates to a model training method and apparatus and an image processing method and apparatus. The model training method includes: obtaining a first sample image and a first standard region proportion corresponding to a first object in the first sample image; obtaining a standard region segmentation result corresponding to the first sample image based on the first standard region proportion; and training a first initial segmentation model based on the first sample image and the standard region segmentation result, to obtain a first target segmentation model.
METHOD OF FUSING IMAGE, AND METHOD OF TRAINING IMAGE FUSION MODEL
A method of fusing an image, a method of training an image fusion model, an electronic device, and a storage medium. The method of fusing the image includes: encoding a stitched image obtained by stitching a foreground image and a background image, so as to obtain a feature map; and decoding the feature map to obtain a fused image, wherein the feature map is decoded by: performing a weighting on the feature map by using an attention mechanism, so as to obtain a weighted feature map; performing a fusion on the feature map according to feature statistical data of the weighted feature map, so as to obtain a fused feature; and decoding the fused feature to obtain the fused image.
METHOD FOR ANALYZING HUMAN TISSUE ON BASIS OF MEDICAL IMAGE AND DEVICE THEREOF
Disclosed are a method and device for analyzing human tissue on the basis of a medical image. A tissue analysis device generates training data including a two-dimensional medical image and volume information of tissue by using a three-dimensional medical image, and trains, by using the training data, an artificial intelligence model that obtains a three-dimensional size, volume, or weight of tissue by dividing at least one or more normal or diseased tissues from a two-dimensional medical image in which a plurality of tissues are displayed overlapping on the same plane. In addition, the tissue analysis device obtains a three-dimensional size, volume, or weight of normal or diseased tissue from an X-ray medical image by using the artificial intelligence model.
MAP INFORMATION UPDATE METHOD, LANDMARK GENERATION METHOD, AND FEATURE POINT DISTRIBUTION ADJUSTMENT METHOD
A map information update method includes: (a) obtaining map information; (b) obtaining landmark observed positions indicating positions of one or more landmarks in a captured image; (c) adding that includes (i) generating added map information by adding information pertaining to the landmark observed positions to the map information, and (ii) updating the map information obtained in (a) to the added map information; (d) predicting that includes (i) calculating predicted map information based on the map information updated in (c), by using a neural network inference engine that has been trained, and (ii) updating the map information to the predicted map information; and updating information that includes (i) calculating updated map information based on the map information updated in (d), by using a gradient method, and (ii) updating the map information to the updated map information.
METHODS AND SYSTEMS FOR GENERATING END-TO-END DE-SMOKING MODEL
The disclosure herein relates to methods and systems for generating an end-to-end de-smoking model for removing smoke present in a video. Conventional data-driven based de-smoking approaches are limited mainly due to lack of suitable training data. Further, the conventional data-driven based de-smoking approaches are not end-to-end for removing the smoke present in the video. The de-smoking model of the present disclosure is trained end-to-end with the use of synthesized smoky video frames that are obtained by source aware smoke synthesis approach. The end-to-end de-smoking model localize and remove the smoke present in the video, using dynamic properties of the smoke. Hence the end-to-end de-smoking model simultaneously identifies the regions affected with the smoke and performs the de-smoking with minimal artifacts. localized smoke removal and color restoration of a real-time video.
TECHNIQUES FOR QUANTITATIVELY ASSESSING TEAR-FILM DYNAMICS
Aspects of the present disclosure provide techniques for quantitatively assessing tear-film dynamics associated with contact lenses. An example method includes projecting an image of one or more shapes on a tear film surface of the contact lens worn on the eye, capturing video data, comprising a plurality of image frames, of the one or more shapes projected on the tear film surface of the contact lens over a period of time, performing image segmentation on a plurality of reflection patterns included in the plurality of image frames, generating a plurality of maps of the tear film surface of the contact lens indicating changes to the tear film surface of the contact lens during the period of time, and outputting, based on the plurality of maps, one or more metrics quantifying the changes to the tear film surface of the contact lens over the period of time.