G06F18/253

COMPUTER PROGRAM, INFORMATION PROCESSING METHOD, AND INFORMATION PROCESSING DEVICE

A non-transitory computer-readable medium (CRM) storing computer program code executed by a computer processor that executes a process of acquiring a medical image generated based on a signal detected by a catheter inserted to a lumen organ, estimating a position of an object at least included in the acquired medical image by inputting the medical image to a first learning model for estimating a position of an object included in the medical image, extracting from the medical image an image portion by using the estimated position of the object as a reference, and recognizing the object included in the extracted image portion by inputting the image portion to a second learning model for recognizing an object included in the image portion.

Communication efficient machine learning of data across multiple sites

In one embodiment, a service receives machine learning-based generative models from a plurality of distributed sites. Each generative model is trained locally at a site using unlabeled data observed at that site to generate synthetic unlabeled data that mimics the unlabeled data used to train the generative model. The service receives, from each of the distributed sites, a subset of labeled data observed at that site. The service uses the generative models to generate synthetic unlabeled data. The service trains a global machine learning-based model using the received subsets of labeled data received from the distributed sites and the synthetic unlabeled data generated by the generative models.

SYSTEM AND METHOD OF UNIQUE IDENTIFYING A GEMSTONE

There is provided a computerized system and method of generating a unique identification associated with a gemstone, usable for unique identification of the gemstone. The method comprises: obtaining one or more images of the gemstone, the one or more images captured at one or more viewing angles relative to the gemstone and to a light pattern, thus giving rise to a representative group of images; processing the representative group of images to generate a set of rotation-invariant values informative of rotational cross-correlation relationship characterizing the images in the representative group; and using the generated set of rotation-invariant values to generate a unique identification associated with the gemstone. The unique identification associated with the gemstone can be further compared with an independently generated unique identification associated with the gemstone in question, or with a class-indicative unique identification.

Image feature combination for image-based object recognition
11551329 · 2023-01-10 · ·

Methods, systems, and articles of manufacture to improve image recognition searching are disclosed. In some embodiments, a first document image of a known object is used to generate one or more other document images of the same object by applying one or more techniques for synthetically generating images. The synthetically generated images correspond to different variations in conditions under which a potential query image might be captured. Extracted features from an initial image of a known object and features extracted from the one or more synthetically generated images are stored, along with their locations, as part of a common model of the known object. In other embodiments, image recognition search effectiveness is improved by transforming the location of features of multiple images of a same known object into a common coordinate system. This can enhance the accuracy of certain aspects of existing image search/recognition techniques including, for example, geometric verification.

Method and apparatus for generating a competition commentary based on artificial intelligence, and storage medium

There is provided a method and apparatus for generating a competition commentary based on artificial intelligence, and a storage medium. The method comprises: obtaining commentator's words commentaries and structured data of historical competitions; generating a commentating model according to obtained information; during live broadcast of a competition, determining a corresponding words commentary according to the commentating model with respect to the structured data obtained each time.

Method of multi-sensor data fusion
11552778 · 2023-01-10 · ·

A method of multi-sensor data fusion includes determining a plurality of first data sets using a plurality of sensors, each of the first data sets being associated with a respective one of a plurality of sensor coordinate systems, and each of the sensor coordinate systems being defined in dependence of a respective one of a plurality of mounting positions for the sensors; transforming the first data sets into a plurality of second data sets using a transformation rule, each of the second data sets being associated with a unified coordinate system, the unified coordinate system being defined in dependence of at least one predetermined reference point; and determining at least one fused data set by fusing the second data sets.

Methods and systems for computer-based determining of presence of objects

A computer-implemented method for processing a 3-D point cloud data and an associated image data to enrich the 3-D point cloud data with relevant portions of the image date. The method comprises generating a 3-D point cloud data tensor representative of information contained in the 3-D point cloud data and generating an image tensor representative of information contained in the image data; and then analyzing the image tensor to identify a relevant data portion of the image information relevant to the at least one object candidate. The method further includes amalgamating the 3-D point cloud data tensor with a relevant portion of the image tensor associated with the relevant data portion of the image information to generate an amalgamated tensor associated with the surrounding area and storing the amalgamated tensor to be used by a machine learning algorithm (MLA) to determine presence of the object in the surrounding area.

Method of performing function of electronic device and electronic device using same

An electronic device includes: a camera; a microphone; a display; a memory; and a processor configured to receive an input for activating an intelligent agent service from a user while at least one application is executed, identify context information of the electronic device, control to acquire image information of the user through the camera, based on the identified context information, detect movement of a user's lips included in the acquired image information to recognize a speech of the user, and perform a function corresponding to the recognized speech.

METHOD, ELECTRONIC DEVICE, AND COMPUTER PROGRAM PRODUCT FOR TRAINING MODEL
20230214450 · 2023-07-06 ·

Embodiments of the present disclosure provide a method, an electronic device, and a computer program product for training a model. The method may include determining image features, audio features, and text features of a reference object based on reference image information, reference audio information, and reference text information associated with the reference object, respectively. The method may also include constructing a feature tensor from the image features, the audio features, and the text features. In addition, the method may further include decomposing the feature tensor into a first feature vector, a second feature vector, and a third feature vector corresponding to the image features, the audio features, and the text features, respectively, to determine a loss function value of the model. The method may also include updating parameters of the model based on the loss function value.

SYSTEMS AND METHODS OF COOPERATIVE DEPTH COMPLETION WITH SENSOR DATA SHARING

Systems and methods are provided for utilizing sensor data from sensors of different modalities and from different vehicles to generate a combined image of an environment. Sensor data, such as a point cloud, generated by a LiDAR sensor on a first vehicle may be combined with sensor data, such as image data, generated by a camera on a second vehicle. The point cloud and image data may be combined to provide benefits over either data individually and processed to provide an improved image of the environment of the first and second vehicles. Either vehicle can perform this processing when receiving the sensor data from the other vehicle. An external system can also do the processing when receiving the sensor data from both vehicles. The improved image can then be used by one or both of the vehicles to improve, for example, automated travel through or obstacle identification in the environment.