Patent classifications
G06V20/647
MULTI-SENSOR OCCLUSION-AWARE TRACKING OF OBJECTS IN TRAFFIC MONITORING SYSTEMS AND METHODS
Systems and methods for tracking objects though a traffic control system include a plurality of sensors configured to capture data associated with a traffic location, and a logic device configured to detect one or more objects in the captured data, determine an object location within the captured data, transform each object location to world coordinates associated with one of the plurality of sensors; and track each object location using the world coordinates using prediction and occlusion-based processes. The plurality of sensors may include a visual image sensor, a thermal image sensor, a radar sensor, and/or another sensor. An object localization process includes a trained deep learning process configured to receive captured data from one of the sensors and determine a bounding box surrounding the detected object and output a classification of the detected object. The tracked objects are further transformed to three-dimensional objects in the world coordinates.
Method and device for vertebra localization and identification
A vertebra localization and identification method includes: receiving one or more images of vertebrae of a spine; applying a machine learning model on the one or more images to generate three-dimensional (3-D) vertebra activation maps of detected vertebra centers; performing a spine rectification process on the 3-D vertebra activation maps to convert each 3-D vertebra activation map into a corresponding one-dimensional (1-D) vertebra activation signal; performing an anatomically-constrained optimization process on each 1-D vertebra activation signal to localize and identify each vertebra center in the one or more images; and outputting the one or more images, wherein on each of the one or more outputted images, a location and an identification of each vertebra center are specified.
INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND PROGRAM
The present technology relates to an information processing apparatus, an information processing method, and a program capable of improving the efficiency and accuracy of clustering. The information processing apparatus estimates a presence area of an object and the number of objects on the basis of a captured image, generates point cloud data from distance measurement information acquired by a distance measuring sensor, and recognizes the object by determining a point cloud, which is a target of clustering, in the point cloud data generated and the number of clusters on the basis of the presence area of the object and the number of the objects, which are estimated, and performing clustering on the point cloud data. The present technology can be applied to an autonomous mobile robot system.
OBJECT RECOGNITION METHOD WITH INCREASED REPRESENTATIVENESS
A method for an object of interest in a degraded 2D digital image of the object is provided. The method includes the following steps: detecting, beforehand, the object of interest in a 2D digital image and assigning it a label; reconstructing a 3D volume of the object thus labeled from a plurality of available 2D digital images of the object of interest; storing, in a database, a record relating to the object thus reconstructed in 3D form and labeled; for each record thus stored, generating a new plurality of 2D digital images according to a plurality of viewing modes from the thus reconstructed 3D volume of each object; training a neural network on a learning set composed of an expanded set of 2D digital images thus generated and corresponding with the label of the object of interest to be recognized; from a degraded 2D digital image of the object of interest to be recognized; using the neural network thus trained to deliver as output the label of the object and a confidence index linked to the recognition of the object of interest.
BARCODE READERS WITH 3D CAMERA(S)
Methods and systems include using three-dimensional imaging apparatuses to capture three-dimensional images and analyze resulting three-dimensional image data to enhance captured two-dimensional images for scanning related processes such as object identification, symbology detection, object recognition model training, and identifying improper scan attempts or other actions performed by an operator. Imaging systems such as bi-optic readers, handheld scanners, and machine vision systems are described using three-dimensional imaging apparatuses and described capturing three-dimensional images and using with captured two-dimensional images.
Advanced driver assist system and method of detecting object in the same
ADAS includes a processing circuit and a memory which stores instructions executable by the processing circuit. The processing circuit executes the instructions to cause the ADAS to receive, from a vehicle that is in motion, a video sequence, generate a position image including at least one object included in the stereo image, generate a second position information associated with the at least one object based on reflected signals received from the vehicle, determine regions each including at least a portion of the at least one object as candidate bounding boxes based on the stereo image and the position image, and selectively adjusting class scores of respective ones of the candidate bounding boxes associated with the at least one object based on whether a respective first position information of the respective ones of the candidate bounding boxes matches the second position information.
Drivable surface identification techniques
The present disclosure relates generally to identification of drivable surfaces in connection with autonomously performing various tasks at industrial work sites and, more particularly, to techniques for distinguishing drivable surfaces from non-drivable surfaces based on sensor data. A framework for the identification of drivable surfaces is provided for an autonomous machine to facilitate it to autonomously detect the presence of a drivable surface and to estimate, based on sensor data, attributes of the drivable surface such as road condition, road curvature, degree of inclination or declination, and the like. In certain embodiments, at least one camera image is processed to extract a set features from which surfaces and objects in a physical environment are identified, and to generate additional images for further processing. The additional images are combined with a 3D representation, derived from LIDAR or radar data, to generate an output representation indicating a drivable surface.
RERENDERING A POSITION OF A HAND TO DECREASE A SIZE OF A HAND TO CREATE A REALISTIC VIRTUAL/AUGMENTED REALITY ENVIRONMENT
The technology disclosed relates to a method of realistic rendering of a real object as a virtual object in a virtual space using an offset in the position of the hand in a three-dimensional (3D) sensory space. An offset between expected positions of the eye(s) of a wearer of a head mounted device and a sensor attached to the head mounted device for sensing a position of at least one hand in a three-dimensional (3D) sensory space is determined. A position of the hand in the three-dimensional (3D) sensory space can be sensed using a sensor. The sensed position of the hand can be transformed by the offset into a re-rendered position of the hand as would appear to the wearer of the head mounted device if the wearer were looking at the actual hand. The re-rendered hand can be depicted to the wearer of the head mounted device.
Classification of synthetic data tasks and orchestration of resource allocation
Various techniques are described for classifying synthetic data tasks and orchestrating a resource allocation between groups of eligible resources for processing the synthetic data tasks. Received synthetic data tasks can be classified by identifying a task category and a corresponding group of eligible resources (e.g., processors) for processing synthetic data tasks in the task category. For example, synthetic data tasks can include generation of source assets, ingestion of source assets, identification of variation parameters, variation of variation parameters, and creation of synthetic data. Certain categories of synthetic data tasks can be classified for processing with a particular group of eligible resources. For example, tasks to ingest synthetic data assets can be classified for processing on a CPU only, while a task to create synthetic data assets can be classified for processing on a GPU only. The synthetic data tasks can be queued and routed for processing by an eligible resource.
IMAGE PROCESSING METHOD AND APPARATUS, ELECTRONIC DEVICE, AND COMPUTER-READABLE STORAGE MEDIUM
The present disclosure provides an image processing method and apparatus, an electronic device, and a computer-readable storage medium. The method includes: obtaining a first three-dimensional image of a target object in a three-dimensional coordinate system; determining a target plane of the target object in the first three-dimensional image, the target plane comprising target three-dimensional points; projecting the target three-dimensional points to a two-dimensional coordinate system defined on the target plane, to obtain target two-dimensional points; determining a target polygon and a minimum circumscribed target graphic of the target polygon according to the target two-dimensional points; and recognizing the minimum circumscribed target graphic as a first target graphic of the target object in the first target three-dimensional image.