G06T7/521

Adaptive sampling for structured light scanning

A system to process images includes a light source configured to emit a first illumination pattern onto one or more first portions of a scene. The system also includes an image sensor configured to capture light reflected from the scene in response to the emitted first illumination pattern. The system also includes an optimizer configured to perform raytracing of the light reflected from the scene. The system further includes a processor operatively coupled to the optimizer. The processor is configured to determine a parameter of a surface of the scene based on the raytracing, cause the light source to emit a second illumination pattern onto one or more second portions of the scene based at least in part on the parameter of the surface, and refine the parameter of the surface of the scene based on additional raytracing performed on reflected light from the second illumination pattern.

Systems and methods for manipulating virtual shapes in three-dimensional space

Examples disclosed herein may involve a computing system that is operable to (i) present, via a visual interface, a virtual shape associated with a three-dimensional (3D) coordinate system, (ii) present, via the visual interface, a visual indicator positioned in proximity to the virtual shape and indicating that a specified spatial parameter of the virtual shape will be modified along a specified dimension of the 3D coordinate system in response to a given type of user input associated with the visual indicator, (iii) while presenting the visual indicator, detect an instance of the given type of user input associated with the visual indicator, and (iv) after detecting the instance of the given type of user input, update the virtual shape that is presented via the visual interface by modifying the specified spatial parameter of the virtual shape along the specified dimension.

Systems and methods for manipulating virtual shapes in three-dimensional space

Examples disclosed herein may involve a computing system that is operable to (i) present, via a visual interface, a virtual shape associated with a three-dimensional (3D) coordinate system, (ii) present, via the visual interface, a visual indicator positioned in proximity to the virtual shape and indicating that a specified spatial parameter of the virtual shape will be modified along a specified dimension of the 3D coordinate system in response to a given type of user input associated with the visual indicator, (iii) while presenting the visual indicator, detect an instance of the given type of user input associated with the visual indicator, and (iv) after detecting the instance of the given type of user input, update the virtual shape that is presented via the visual interface by modifying the specified spatial parameter of the virtual shape along the specified dimension.

Alternating light distributions for active depth sensing

Aspects of the present disclosure relate to systems and methods for active depth sensing. An example apparatus configured to perform active depth sensing includes a projector. The projector is configured to emit a first distribution of light during a first time and emit a second distribution of light different from the first distribution of light during a second time. A set of final depth values of one or more objects in a scene is based on one or more reflections of the first distribution of light and one or more reflections of the second distribution of light. The projector may include a laser array, and the apparatus may be configured to switch between a first plurality of lasers of the laser array to emit light during the first time and a second plurality of laser to emit light during the second time.

Alternating light distributions for active depth sensing

Aspects of the present disclosure relate to systems and methods for active depth sensing. An example apparatus configured to perform active depth sensing includes a projector. The projector is configured to emit a first distribution of light during a first time and emit a second distribution of light different from the first distribution of light during a second time. A set of final depth values of one or more objects in a scene is based on one or more reflections of the first distribution of light and one or more reflections of the second distribution of light. The projector may include a laser array, and the apparatus may be configured to switch between a first plurality of lasers of the laser array to emit light during the first time and a second plurality of laser to emit light during the second time.

Surveying data processing device, surveying data processing method, and surveying data processing program
11580696 · 2023-02-14 · ·

A surveying data processing device includes a point cloud data acquiring unit, a three-dimensional model acquiring unit, a first correspondence relationship determining unit, an extended three-dimensional data generating unit, and a second correspondence relationship determining unit. The point cloud data acquiring unit acquires first point cloud data obtained by laser scanning, at a first viewpoint, and acquires second point cloud data obtained by laser scanning, at a second viewpoint. The three-dimensional model acquiring unit acquires data of a three-dimensional model. The first correspondence relationship determining unit obtains a correspondence relationship between the first point cloud data and the three-dimensional model. The extended three-dimensional data generating unit generates extended three-dimensional data in which the first point cloud data is extended, on the basis of the correspondence relationship. The second correspondence relationship determining unit determines a correspondence relationship between the extended three-dimensional data and the second point cloud data.

Surveying data processing device, surveying data processing method, and surveying data processing program
11580696 · 2023-02-14 · ·

A surveying data processing device includes a point cloud data acquiring unit, a three-dimensional model acquiring unit, a first correspondence relationship determining unit, an extended three-dimensional data generating unit, and a second correspondence relationship determining unit. The point cloud data acquiring unit acquires first point cloud data obtained by laser scanning, at a first viewpoint, and acquires second point cloud data obtained by laser scanning, at a second viewpoint. The three-dimensional model acquiring unit acquires data of a three-dimensional model. The first correspondence relationship determining unit obtains a correspondence relationship between the first point cloud data and the three-dimensional model. The extended three-dimensional data generating unit generates extended three-dimensional data in which the first point cloud data is extended, on the basis of the correspondence relationship. The second correspondence relationship determining unit determines a correspondence relationship between the extended three-dimensional data and the second point cloud data.

APPARATUS FOR ACQUIRING DEPTH IMAGE, METHOD FOR FUSING DEPTH IMAGES, AND TERMINAL DEVICE
20230042846 · 2023-02-09 · ·

Provided are an apparatus for acquiring a depth image, a method for fusing depth images, and a terminal device. The apparatus for acquiring a depth image includes an emitting module, a receiving module, and a processing unit. The emitting module is configured to emit a speckle array to an object, where the speckle array includes p mutually spaced apart speckles. The receiving module includes an image sensor. The processing unit is configured to receive the pixel signal and generate a sparse depth image based on the pixel signal, align an RGB image at a resolution of a*b with the sparse depth image, and fuse the aligned sparse depth image with the RGB image using a pre-trained image fusion model to obtain a dense depth image at a resolution of a*b.

APPARATUS FOR ACQUIRING DEPTH IMAGE, METHOD FOR FUSING DEPTH IMAGES, AND TERMINAL DEVICE
20230042846 · 2023-02-09 · ·

Provided are an apparatus for acquiring a depth image, a method for fusing depth images, and a terminal device. The apparatus for acquiring a depth image includes an emitting module, a receiving module, and a processing unit. The emitting module is configured to emit a speckle array to an object, where the speckle array includes p mutually spaced apart speckles. The receiving module includes an image sensor. The processing unit is configured to receive the pixel signal and generate a sparse depth image based on the pixel signal, align an RGB image at a resolution of a*b with the sparse depth image, and fuse the aligned sparse depth image with the RGB image using a pre-trained image fusion model to obtain a dense depth image at a resolution of a*b.

OBJECT POSITION DETERMINING SYSTEM AND OBJECT DETECTION SYSTEM

An object position determining system comprising: at least one light source, configured to emit light; at least one optical sensor, configured to sense optical data generated based on reflected light of the light; and a processing circuit, configured to compute distance information between the optical sensor and an object which generates the reflected light. The processing circuit further determines a position of the object according to the distance information.