G06T7/596

Systems and methods for encoding image files containing depth maps stored as metadata

Systems and methods in accordance with embodiments of the invention are configured to render images using light field image files containing an image synthesized from light field image data and metadata describing the image that includes a depth map. One embodiment of the invention includes a processor and memory containing a rendering application and a light field image file including an encoded image, a set of low resolution images, and metadata describing the encoded image, where the metadata comprises a depth map that specifies depths from the reference viewpoint for pixels in the encoded image. In addition, the rendering application configures the processor to: locate the encoded image within the light field image file; decode the encoded image; locate the metadata within the light field image file; and post process the decoded image by modifying the pixels based on the depths indicated within the depth map and the set of low resolution images to create a rendered image.

IMAGE CAPTURING APPARATUS, MONITORING SYSTEM, IMAGE PROCESSING APPARATUS, IMAGE CAPTURING METHOD, AND NON-TRANSITORY COMPUTER READABLE RECORDING MEDIUM
20190073535 · 2019-03-07 ·

There is provided an image capturing apparatus that captures a plurality of images, calculates a three-dimensional position from the plurality of images, and outputs the plurality of images and information about the three-dimensional position. The image capturing apparatus includes an image capturing unit, a camera parameter storage unit, a position calculation unit, a position selection unit, and an image complementing unit. The image capturing unit outputs the plurality of images using at least three cameras. The camera parameter storage unit stores in advance camera parameters including occlusion information. The position calculation unit calculates three dimensional positions of a plurality of points. The position selection unit selects a piece of position information relating to a subject area that does not have an occlusion, and outputs selected position information. The image complementing unit generates a complementary image, and outputs the complementary image and the selected position information.

Systems and Methods for Estimating Depth from Projected Texture using Camera Arrays
20190063905 · 2019-02-28 · ·

Systems and methods in accordance with embodiments of the invention estimate depth from projected texture using camera arrays. One embodiment of the invention includes: at least one two-dimensional array of cameras comprising a plurality of cameras; an illumination system configured to illuminate a scene with a projected texture; a processor; and memory containing an image processing pipeline application and an illumination system controller application. In addition, the illumination system controller application directs the processor to control the illumination system to illuminate a scene with a projected texture. Furthermore, the image processing pipeline application directs the processor to: utilize the illumination system controller application to control the illumination system to illuminate a scene with a projected texture capture a set of images of the scene illuminated with the projected texture; determining depth estimates for pixel locations in an image from a reference viewpoint using at least a subset of the set of images. Also, generating a depth estimate for a given pixel location in the image from the reference viewpoint includes: identifying pixels in the at least a subset of the set of images that correspond to the given pixel location in the image from the reference viewpoint based upon expected disparity at a plurality of depths along a plurality of epipolar lines aligned at different angles; comparing the similarity of the corresponding pixels identified at each of the plurality of depths; and selecting the depth from the plurality of depths at which the identified corresponding pixels have the highest degree of similarity as a depth estimate for the given pixel location in the image from the reference viewpoint.

INFORMATION PROCESSING APPARATUS AND INFORMATION PROCESSING METHOD
20190028616 · 2019-01-24 · ·

The present disclosure relates to an information processing apparatus and an information processing method that are configured to be capable of efficiently acquiring information for use in generating three-dimensional data from two-dimensional image data. A grouping block sorts two or more virtual cameras for acquiring two-dimensional image data into two or more groups. A global table generation block generates a global table in which group information related with each of two or more groups is registered. A group table generation block generates, for each group, a group table in which camera information for use in generating three-dimensional data from two-dimensional image data acquired by a virtual camera sorted into a group is registered. The present disclosure is applicable to an encoding apparatus and the like, for example.

DETERMINING A MARK IN A DATA RECORD WITH THREE-DIMENSIONAL SURFACE COORDINATES OF A SCENE, CAPTURED BY AT LEAST ONE LASER SCANNER
20190026899 · 2019-01-24 ·

A method for determining a mark in a data record with three-dimensional surface coordinates of a scene includes ascertaining a first collection of edge points in a three-dimensional coordinate system of the data record, fitting an equalization area into at least a subset of the edge points of the first collection of edge points to permit the edge points in the three-dimensional coordinate system to be partly positioned on a first side of the equalization area and partly positioned on a second side, lying opposite the first side, of the equalization area, displacing edge points of the first collection of edge points into the equalization area to permit a corrected collection of edge points to be formed, and determining the mark in the three-dimensional coordinate system based on the corrected collection of edge points or the corrected closed circumferential edge line.

ELECTRONIC DEVICE MOVED BASED ON DISTANCE FROM EXTERNAL OBJECT AND CONTROL METHOD THEREOF
20190023395 · 2019-01-24 ·

An electronic device is disclosed. The electronic device includes a sensor, an actuator, and a processor. The sensor is configured to sense at least one external object in a direction of 360 degrees outside the electronic device. The actuator configured to allow the electronic device to move or yaw. The processor is configured to verify an angle corresponding to a location of the at least one external object among the 360 degrees and a distance between the at least one external object and the electronic device using the sensor. When the distance does not belong to a specified range, the processor is also configured to move the electronic device in a direction corresponding to the angle using the actuator such that the distance belongs to the specified range.

METHODS, SYSTEMS, AND COMPUTER-READABLE STORAGE MEDIA FOR GENERATING THREE-DIMENSIONAL (3D) IMAGES OF A SCENE

Disclosed herein are methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene. According to an aspect, a method includes capturing a real-time image and a first still image of a scene. Further, the method includes displaying the real-time image of the scene on a display. The method also includes determining one or more properties of the captured images. The method also includes calculating an offset in a real-time display of the scene to indicate a target camera positional offset with respect to the first still image. Further, the method includes determining that a capture device is in a position of the target camera positional offset. The method also includes capturing a second still image. Further, the method includes correcting the captured first and second still images. The method also includes generating the three-dimensional image based on the corrected first and second still images.

Systems and Methods for Object Replacement
20180374036 · 2018-12-27 ·

Described in detail herein are systems and methods for an object replacement system. A central computing system can receives data associated with quantities of like physical objects from remote systems. The central computing system can adjust the first quantity of the like physical objects stored in the first one of the remote systems based on the second quantity of the like physical objects stored in the at least another one of the remote systems. The central computing system can determine the like physical objects are absent from the facility. An autonomous robot device can detect a vacant space at the designated location at which the like physical objects are supposed to be disposed. The autonomous robot device using the image capturing device can capture an image of the vacant space. The central computing system can determine a set of like replacement physical objects to be disposed in the vacant space.

Indoor navigation via multi-beam laser projection

An indoor navigation system is based on a multi-beam laser projector, a set of calibrated cameras, and a processor that uses knowledge of the projector design and data on laser spot locations observed by the cameras to solve the space resection problem to find the location and orientation of the projector.

IMAGE PROCESSING APPARATUS, OBJECT SHAPE ESTIMATION METHOD, AND STORAGE MEDIUM
20180367709 · 2018-12-20 ·

Highly accurate estimation results are obtained even though cameras used for shape estimation of an object are distributed in accordance with a plurality of points of interest. The image processing apparatus of the present invention includes: an estimation unit configured to estimate an object shape of an object within a multi-viewpoint video image captured by each of a plurality of camera groups in units of camera groups; and an integration unit configured to integrate estimation results of the object shapes estimated in units of camera groups based on a camera map indicating a position relationship between common image capturing areas in each of the plurality of camera groups.