G06T7/596

Image processing apparatus, image processing method and storage medium
11037323 · 2021-06-15 · ·

The image processing apparatus of the present invention is an image processing apparatus that generates a virtual viewpoint image based on image data obtained by capturing an image capturing area from a plurality of directions by a plurality of cameras, the image processing apparatus including: an acquisition unit configured to acquire viewpoint information of a virtual viewpoint; an area determination unit configured to determine a three-dimensional area in accordance with a position and a size of a specific object within the image capturing area; and a generation unit configured to generate the virtual viewpoint image in accordance with the virtual viewpoint indicated by the viewpoint information based on determination by the area determination unit such that an object within a field of view in accordance with the virtual viewpoint and not included in the three-dimensional area determined by the area determination unit is not displayed in the virtual viewpoint image.

Real-time views of mixed-reality environments responsive to motion-capture data

An immersive content presentation system can capture the motion or position of a performer in a real-world environment. A game engine can be modified to receive the position or motion of the performer and identify predetermined gestures or positions that can be used to trigger actions in a 3-D virtual environment, such as generating a digital effect, transitioning virtual assets through an animation graph, adding new objects, and so forth. The use of the 3-D environment can be rendered and composited views can be generated. Information for constructing the composited views can be streamed to numerous display devices in many different physical locations using a customized communication protocol. Multiple real-world performers can interact with virtual objects through the game engine in a shared mixed-reality experience.

SYSTEMS AND METHODS FOR DEPTH ESTIMATION BY LEARNING TRIANGULATION AND DENSIFICATION OF SPARSE POINTS FOR MULTI-VIEW STEREO
20210279904 · 2021-09-09 · ·

Systems and methods for estimating depths of features in a scene or environment surrounding a user of a spatial computing system, such as a virtual reality, augmented reality or mixed reality (collectively, cross reality) system, in an end-to-end process. The estimated depths can be utilized by a spatial computing system, for example, to provide an accurate and effective 3D cross reality experience.

Systems and methods for image texture uniformization for multiview object capture
11120606 · 2021-09-14 · ·

An image processing method is disclosed, comprising: receiving a plurality of images of an object; and, for each image, determining a brightness value for each point on the object that is visible, and generating a tile grid for the image. Each point that is visible in an image corresponds to one tile in the tile grid. For each tile pair where at least one point is visible in both tiles of the tile pair, a brightness offset value is determined. For each image, an offset matrix is generated. Values of the offset matrix comprise brightness offsets to apply to each tile of a tile grid for the image to normalize the brightness of the image with the other images. A plurality of reconstructed images is generated based on performing, for each image, transformation operations on the given image based on the corresponding offset matrix to generate a reconstructed image.

RECONSTRUCTION OF OBSCURED VIEWS OF CAPTURED IMAGERY USING ARBITRARY CAPTURED INPUTS

An imagery processing system obtains capture inputs from capture devices that might have capture parameters and characteristics that differ from those of a main imagery capture device. By normalizing outputs of those capture devices, potentially arbitrary capture devices could be used for reconstructing portions of a scene captured by the main imagery capture device when reconstructing a plate of the scene to replace an object in the scene with what the object obscured in the scene. Reconstruction could be of one main image, a stereo pair of images, or some number, N, of images where N>2.

PLATE RECONSTRUCTION OF OBSCURED VIEWS OF CAPTURED IMAGERY USING ARBITRARY CAPTURED INPUTS

An imagery processing system obtains capture inputs from capture devices that might have capture parameters and characteristics that differ from those of a main imagery capture device. By normalizing outputs of those capture devices, potentially arbitrary capture devices could be used for reconstructing portions of a scene captured by the main imagery capture device when reconstructing a plate of the scene to replace an object in the scene with what the object obscured in the scene. Reconstruction could be of one main image, a stereo pair of images, or some number, N, of images where N>2.

MULTI-VIEW THREE-DIMENSIONAL POSITIONING
20210183098 · 2021-06-17 ·

A device determines positions of objects in a scene. The device obtains object detection data (ODD) which identifies the objects and locations of reference points of the objects in 2D images of the scene. The device processes the ODD to generate candidate association data (CAD) which associates pairs of objects between the images, computes estimated 3D positions in the scene for associated pairs of objects in the CAD, and performs clustering of the estimated positions. The device further generates, based on estimated 3D positions in one or more clusters, final association data (FAD) which associates one or more objects between the images, and computes one or more final 3D positions in the scene for one or more reference points of the one or more objects in the FAD. The final 3D position(s) represent the 3D position or the 3D pose of the respective object in the scene.

Systems and methods for encoding image files containing depth maps stored as metadata

Systems and methods in accordance with embodiments of the invention are configured to render images using light field image files containing an image synthesized from light field image data and metadata describing the image that includes a depth map. One embodiment of the invention includes a processor and memory containing a rendering application and a light field image file including an encoded image, a set of low resolution images, and metadata describing the encoded image, where the metadata comprises a depth map that specifies depths from the reference viewpoint for pixels in the encoded image. In addition, the rendering application configures the processor to: locate the encoded image within the light field image file; decode the encoded image; locate the metadata within the light field image file; and post process the decoded image by modifying the pixels based on the depths indicated within the depth map and the set of low resolution images to create a rendered image.

Image processing apparatus, image capturing apparatus, image processing method, and storage medium
11032533 · 2021-06-08 · ·

An image processing apparatus processes a first image and a second image so as to detect a corresponding pixel in the second image which corresponds to a target pixel in the first image. The first image has a first parameter value, and the second image has a second parameter value different from the first parameter value. The first parameter value and the second parameter value are values of optical parameters of image capturing systems used to capture the first image and the second image. The image processing apparatus includes an area setter that sets a two-dimensional search area as a partial area in which the corresponding pixel is to be searched in the second image, based on a predetermined range in which each of the first and second parameter values can change, and a detector that detects the corresponding pixel by searching the two-dimensional search area.

SYSTEM AND METHOD FOR 3D PROFILE DETERMINATION USING MODEL-BASED PEAK SELECTION
20210148694 · 2021-05-20 ·

This invention provides a system and method for selecting the correct profile from a range of peaks generated by analyzing a surface with multiple exposure levels applied at discrete intervals. The cloud of peak information is resolved by comparison to a model profile into a best candidate to represent an accurate representation of the object profile. Illustratively, a displacement sensor projects a line of illumination on the surface and receives reflected light at a sensor assembly at a set exposure level. A processor varies the exposure level setting in a plurality of discrete increments, and stores an image of the reflected light for each of the increments. A determination process combines the stored images and aligns the combined images with respect to a model image. Points from the combined images are selected based upon closeness to the model image to provide a candidate profile of the surface.