Patent classifications
G06T7/596
Technique For Determining A Position of One Or More Imaged Markers In An Image Coordinate System
A method and a device for determining a respective position of one or more markers in a 3D image coordinate system are provided. A plurality of image data sets taken from a 3D volume in which an object and one or more markers are disposed. The 3D volume comprises a central volume containing at least a portion of the object and further comprises a peripheral volume adjacent to the central volume and containing the one or more markers. The image data sets have been taken from at least one of different positions and different orientations relative to the object. A first subset comprises image data sets that each includes at least one dedicated marker of the one or more markers and a second subset comprises at least one image data set that does not include the at least one dedicated marker. The method further comprises determining, from the image data sets, a position of the at least one dedicated marker in a 3D image coordinate system of a 3D reconstruction of the central volume with the object portion.
Plate reconstruction of obscured views of captured imagery using arbitrary captured inputs
An imagery processing system obtains capture inputs from capture devices that might have capture parameters and characteristics that differ from those of a main imagery capture device. By normalizing outputs of those capture devices, potentially arbitrary capture devices could be used for reconstructing portions of a scene captured by the main imagery capture device when reconstructing a plate of the scene to replace an object in the scene with what the object obscured in the scene. Reconstruction could be of one main image, a stereo pair of images, or some number, N, of images where N>2.
Motion correction of angiography images for 3D reconstruction of coronary arteries
Systems and methods for computing a transformation for correction motion between a first medical image and a second medical image are provided. One or more landmarks are detected in the first medical image and the second medical image. A first tree of the anatomical structure is generated from the first medical image based on the one or more landmarks detected in the first medical image and a second tree of the anatomical structure is generated from the second medical image based on the one or more landmarks detected in the second medical image. The one or more landmarks detected in the first medical image are mapped to the one or more landmarks detected in the second medical image based on the first tree and the second tree. A transformation to align the first medical image and the second medical image is computed based on the mapping.
METHODS, SYSTEMS, AND COMPUTER-READABLE STORAGE MEDIA FOR GENERATING THREE-DIMENSIONAL (3D) IMAGES OF A SCENE
Disclosed herein are methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene. According to an aspect, a method includes capturing a real-time image and a first still image of a scene. Further, the method includes displaying the real-time image of the scene on a display. The method also includes determining one or more properties of the captured images. The method also includes calculating an offset in a real-time display of the scene to indicate a target camera positional offset with respect to the first still image. Further, the method includes determining that a capture device is in a position of the target camera positional offset. The method also includes capturing a second still image. Further, the method includes correcting the captured first and second still images. The method also includes generating the three-dimensional image based on the corrected first and second still images.
Communication protocol for streaming mixed-reality environments between multiple devices
An immersive content presentation system can capture the motion or position of a performer in a real-world environment. A game engine can be modified to receive the position or motion of the performer and identify predetermined gestures or positions that can be used to trigger actions in a 3-D virtual environment, such as generating a digital effect, transitioning virtual assets through an animation graph, adding new objects, and so forth. The use of the 3-D environment can be rendered and composited views can be generated. Information for constructing the composited views can be streamed to numerous display devices in many different physical locations using a customized communication protocol. Multiple real-world performers can interact with virtual objects through the game engine in a shared mixed-reality experience.
Systems and Methods for Encoding Image Files Containing Depth Maps Stored as Metadata
Systems and methods in accordance with embodiments of the invention are configured to render images using light field image files containing an image synthesized from light field image data and metadata describing the image that includes a depth map. One embodiment of the invention includes a processor and memory containing a rendering application and a light field image file including an encoded image, a set of low resolution images, and metadata describing the encoded image, where the metadata comprises a depth map that specifies depths from the reference viewpoint for pixels in the encoded image. In addition, the rendering application configures the processor to: locate the encoded image within the light field image file; decode the encoded image; locate the metadata within the light field image file; and post process the decoded image by modifying the pixels based on the depths indicated within the depth map and the set of low resolution images to create a rendered image.
Depth map generation device for merging multiple depth maps
A depth map generation device for merging multiple depth maps includes at least three image capturers, a depth map generator, and a mixer. The at least three image capturers form at least two image capture pairs. The depth map generator is coupled to the at least three image capturers for generating a depth map corresponding to each image capturer pair of the at least two image capture pairs according to an image pair captured by the each image capturer. The mixer is coupled to the depth map generator for merging at least two depth maps corresponding to the at least two image capturer pairs to generate a final depth map, wherein the at least two depth maps have different characteristics.
FOREGROUND-BACKGROUND-AWARE ATROUS MULTISCALE NETWORK FOR DISPARITY ESTIMATION
A system for disparity estimation includes one or more feature extractor modules configured to extract one or more feature maps from one or more input images; and one or more semantic information modules connected at one or more outputs of the one or more feature extractor modules, wherein the one or more semantic information modules are configured to generate one or more foreground semantic information to be provided to the one or more feature extractor modules for disparity estimation at a next training epoch.
TIME-LAPSE STEREO MACRO PHOTOGRAPHY SYSTEMS AND METHODS AND STEREO TIME-LAPSE VIDEO MADE WITH SAME
Systems and methods for macro stereo time-lapse photography, producing a stereographic time-lapse digital video, and macro stereographic time-lapse digital videos. A method of producing a sequence of time-lapse stereographic images of a subject, by positioning a camera with a macro lens at a first position relative to the subject; using the camera to obtain a first stack of images of the subject from the first position; positioning the camera at a second position relative to the subject; using the camera to obtain a second stack of images of the subject from the second position; and storing the first stack of images and the second stack of images as a stack pair; and then selectively repeating.
Information processing device, system, information processing method, and storage medium
An information processing device has at least one memory that stores instructions, and at least one processor coupled to the at least one memory, and configured to set, on a combined image in which a plurality of images captured through a plurality of imaging units are combined, an object region that is a region to be an object of a predetermined process, and to perform, when a boundary between images included in the plurality of images in the combined image is included in the set object region, a predetermined display process on a display unit.