Patent classifications
G06T7/596
Systems and methods for encoding image files containing depth maps stored as metadata
Systems and methods for storing images synthesized from light field image data and metadata describing the images in electronic files in accordance with embodiments of the invention are disclosed. One embodiment includes a processor and memory containing an encoding application and light field image data, where the light field image data comprises a plurality of low resolution images of a scene captured from different viewpoints. In addition, the encoding application configures the processor to synthesize a higher resolution image of the scene from a reference viewpoint using the low resolution images, where synthesizing the higher resolution image involves creating a depth map that specifies depths from the reference viewpoint for pixels in the higher resolution image; encode the higher resolution image; and create a light field image file including the encoded image, the low resolution images, and metadata including the depth map.
ELECTRONIC APPARATUS AND METHOD FOR CONTROLLING THEREOF
An electronic apparatus and a method for controlling thereof are provided. The electronic apparatus includes a first display comprising a first display panel for displaying a multi-view image and a micro-lens array which is arranged on a front surface of the first display panel and is for adjusting directions of lights output through the first display panel, a second display comprising a second display panel which is arranged on the front surface of the first display and displays a two-dimensional image corresponding to the multi-view image and has a transmittance greater than or equal to a predetermined value, memory storing one or more computer program, and one or more processors communicatively coupled to the first display, the second display, and the memory.
OBJECT POSE ESTIMATION IN VISUAL DATA
The pose of an object may be estimated based on fiducial points identified in a visual representation of the object. Each fiducial point may correspond with a component of the object, and may be associated with a first location in an image of the object and a second location in a 3D coordinate pace. A 3D skeleton of the object may be determined by connecting the locations in the 3D space, and the object's pose may be determined based on the 3D skeleton.
SYSTEMS AND METHODS FOR ELECTRON CRYOTOMOGRAPHY RECONSTRUCTION
Described herein are methods and non-transitory computer-readable media of a computing system configured to obtain a plurality of images of an object from a plurality of orientations at a plurality of times. A machine learning model is encoded to represent a continuous density field of the object that maps a spatial coordinate to a density value. The machine learning model comprises a deformation module configured to deform the spatial coordinate in accordance with a timestamp and a trained deformation weight. The machine learning model further comprises a neural radiance module configured to derive the density value in accordance with the deformed spatial coordinate, the timestamp, a direction, and a trained radiance weight. The machine learning model is trained using the plurality of images. A three-dimensional structure of the object is constructed based on the trained machine learning model.
Perception uncertainty
A computer-implemented method of perceiving structure in an environment comprises steps of: receiving at least one structure observation input pertaining to the environment; processing the at least one structure observation input in a perception pipeline to compute a perception output; determining one or more uncertainty source inputs pertaining to the structure observation input; and determining for the perception output an associated uncertainty estimate by applying, to the one or more uncertainty source inputs, an uncertainty estimation function learned from statistical analysis of historical perception outputs.
Method and apparatus for obtaining extended depth of field image and electronic device
A method and an apparatus for obtaining an extended depth of field image, where the method includes: determining a target focal length range based on an initial focal length, where the target focal length range includes the initial focal length; obtaining a plurality of images of a photographed object for a plurality of focal lengths in the target focal length range; and registering and fusing the plurality of images to obtain an extended depth of field image. The target focal length range of concern to a user is selected based on the initial focal length, such that there is no need to obtain images of all focal lengths of a lens, a quantity of obtained images and processing time of registration and fusion can be reduced.
Systems and methods for encoding light field image files
Systems and methods configured to store images synthesized from light field image data and metadata describing the images in electronic files and render images using the stored image and the metadata in accordance with embodiments of the invention are disclosed. One embodiment includes a processor and memory containing an encoding application and light field image data, where the light field image data comprises a plurality of low resolution images of a scene captured from different viewpoints. In addition, the encoding application configures the processor to synthesize a higher resolution image of the scene from a reference viewpoint using the low resolution images, where synthesizing the higher resolution image involves creating a depth map that specifies depths from the reference viewpoint for pixels in the higher resolution image; encode the higher resolution image; and create a light field image file including the encoded image and metadata including the depth map.
Off-band resolution emhancement
A method of enhancing an image includes increasing sampling rate of a first image to a target sampling rate to form an interpolated image. The method also includes processing a second image through a high pass filter to form a high pass features image, wherein the second image is at the target sampling rate. The method also includes extracting detail from the high pass features image relevant to the first image, merging the detail from the high pass features image with the interpolated image to form a prediction image at the target sampling rate, and outputting the prediction image.
IMAGE PROCESSING APPARATUS, IMAGING APPARATUS, AND IMAGE PROCESSING METHOD
There is provided an image processing apparatus including a distance information generation portion configured to generate first distance information about an object to be measured based on a phase difference between images provided by a plurality of first cameras having a first base length, and generate second distance information about the object to be measured based on a phase difference between images provided by a plurality of second cameras having a second base length that is different from the first base length; and a distance extraction portion configured to extract distance information from an imaging position and the object to be measured based on the first distance information and the second distance information.
Image processing apparatus, imaging apparatus, and image processing method
There is provided an image processing apparatus including a distance information generation portion configured to generate first distance information about an object to be measured based on a phase difference between images provided by a plurality of first cameras having a first base length, and generate second distance information about the object to be measured based on a phase difference between images provided by a plurality of second cameras having a second base length that is different from the first base length; and a distance extraction portion configured to extract distance information from an imaging position and the object to be measured based on the first distance information and the second distance information.