H04N2013/0088

Real-world object holographic transport and communication room system
11181862 · 2021-11-23 · ·

A novel holographic transport and communication room system utilizes a single red-green-blue (RGB)-depth (RGB-D) camera to capture the motion of a dynamic target, which is required to rotate around the RGB-D camera, instead of capturing three-dimensional volume of the dynamic target conventionally with a plurality of multi-angle cameras positioned around the dynamic target. The captured 3D volume of the dynamic target subject undergoes relighting, subject depth calculations, geometrical extrapolations, and volumetric reconstructions in a machine-learning graphical transformation feedback loop to synthesize a refined real-time hologram. The resulting hologram in one holographic room system is shared with other users occupying other holographic room systems equipped with similar holographic capabilities for live bilateral or multilateral holographic visualization and collaboration. Preferably, each holographic room system also integrates a mixed-reality content synthesis table for real-time remote participant collaboration in manipulating holographic contents and a one-to-one ratio life-size holographic display and capture tubular device.

Modification of a live-action video recording using volumetric scene reconstruction to replace a designated region

A main video sequence of a live action scene is captured along with ancillary device data to provide corresponding volumetric information about the scene. The volumetric data can then be used to visually remove or replace objects in the main video sequence. A removed object is replaced by the view that would have been captured by the main video sequence had the removed object not been present in the live action scene at the time of capturing.

LAYERED SCENE DECOMPOSITION CODEC WITH VIEW INDEPENDENT RASTERIZATION

A system and methods for a CODEC driving a real-time light field display for multi-dimensional video streaming, interactive gaming and other light field display applications is provided applying a layered scene decomposition strategy. Multi-dimensional scene data is divided into a plurality of data layers. One or more polygons representative of corresponding portions of objects in the scene are obtained for each layer and used to determine a view-independent representation. This view independent representation and data layers are sampled using a plenoptic sampling scheme and rendered using hybrid rendering to encode light fields corresponding to each data layer. The resulting compressed, (layered) core representation of the multi-dimensional scene data is produced at predictable rates, reconstructed and merged at the light field display in real-time by applying view synthesis protocols, including edge adaptive interpolation, to reconstruct pixel arrays in stages (e.g. columns then rows) from reference elemental images.

Layered scene decomposition codec with volume rendering

A system and methods for a CODEC driving a real-time light field display for multi-dimensional video streaming, interactive gaming and other light field display applications is provided applying a layered scene decomposition strategy. Multi-dimensional scene data is divided into a plurality of data layers of increasing depths as the distance between a given layer and the display surface increases. Data layers which are sampled using an effective resolution function to determine a suitable sampling rate and rendered using hybrid rendering, such as perspective and oblique rendering, to encode light fields corresponding to each data layer. The resulting compressed, (layered) core representation of the multi-dimensional scene data is produced at predictable rates, reconstructed and merged at the light field display in real-time by applying view synthesis protocols, including edge adaptive interpolation, to reconstruct pixel arrays in stages (e.g. columns then rows) from reference elemental images.

Real-Time Novel View Synthesis With Forward Warping And Depth
20230362347 · 2023-11-09 ·

A fast and generalizable novel view synthesis method with sparse inputs is disclosed. The method may comprise: accessing at least a first input image with a first view of a subject in the first input image, and a second input image with a second view of the subject in the second input image using a computer system; estimating depths for pixels in the at least first and second input images; constructing a point cloud of image features from the estimated depths; and synthesizing a novel view by forward warping by using a point cloud rendering of the constructed point cloud.

Techniques for generating light field data by combining multiple synthesized viewpoints
11570418 · 2023-01-31 · ·

Techniques for efficiently generating and displaying light-field data are disclosed. In one particular embodiment, the techniques may be realized as a method for generating light-field data, the method comprising receiving input image data, synthesizing a first plurality of viewpoints based on the input image data, synthesizing a second plurality of viewpoints based on cached image data, combining the first and second plurality of viewpoints, yielding a plurality of blended viewpoints, displaying the plurality of blended viewpoints, and caching image data associated with the plurality of blended viewpoints.

METHODS AND APPARATUS FOR PROCESSING CONTENT BASED ON VIEWING INFORMATION AND/OR COMMUNICATING CONTENT
20220321860 · 2022-10-06 ·

Methods and apparatus for collecting user feedback information from viewers of content are described. Feedback information is received from viewers of content. The feedback indicates, based on head tracking information in some embodiments, where users are looking in a simulated environment during different times of a content presentation, e.g., different frame times. The feedback information is used to prioritize different portions of an environment represented by the captured image content. Resolution allocation is performed based on the feedback information and the content re-encoded based on the resolution allocation. The resolution allocation may and normally does change as the priority of different portions of the environment change.

METHODS, SYSTEMS, AND MEDIA FOR RENDERING IMMERSIVE VIDEO CONTENT WITH FOVEATED MESHES
20220321858 · 2022-10-06 ·

Methods, systems, and media for rendering immersive video content with foveated meshes are provided. In some embodiments, the method comprises: receiving a video content item; determining, using a hardware processor, whether the video content item meets at least one criterion; in response to determining that the video content item meets the at least one criterion, generating, using the hardware processor, a foveated mesh in accordance with a foveation ratio parameter on which frames of the video content item are to be projected, wherein the foveated mesh has a non-uniform position map that increases pixel density in a central portion of each frame of the video content item in comparison with peripheral portions of each frame of the video content item; and storing the video content item in a file format that includes the generated foveated mesh, wherein the immersive video content is rendered by applying the video content item as a texture to the generated foveated mesh.

THREE-DIMENSIONAL NOISE REDUCTION

Systems and methods are disclosed for image signal processing. For example, methods may include receiving a current image of a sequence of images from an image sensor; combining the current image with a recirculated image to obtain a noise reduced image, where the recirculated image is based on one or more previous images of the sequence of images from the image sensor; determining a noise map for the noise reduced image, where the noise map is determined based on estimates of noise levels for pixels in the current image, a noise map for the recirculated image, and a set of mixing weights; recirculating the noise map with the noise reduced image to combine the noise reduced image with a next image of the sequence of images from the image sensor; and storing, displaying, or transmitting an output image that is based on the noise reduced image.

Methods and apparatus for receiving and/or using reduced resolution images
11388384 · 2022-07-12 · ·

Methods and apparatus for using selective resolution reduction on images to be transmitted and/or used by a playback device are described. Prior to transmission one or more images of an environment are captured. Based on image content, motion detection and/or user input a resolution reduction operation is selected and performed. The reduced resolution image is communicated to a playback device along with information indicating a UV map corresponding to the selected resolution allocation that should be used by the playback device for rendering the communicated image. By changing the resolution allocation used and which UV map is used by the playback device different resolution allocations can be made with respect to different portions of the environment while allowing the number of pixels in transmitted images to remain constant. The playback device renders the individual images with the UV map corresponding to the resolution allocation used to generate the individual images.