H04N13/161

CODING AND DECODING OF AN OMNIDIRECTIONAL VIDEO
20230239503 · 2023-07-27 ·

A process for coding an image of a view from among a plurality of views, including the following steps: selecting a first or a second coding method to code image data from the image; generating a data signal containing information indicating whether it is the first or the second coding method that has been selected, and, if it is the first coding method, coding the original image data so as to provide coded original data, and, if it is the second coding method, coding processed image data from the image obtained by image processing of the original image data so as to provide coded processed data; and coding information describing the image processing which has been applied.

A METHOD AND APPARATUS FOR ENCODING AND DECODING VOLUMETRIC CONTENT IN AND FROM A DATA STREAM
20230239451 · 2023-07-27 ·

Methods and apparatus for encoding and decoding a volumetric scene are disclosed. A set of attribute and geometry patches is obtained by projecting samples of the volumetric scene onto the patches according to projection parameters. If the geometry patch is comparable to a planar layer located at a constant depth according to the projection parameters, only the attribute patch is packed in an attribute atlas image and the depth value is encoded in metadata. Otherwise, both attribute and geometry patches are packed in an atlas. At the decoding, if metadata for an attribute patch indicates that its geometry may be determined from the projection parameters and a constant depth, the attributes are inverse projected on a planar layer. Otherwise, attributes are inverse projected according to the associated geometry patch.

A METHOD AND APPARATUS FOR ENCODING AND DECODING VOLUMETRIC CONTENT IN AND FROM A DATA STREAM
20230239451 · 2023-07-27 ·

Methods and apparatus for encoding and decoding a volumetric scene are disclosed. A set of attribute and geometry patches is obtained by projecting samples of the volumetric scene onto the patches according to projection parameters. If the geometry patch is comparable to a planar layer located at a constant depth according to the projection parameters, only the attribute patch is packed in an attribute atlas image and the depth value is encoded in metadata. Otherwise, both attribute and geometry patches are packed in an atlas. At the decoding, if metadata for an attribute patch indicates that its geometry may be determined from the projection parameters and a constant depth, the attributes are inverse projected on a planar layer. Otherwise, attributes are inverse projected according to the associated geometry patch.

DIRECTED INTERPOLATION AND DATA POST-PROCESSING

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images.

DIRECTED INTERPOLATION AND DATA POST-PROCESSING

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images.

A METHOD AND APPARATUS FOR DECODING A 3D VIDEO

A method and an apparatus for decoding a volumetric video are disclosed. Such a method comprises receiving a data stream representative of a file comprising information for selecting, according to a rendering viewpoint, at least one atlas comprising color and depth data patches associated with a viewpoint in said volumetric video, said color and depth data patches being generated with respect to depth and color reference data acquired from a reference viewpoint in said volumetric video.

Efficient multi-view coding using depth-map estimate and update

This disclosure is directed to coding a multi-view signal, which includes processing a list of plurality of motion vector candidates associated with a coding block of a current picture in a dependent view of the multi-view signal. Such processing includes estimating a first motion vector based on a second motion vector associated with a reference block in a current picture of a reference view of the multi-view signal, the reference block corresponding to the coding block of the current picture in the dependent view. The first motion vector is added into the list, and an index is used that specifies at least one candidate from the list to be used for motion-compensated prediction. The coding block in the current picture is coded by performing the motion-compensated prediction based on the at least one candidate indicated by the index.

Efficient multi-view coding using depth-map estimate and update

This disclosure is directed to coding a multi-view signal, which includes processing a list of plurality of motion vector candidates associated with a coding block of a current picture in a dependent view of the multi-view signal. Such processing includes estimating a first motion vector based on a second motion vector associated with a reference block in a current picture of a reference view of the multi-view signal, the reference block corresponding to the coding block of the current picture in the dependent view. The first motion vector is added into the list, and an index is used that specifies at least one candidate from the list to be used for motion-compensated prediction. The coding block in the current picture is coded by performing the motion-compensated prediction based on the at least one candidate indicated by the index.

Systems and methods for improved delivery and display of 360-degree content

Systems and methods are provided for generating a viewport for display. A user preference for a character and/or a genre of a scene in a spherical media content item is determined, wherein the spherical media content item comprises a plurality of tiles. A tile of the plurality of tiles is identified based on the determined user preference. A viewport to be generated for display at a computing device is predicted, based on the identified tile. A first tile to be transmitted to a computing device at a first resolution is identified, based on the predicted viewport to be generated for display. The tile is transmitted, to the computing device, at the first resolution.

Systems and methods for improved delivery and display of 360-degree content

Systems and methods are provided for generating a viewport for display. A user preference for a character and/or a genre of a scene in a spherical media content item is determined, wherein the spherical media content item comprises a plurality of tiles. A tile of the plurality of tiles is identified based on the determined user preference. A viewport to be generated for display at a computing device is predicted, based on the identified tile. A first tile to be transmitted to a computing device at a first resolution is identified, based on the predicted viewport to be generated for display. The tile is transmitted, to the computing device, at the first resolution.