Patent classifications
H04N19/85
METHODS AND DEVICES FOR ENCODING AND DECODING A DATA STREAM REPRESENTING AT LEAST ONE IMAGE THAT DISABLES POST-PROCESSING OF RECONSTRUCTED BLOCK BASED ON PREDICTION MODE
A method for decoding a data stream representative of an image split into blocks. For a current block of the image, an item of information indicating a coding mode among a first and a second coding mode of the current block is decoded from the data stream and the current block is decoded depending on this information. When the coding mode of the current block corresponds to the second coding mode, the current block is reconstructed from a prediction obtained, for each pixel, from another previously decoded pixel belonging to the current block or to a previously decoded block of the image, and from a decoded residue associated with the pixel. At least one processing method is applied to the reconstructed current block for at least one pixel of the current block depending on the coding mode of the current block and/or the coding mode of the neighbouring blocks.
METHOD, APPARATUS, AND RECORDING MEDIUM FOR REGION-BASED DIFFERENTIAL IMAGE ENCODING/DECODING
Disclosed herein are a video-decoding method and apparatus and a video encoding method and apparatus, and more particularly a method and an apparatus which perform region-differential image encoding/decoding using a recovered image. In accordance with an encoding method according to an embodiment, a recovered low-quality image is generated by performing encoding on an original image and a recovered high-quality image is generated using the recovered low-quality image. An image is segmented into multiple regions, and encoded reconstruction information for generating a reconstructed high-quality image is generated by performing encoding on the image.
Generative adversarial neural network assisted video reconstruction
A latent code defined in an input space is processed by the mapping neural network to produce an intermediate latent code defined in an intermediate latent space. The intermediate latent code may be used as appearance vector that is processed by the synthesis neural network to generate an image. The appearance vector is a compressed encoding of data, such as video frames including a person's face, audio, and other data. Captured images may be converted into appearance vectors at a local device and transmitted to a remote device using much less bandwidth compared with transmitting the captured images. A synthesis neural network at the remote device reconstructs the images for display.
Encoding and decoding image data
Certain aspects of the present disclosure provide techniques for encoding image data for one or more images. In one embodiment, a method includes the steps of downscaling the one or more images, and encoding the one or more downscaled images using an image codec. Another embodiment concerns a computer-implemented method of decoding encoded image data, and a computer-implemented method of encoding and decoding image data.
Encoding and decoding image data
Certain aspects of the present disclosure provide techniques for encoding image data for one or more images. In one embodiment, a method includes the steps of downscaling the one or more images, and encoding the one or more downscaled images using an image codec. Another embodiment concerns a computer-implemented method of decoding encoded image data, and a computer-implemented method of encoding and decoding image data.
Techniques for video compression
A method is disclosed. In the method, color differences are calculated between a current video frame and a motion predicted version of the current video frame based on a human visual system's ability to perceive the color differences. Also, information in a difference frame is discarded based on the color differences. The difference frame includes differences between the current video frame and the motion predicted version of the current video frame.
Techniques for video compression
A method is disclosed. In the method, color differences are calculated between a current video frame and a motion predicted version of the current video frame based on a human visual system's ability to perceive the color differences. Also, information in a difference frame is discarded based on the color differences. The difference frame includes differences between the current video frame and the motion predicted version of the current video frame.
Constraints for inter-layer referencing
A video coding method using inter-layer prediction or referencing is provided. A video decoder receives data from a bitstream carrying data for video pictures in a plurality of different layers. At least one of the plurality of layers comprises temporal sublayers that correspond to levels in a hierarchical temporal prediction structure. Each temporal sublayer is associated with a temporal identifier. The video decoder receives an inter-layer prediction constraint parameter constraining a maximum temporal sublayer used in inter-layer prediction. The video decoder reconstructs a first picture in a first layer by referencing data of a second picture in a second layer. A temporal identifier of the referenced data satisfies the received inter-layer prediction constraint parameter.
Signal reshaping for high dynamic range signals
In a method to improve backwards compatibility when decoding high-dynamic range images coded in a wide color gamut (WCG) space which may not be compatible with legacy color spaces, hue and/or saturation values of images in an image database are computed for both a legacy color space (say, YCbCr-gamma) and a preferred WCG color space (say, IPT-PQ). Based on a cost function, a reshaped color space is computed so that the distance between the hue values in the legacy color space and rotated hue values in the preferred color space is minimized HDR images are coded in the reshaped color space. Legacy devices can still decode standard dynamic range images assuming they are coded in the legacy color space, while updated devices can use color reshaping information to decode HDR images in the preferred color space at full dynamic range.
Signal reshaping for high dynamic range signals
In a method to improve backwards compatibility when decoding high-dynamic range images coded in a wide color gamut (WCG) space which may not be compatible with legacy color spaces, hue and/or saturation values of images in an image database are computed for both a legacy color space (say, YCbCr-gamma) and a preferred WCG color space (say, IPT-PQ). Based on a cost function, a reshaped color space is computed so that the distance between the hue values in the legacy color space and rotated hue values in the preferred color space is minimized HDR images are coded in the reshaped color space. Legacy devices can still decode standard dynamic range images assuming they are coded in the legacy color space, while updated devices can use color reshaping information to decode HDR images in the preferred color space at full dynamic range.