H04N19/51

Video encoding technique utilizing user guided information in cloud environment

The present disclosure relates to a computer-implemented method for processing video data. The method comprises receiving a user input corresponding to a first picture of the video data, generating, based on the user input, prediction information of the first picture with respect a reference picture of the video data, and encoding the first picture using the prediction information.

Multi-model selection for neural network based tools in video coding

A method, computer program, and computer system is provided for video encoding and decoding. Video data including one or more frames is received. One or more quantization parameters associated with the received video data are determined for frame generation or enhancement of a target frame from among the one or more frames. The video data is decoded based on the determined quantization parameters.

IMAGE ENCODING METHOD AND IMAGE DECODING METHOD
20230038870 · 2023-02-09 · ·

An image encoding method for encoding an image, includes: a prediction image generation step generating a synthesis-prediction image by performing synthesis processing synthesizing an inter-prediction image and an intra-prediction image to an encoding-target block; and an encoding step encoding a difference between the prediction image generated in the prediction image generation step and a pixel value of an image of the encoding-target block, a plurality of intra prediction types are allowed to be used as the intra prediction used for the synthesis processing, and the plurality of intra prediction types include an in-screen block copy prediction, the synthesis processing includes weighting processing performed to the inter-prediction image and the intra-prediction image, and a weighting parameter of the intra-prediction image in the weighting processing is determined in accordance with an intra prediction type of the encoding-target block and prediction mode combination of a plurality of blocks adjacent to the encoding-target block.

Prediction refinement for combined inter intra prediction mode

A method for processing a video includes performing a conversion between a current block of visual media data and a corresponding coded representation of the visual media data, wherein the conversion of the current block includes determining whether a use of one or both of a bi-directional optical flow (BIO) technique or a decoder-side motion vector refinement (DMVR) technique to the current block is enabled or disabled, and wherein the determining the use of the BIO technique or the DMVR technique is based on a cost criterion associated with the current block.

Sub-block motion derivation and decoder-side motion vector refinement for merge mode
11558633 · 2023-01-17 · ·

Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.

Sub-block motion derivation and decoder-side motion vector refinement for merge mode
11558633 · 2023-01-17 · ·

Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.

Method and apparatus for reconstructing 360-degree image according to projection format

Disclosed are methods and apparatuses for image data encoding/decoding. A method for decoding a 360-degree image includes the steps of: receiving a bitstream obtained by encoding a 360-degree image; generating a prediction image by making reference to syntax information obtained from the received bitstream; adding the generated prediction image to a residual image obtained by dequantizing and inverse-transforming the bitstream, so as to obtain a decoded image; and reconstructing the decoded image into a 360-degree image according to a projection format. Therefore, the performance of image data compression can be improved.

Method and apparatus for reconstructing 360-degree image according to projection format

Disclosed are methods and apparatuses for image data encoding/decoding. A method for decoding a 360-degree image includes the steps of: receiving a bitstream obtained by encoding a 360-degree image; generating a prediction image by making reference to syntax information obtained from the received bitstream; adding the generated prediction image to a residual image obtained by dequantizing and inverse-transforming the bitstream, so as to obtain a decoded image; and reconstructing the decoded image into a 360-degree image according to a projection format. Therefore, the performance of image data compression can be improved.

Method and apparatus for reconstructing 360-degree image according to projection format

Disclosed are methods and apparatuses for image data encoding/decoding. A method for decoding a 360-degree image includes the steps of: receiving a bitstream obtained by encoding a 360-degree image; generating a prediction image by making reference to syntax information obtained from the received bitstream; adding the generated prediction image to a residual image obtained by dequantizing and inverse-transforming the bitstream, so as to obtain a decoded image; and reconstructing the decoded image into a 360-degree image according to a projection format. Therefore, the performance of image data compression can be improved.

Method and apparatus for reconstructing 360-degree image according to projection format

Disclosed are methods and apparatuses for image data encoding/decoding. A method for decoding a 360-degree image includes the steps of: receiving a bitstream obtained by encoding a 360-degree image; generating a prediction image by making reference to syntax information obtained from the received bitstream; adding the generated prediction image to a residual image obtained by dequantizing and inverse-transforming the bitstream, so as to obtain a decoded image; and reconstructing the decoded image into a 360-degree image according to a projection format. Therefore, the performance of image data compression can be improved.