Patent classifications
H04N19/577
HYBRID INTER BI-PREDICTION IN VIDEO CODING
A video decoder can be configured to determine that a current block of the video data is coded in a bi-prediction inter mode; receive a first syntax element identifying a motion vector predictor from a first candidate list of motion vector predictors; receive a second syntax element identifying a motion vector difference; determine a first motion vector for the current block based on the motion vector predictor and the motion vector difference; determine a second motion vector for the current block from a second list of candidate motion vector predictors based on bilateral matching; and determine a prediction block for the current block using the first motion vector and the second motion vector.
Predictive motion vector coding
Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.
Predictive motion vector coding
Overlapped block disparity estimation and compensation is described. Compensating for images with overlapped block disparity compensation (OBDC) involves determining if OBDC is enabled in a video bit stream, and determining if OBDC is enabled for one or more macroblocks that neighbor a first macroblock within the video bit stream. The neighboring macroblocks may be transform coded. If OBDC is enabled in the video bit stream and for the one or more neighboring macroblocks, predictions may be made for a region of the first macroblock that has an edge adjacent with the neighboring macroblocks. OBDC can be causally applied. Disparity compensation parameters or modes may be shared amongst views or layers. A variety of predictions may be used with causally-applied OBDC.
Method and apparatus of multi-hypothesis in video coding
A method and apparatus of prediction for video coding using MH (Multiple Hypothesis) mode are disclosed. According to this method, a block is partitioned into a first partition and a second partition. A first candidate and a second candidate are derived for the first and second partitions respectively. At least one of the first candidate and the second candidate is derived using a candidate list derived for a regular Merge mode (or also called as normal Merge mode). An MH prediction generated by blending a first prediction corresponding to a first candidate and a second prediction corresponding to a second candidate, and the MH prediction is applied to a part of the current block.
ENCODING AND DECODING METHODS AND APPARATUS
A method for decoding or encoding comprising: determining if an inter prediction mode called weighted prediction is enabled for a current block; and, if weighted prediction is enabled, modifying a part of a reconstruction process of said current block related to the weighted prediction and/or to at least another prediction tool of a first set of prediction tools, the first set comprising a triangle prediction mode, a geometrical prediction mode, a local illumination compensation mode, a combined intra inter prediction mode, and modes in which motion information used for motion compensation may be refined based on an optical flow model.
Reducing motion vector information transmission in bi-directional temporal prediction
A method for inter-coding video is provided in which transmission bandwidth requirements associated with second motion vectors for bi-directional temporal prediction is reduced. In the method motion vector information for only one of the two motion vectors for bi-directional temporal prediction can be transmitted together with information on how to derive or construct the second motion vector. Thus, rather than sending express information regarding two motion vectors, express information related to only one motion vector along with information related to reconstruction/derivation of the second motion vector is transmitted, thus reducing bandwidth requirements and increasing coding efficiency.
METHOD AND APPARATUS OF HARMONIZING WEIGHTED PREDICTION WITH NON-RECTANGULAR MERGE MODES
A method of coding is described. The method can include obtaining a bitstream for a current picture, obtaining a value of a first indicator for the current picture according to the bitstream indicating a slice type, and obtaining a value of a second indicator for the current picture according to the bitstream indicating whether a weighted prediction parameter is present in a picture header or slice header of the bitstream. The method can also include parsing a value of the weighted prediction parameter for a current block of a current slice of the current picture from the bitstream. Furthermore, the method can include predicting the current block according to the value of the weighted prediction parameter.
SIGNALLING OF WEIGHTS OF A REFERENCE PICTURE LIST
A method includes performing a conversion, according to a rule, between a current slice of a current picture of a video and a bitstream of the video, wherein the rule specifies that a first syntax element of a picture parameter set (PPS) and a second syntax element of the PPS control whether a third syntax element is included in the bitstream, and wherein the first syntax element indicates whether a weighted prediction is enabled for bi-directional slices of coded pictures referring to the PPS, the second syntax element indicates whether information related to the weighted prediction is present in picture headers or slice headers of coded pictures referring to the PPS, and the third syntax element indicates a number of weights associated with a reference picture list 1 of the current slice.
Picture prediction method and apparatus, and codec
A picture prediction method is provided, which includes: obtaining initial motion information of a current picture block; determining, based on a matching cost criterion, that positions of a pair of reference blocks are a position of a target forward reference block of the current picture block and a position of a target backward reference block of the current picture block, where positions of each pair of reference blocks include a position of a forward reference block and a position of a backward reference block; and for the positions of each pair of reference blocks, a first position offset and a second position offset are in a mirror relationship; and obtaining a predicted value of a pixel value of the current picture block based on a pixel value of the target forward reference block and a pixel value of the target backward reference block.
Neural Network-Based Video Compression with Spatial-Temporal Adaptation
A mechanism for processing video data is disclosed. A determination is made to apply an end-to-end neural network-based video codec to a current video unit of a video. The end-to-end neural network-based video codec comprises a spatial-temporal adaptive compression (STAC) component including a frame extrapolative compression (FEC) branch and an image compression branch. A conversion is performed between the current video unit and a bitstream of the video via the end-to-end neural network-based video codec.