Patent classifications
H04N19/30
Video encoding method, video decoding method, video encoding apparatus, and video decoding apparatus
A video encoding method of performing scalable encoding on input video includes: determining a total number of layers of the scalable encoding to be less than or equal to a maximum layer count determined according to a frame rate; and performing the scalable encoding on the input video to generate a bitstream, using the determined total number of layers.
Gradual Decoding Refresh In Video Coding
A method of decoding a coded video bitstream implemented by a video decoder is provided. The method includes the video decoder receiving the coded video bitstream, the coded video bitstream containing a gradual decoding refresh (GDR) flag corresponding to a coded video sequence (CVS); determining, by the video decoder, whether a GDR picture is present in the CVS based on a value of the GDR flag; initiating, by the video decoder, decoding of the CVS at the GDR picture when the value of the GDR flag indicates that the GDR picture is present; and generating, by the video decoder, an image according to the CVS as decoded. A corresponding method of encoding implemented by a video encoder is also disclosed.
Gradual Decoding Refresh In Video Coding
A method of decoding a coded video bitstream implemented by a video decoder is provided. The method includes the video decoder receiving the coded video bitstream, the coded video bitstream containing a gradual decoding refresh (GDR) flag corresponding to a coded video sequence (CVS); determining, by the video decoder, whether a GDR picture is present in the CVS based on a value of the GDR flag; initiating, by the video decoder, decoding of the CVS at the GDR picture when the value of the GDR flag indicates that the GDR picture is present; and generating, by the video decoder, an image according to the CVS as decoded. A corresponding method of encoding implemented by a video encoder is also disclosed.
Utilization of non-sub block spatial-temporal motion vector prediction in inter mode
A method of video processing includes determining, for a current block, at least one motion candidate list; and performing a conversion between the current block and a bitstream representation of the current block using the at least one motion candidate list, the at least one motion candidate list including at least one motion candidate derived from a set of neighboring blocks including one or more spatial and temporal neighboring blocks.
Utilization of non-sub block spatial-temporal motion vector prediction in inter mode
A method of video processing includes determining, for a current block, at least one motion candidate list; and performing a conversion between the current block and a bitstream representation of the current block using the at least one motion candidate list, the at least one motion candidate list including at least one motion candidate derived from a set of neighboring blocks including one or more spatial and temporal neighboring blocks.
Triangle prediction with applied-block settings and motion storage settings
A video coder receives data from a bitstream for a block of pixels to be encoded or decoded as a current block of a current picture of a video. Upon determining that an applied block setting of the current block satisfies a threshold condition, the video coder generates a first prediction based on a first motion information for a first prediction unit of the current block. The video coder generates a second prediction based on a second motion information for a second prediction unit of the current block. The video coder generates a third prediction based on the first and second motion information for an overlap prediction region that is defined based on a partitioning between the first prediction unit and the second prediction unit. The video coder encodes or decodes the current block by using the first, second, and third predictions.
Triangle prediction with applied-block settings and motion storage settings
A video coder receives data from a bitstream for a block of pixels to be encoded or decoded as a current block of a current picture of a video. Upon determining that an applied block setting of the current block satisfies a threshold condition, the video coder generates a first prediction based on a first motion information for a first prediction unit of the current block. The video coder generates a second prediction based on a second motion information for a second prediction unit of the current block. The video coder generates a third prediction based on the first and second motion information for an overlap prediction region that is defined based on a partitioning between the first prediction unit and the second prediction unit. The video coder encodes or decodes the current block by using the first, second, and third predictions.
Coded video sequence start access unit in video coding
A video decoder may be configured to output layers of video data according to layers specified in an output layer set mode as well as based on the actual layers received in an access unit. Accordingly, the video coder may be configured to output layers actually received rather than restart a video bitstream in situations where the layers actually received do not match the layers specified in the output layer set mode. In another example, a video encoder may disable inter prediction when the sub-picture ID of a sub-picture in a currently coded picture does not match the sub-picture ID of a sub-picture in a reference picture.
Encoder and decoder for encoding and decoding images
There are disclosed techniques for encoding and/or decoding multiple image information, e.g. through recurrent neural network based stereo compression. In an image encoder (1), a primary block (100) encodes a primary image information, and a secondary block (300) encodes a secondary image information. States for the primary block are transformed onto states for the secondary block at a transformation block (200), which keeps into account correspondence information (e.g. disparity information) between the first image information and the second image information. In an image decoder (1b), a primary block (100) decodes an encoded version of a primary image information, and a secondary block (300) encodes an encoded version of a secondary image information. States for the primary block are transformed onto states for the secondary block at a transformation block (200), which keeps into account correspondence information (e.g. disparity information) between the first image information and the second image information.
Encoder and decoder for encoding and decoding images
There are disclosed techniques for encoding and/or decoding multiple image information, e.g. through recurrent neural network based stereo compression. In an image encoder (1), a primary block (100) encodes a primary image information, and a secondary block (300) encodes a secondary image information. States for the primary block are transformed onto states for the secondary block at a transformation block (200), which keeps into account correspondence information (e.g. disparity information) between the first image information and the second image information. In an image decoder (1b), a primary block (100) decodes an encoded version of a primary image information, and a secondary block (300) encodes an encoded version of a secondary image information. States for the primary block are transformed onto states for the secondary block at a transformation block (200), which keeps into account correspondence information (e.g. disparity information) between the first image information and the second image information.