Patent classifications
H04N19/132
Partitioning Information In Neural Network-Based Video Coding
A method implemented by a video coding apparatus. The method includes applying a neural network (NN) filter to an unfiltered sample of a video unit to generate a filtered sample, where the NN filter includes an NN filter model generated based on partitioning information of the video unit; and performing a conversion between a video media file and a bitstream based on the filtered sample.
Partitioning Information In Neural Network-Based Video Coding
A method implemented by a video coding apparatus. The method includes applying a neural network (NN) filter to an unfiltered sample of a video unit to generate a filtered sample, where the NN filter includes an NN filter model generated based on partitioning information of the video unit; and performing a conversion between a video media file and a bitstream based on the filtered sample.
Video coding method on basis of secondary transform, and device for same
A video decoding method according to the present document is characterized by comprising: a step for deriving transform coefficients through inverse quantization on the basis of quantized transform coefficients for a target block; a step for deriving modified transform coefficients on the basis of an inverse reduced secondary transform (RST) of the transform coefficients; and a step for generating a reconstructed picture on the basis of residual samples for the target block on the basis of an inverse primary transform of the modified transform coefficients, wherein the inverse RST using a transform kernel matrix is performed on transform coefficients of the upper-left 4×4 region of an 8×8 region of the target block, and the modified transform coefficients of the upper-left 4×4 region, upper-right 4×4 region, and lower-left 4×4 region of the 8×8 region are derived through the inverse RST.
Video coding method on basis of secondary transform, and device for same
A video decoding method according to the present document is characterized by comprising: a step for deriving transform coefficients through inverse quantization on the basis of quantized transform coefficients for a target block; a step for deriving modified transform coefficients on the basis of an inverse reduced secondary transform (RST) of the transform coefficients; and a step for generating a reconstructed picture on the basis of residual samples for the target block on the basis of an inverse primary transform of the modified transform coefficients, wherein the inverse RST using a transform kernel matrix is performed on transform coefficients of the upper-left 4×4 region of an 8×8 region of the target block, and the modified transform coefficients of the upper-left 4×4 region, upper-right 4×4 region, and lower-left 4×4 region of the 8×8 region are derived through the inverse RST.
Using morphological operations to process frame masks in video content
A computer implemented method can decode a frame of video data comprising an array of pixels to obtain decoded luma values and decoded chroma values corresponding to the array of pixels, and extract a frame mask based on the decoded luma values. The frame mask can include an array of mask values respectively corresponding to the array of pixels. A mask value indicates whether a corresponding pixel is in foreground or background of the frame. The method can perform a morphological operation to the frame mask to change one or more mask values to indicate their corresponding pixels are removed from the foreground and added to the background of the frame. The method can also identify foreground pixels after performing the morphological operation to the frame mask, and render a foreground image for display based on the decoded luma values and decoded chroma values of the foreground pixels.
Using morphological operations to process frame masks in video content
A computer implemented method can decode a frame of video data comprising an array of pixels to obtain decoded luma values and decoded chroma values corresponding to the array of pixels, and extract a frame mask based on the decoded luma values. The frame mask can include an array of mask values respectively corresponding to the array of pixels. A mask value indicates whether a corresponding pixel is in foreground or background of the frame. The method can perform a morphological operation to the frame mask to change one or more mask values to indicate their corresponding pixels are removed from the foreground and added to the background of the frame. The method can also identify foreground pixels after performing the morphological operation to the frame mask, and render a foreground image for display based on the decoded luma values and decoded chroma values of the foreground pixels.
Techniques for decoding or coding images based on multiple intra-prediction modes
Aspects of the present disclosure provide techniques for derive one or more intra prediction modes (IPMs) from a subset of IPM candidates in order to determine a predictor to use for decoding a block of an image. In some aspects, the subset of IPM candidates may include IPMs that are less than the number of IPMs in a full set of all available IPM candidates (e.g., 67 IPMs in VVC or 35 in HEVC). In some aspects, the subset of IPM candidates may be based on a most probable mode (MPM) list that can be used to determine or signal an IPM based on IPMs previously used in decoding other blocks.
Techniques for decoding or coding images based on multiple intra-prediction modes
Aspects of the present disclosure provide techniques for derive one or more intra prediction modes (IPMs) from a subset of IPM candidates in order to determine a predictor to use for decoding a block of an image. In some aspects, the subset of IPM candidates may include IPMs that are less than the number of IPMs in a full set of all available IPM candidates (e.g., 67 IPMs in VVC or 35 in HEVC). In some aspects, the subset of IPM candidates may be based on a most probable mode (MPM) list that can be used to determine or signal an IPM based on IPMs previously used in decoding other blocks.
System and method for content-layer based video compression
Embodiments of the present invention disclose a method of content-layer based compression of a video being broadcasted over a network. The method may include: receiving a video stream comprising a plurality of video stream frames; identifying in at least some of the plurality of video stream frames at least two content-layers of predefined content-layers to yield corresponding at least two content-layer streams, wherein each of the at least two content-layer streams is associated with one of the at least two content-layers; and compressing each of the at least two content-layer video streams according to predetermined parameters of the content-layer associated with the respective content-layer video stream and according to available resources of the network to yield corresponding at least two compressed content-layer streams.
Method for alignment across layers in coded video stream
A method, computer program, and computer system is provided for aligning across layers in a coded video stream. A video bitstream having multiple layers is decoded. One or more subpicture regions are identified from among the multiple layers of the decoded video bitstream, the subpicture regions including a background region and one or more foreground subpicture regions. An enhanced subpicture is decoded and displayed based on a determination that a foreground subpicture region is selected. The background region is decoded and displayed based on a determination that a foreground subpicture region was not selected.