H04N19/134

SYSTEMS AND METHODS FOR REDUCING BLOCKING ARTIFACTS
20230098413 · 2023-03-30 ·

Several systems, methods and integrated circuits capable of reducing blocking artifacts in video data are disclosed. In an embodiment, a system for reducing blocking artifacts in video data includes a processing module and a deblocking module. The deblocking module comprises a luma deblocking filter and a chroma deblocking filter configured to filter an edge between adjacent blocks associated with the video data, where a block of the adjacent blocks corresponds to one of a prediction block and a transform block. The processing module is communicatively associated with the deblocking module and is operable to configure at least one filter coefficient corresponding to the chroma deblocking filter based on one or more filter coefficients corresponding to the luma deblocking filter. The processing module is further configured to cause the chroma deblocking filter to filter the edge between the adjacent blocks based on the configured at least one filter coefficient.

Cascaded prediction-transform approach for mixed machine-human targeted video coding

Data may be encoded to minimize distortion after decoding, but the quality required for presentation of the decoded data to a machine and the quality required for presentation to a human may be different. To accommodate different quality requirements, video data may be encoded to produce a first set of encoded data and a second set of encoded data, where the first set may be decoded for use by one of a machine consumer or a human consumer, and a combination of the first set and the second set may be decoded for use by the other of a machine consumer or a human consumer. The first and second set may be produced with a neural encoder and a neural decoder, and/or may be produced with the use of prediction and transform neural network modules. A human-targeted structure and a machine-targeted structure may produce the sets of encoded data.

Cascaded prediction-transform approach for mixed machine-human targeted video coding

Data may be encoded to minimize distortion after decoding, but the quality required for presentation of the decoded data to a machine and the quality required for presentation to a human may be different. To accommodate different quality requirements, video data may be encoded to produce a first set of encoded data and a second set of encoded data, where the first set may be decoded for use by one of a machine consumer or a human consumer, and a combination of the first set and the second set may be decoded for use by the other of a machine consumer or a human consumer. The first and second set may be produced with a neural encoder and a neural decoder, and/or may be produced with the use of prediction and transform neural network modules. A human-targeted structure and a machine-targeted structure may produce the sets of encoded data.

Encoder and decoder

An encoder includes circuitry and memory coupled to the circuitry. The circuitry, in operation, determines whether a first block is available and whether a second block is available, the first block and the second block being defined relative to a current block to be processed; selects a context model based on whether the first block is available, whether the second block is available, which of inter prediction and intra prediction is to be applied to the first block, and which of inter prediction and intra prediction is to be applied to the second block; and encodes, using the context model selected, a parameter indicating which of intra prediction and inter prediction is to be applied to the current block.

Inference Processing of Data
20220351422 · 2022-11-03 ·

A method for processing data in a system configured to operate in either of at least a first power mode and a second power mode, wherein the first power mode is associated with a first power level and the second power mode is associated with a second power level, the second power level being higher than the first power level, wherein the first and second power modes each are configured to prepare a respective model for inference processing is disclosed. The method comprises acquiring (101) compressed data, determining (102) whether the system operates in the first power mode or in the second power mode. The method further comprises, when the system operates in the first power mode, determining (103) whether the acquired compressed data comprises a self-contained frame, and if so partly decoding (104) the self-contained frame, performing (105) feature extraction of the decoded self-contained frame, preparing (107) the model for inference processing in the first power mode in the system, wherein the model comprises inference parameters for the first power mode, and performing (108) inference processing by a neural network based on the extracted features and the prepared model for inference processing. Corresponding computer program product, apparatus, and system are also disclosed.

METHODS AND APPARATUS OF VIDEO CODING USING PALETTE MODE

An electronic apparatus performs a method of decoding video data. The method comprises: receiving, from bitstream, a plurality of syntax elements associated with a coding unit, wherein the plurality of syntax elements indicate a size of the coding unit and a coding tree type of the coding unit; determining a minimum palette mode block size for the coding unit in accordance with the coding tree type of the coding unit; in accordance with a determination that the size of the coding unit is greater than the minimum palette mode block size: receiving, from the bitstream, a palette mode enable flag associated with the coding unit; and decoding, from the bitstream, the coding unit in accordance with the palette mode enable flag.

Image encoder, image decoder, image encoding method, and image decoding method

An image encoder is provided including circuitry and a memory coupled to the circuitry. The circuitry, in operation, responds to a size of a block satisfying a size condition by generating a prediction image using a prediction mode selected from a plurality of prediction modes. The plurality of prediction modes include a first prediction mode in which a prediction process uses a motion vector and a reference block in a same picture as the block. The circuitry encodes the block using the prediction image.

Intra-prediction apparatus, encoding apparatus, decoding apparatus and methods for non-square video coding blocks

A device and method for intra-predicting a rectangular and non-square video coding block, the device including an intra-prediction apparatus configured for intra-predicting a rectangular and non-square video coding block according to a set of directional intra-prediction modes distributed over a set of directional ranges, where the intra-prediction apparatus being configured for configured for intra-predicting the rectangular and non-square video coding block includes the intra-prediction apparatus being configured to select a directional range from the set of directional ranges, and determine a number of directional intra-prediction modes of the selected directional range according to a size of a block.

Image data encoding/decoding method and apparatus

Disclosed is an image data encoding/decoding method and apparatus. A method for decoding a 360-degree image comprises the steps of: receiving a bitstream obtained by encoding a 360-degree image; generating a prediction image by making reference to syntax information obtained from the received bitstream; combining the generated prediction image with a residual image obtained by dequantizing and inverse-transforming the bitstream, so as to obtain a decoded image; and reconstructing the decoded image into a 360-degree image according to a projection format.

Image data encoding/decoding method and apparatus

Disclosed is an image data encoding/decoding method and apparatus. A method for decoding a 360-degree image comprises the steps of: receiving a bitstream obtained by encoding a 360-degree image; generating a prediction image by making reference to syntax information obtained from the received bitstream; combining the generated prediction image with a residual image obtained by dequantizing and inverse-transforming the bitstream, so as to obtain a decoded image; and reconstructing the decoded image into a 360-degree image according to a projection format.