H04N19/85

Per-Title Encoding Using Spatial and Temporal Resolution Downscaling
20230007298 · 2023-01-05 · ·

Techniques relating to per-title encoding using spatial and temporal resolution downscaling is disclosed. A method for per-title encoding includes receiving a video input comprised of video segments, spatially downscaling the video input, temporally downscaling the video input, encoding the video input to generate an encoded video, then temporally and spatially upscaling the encoded video. Spatially downscaling may include reducing a resolution of the video input, and temporally downscaling may include reducing a framerate of the video input. Objective metrics for the upscaled encoded video show improved quality over conventional methods.

Data stream encoder configuration

A media encoder for encoding a stream of media data blocks has an encoder pipeline including a sequence of processing modules for processing a stream of media data blocks, and a pipeline configurator configured effect a switch in the encoder pipeline from one or more first encode parameters to one or more second encode parameters. The first processing module of the pipeline can be configured to associate a trigger value with at least a first media data block processed at the first processing module in accordance with second encode parameters, the trigger value passing to subsequent modules so as to cause those modules to adopt the second encode parameters.

Data stream encoder configuration

A media encoder for encoding a stream of media data blocks has an encoder pipeline including a sequence of processing modules for processing a stream of media data blocks, and a pipeline configurator configured effect a switch in the encoder pipeline from one or more first encode parameters to one or more second encode parameters. The first processing module of the pipeline can be configured to associate a trigger value with at least a first media data block processed at the first processing module in accordance with second encode parameters, the trigger value passing to subsequent modules so as to cause those modules to adopt the second encode parameters.

Method and device for generating a second image from a first image

Described are methods and devices for applying a color gamut mapping process on a first image to generate a second image, where the content of the first and second images is similar but the respective color spaces of the first and second images are different. The color gamut mapping process may be controlled using a color gamut mapping mode obtained from a bitstream where the color gamut mapping mode belongs to a set comprising at least two preset modes and an explicit parameters mode. If the obtained color gamut mapping mode is the explicit parameters mode and the color gamut mapping process is not enabled for the explicit parameters mode, the color gamut mapping process may be controlled by a substitute color gamut mapping mode determined from additional data.

Method and device for generating a second image from a first image

Described are methods and devices for applying a color gamut mapping process on a first image to generate a second image, where the content of the first and second images is similar but the respective color spaces of the first and second images are different. The color gamut mapping process may be controlled using a color gamut mapping mode obtained from a bitstream where the color gamut mapping mode belongs to a set comprising at least two preset modes and an explicit parameters mode. If the obtained color gamut mapping mode is the explicit parameters mode and the color gamut mapping process is not enabled for the explicit parameters mode, the color gamut mapping process may be controlled by a substitute color gamut mapping mode determined from additional data.

Image processing apparatus, image processing method and image processing program

An image processing device that updates a pixel value of a processing target image and generates a new image generates a first feature vector based on the processing target image and a first feature map generated with at least one pre-decided filter; updates the processing target image to generate an updated image; generates a second feature vector based on the updated image and a second feature map generated with at least one pre-decided filter; performs quality evaluation of the updated image based on the first and second feature vectors and generates a quality feedback vector which is a vector based on a result of the quality evaluation; performs an encoding amount evaluation on the updated image and generates an encoding amount feedback vector which is a vector based on a result of the encoding amount evaluation; and determines an updating amount in updating of the updated image based on the quality feedback vector and the encoding amount feedback vector.

Method for optimizing structure similarity index in video coding
11546607 · 2023-01-03 · ·

The present disclosure provides a computer-implemented method for encoding video. The method includes: generating training data based on one or more video sequences, the training data including a structure similarity index comprising at least one of structure similarity index (SSIM) or multi-scale-structural similarity index (MS-SSIM); training a rate-distortion optimization (RDO) model using the training data; processing the one or more video sequences using the rate-distortion optimization model.

Method for optimizing structure similarity index in video coding
11546607 · 2023-01-03 · ·

The present disclosure provides a computer-implemented method for encoding video. The method includes: generating training data based on one or more video sequences, the training data including a structure similarity index comprising at least one of structure similarity index (SSIM) or multi-scale-structural similarity index (MS-SSIM); training a rate-distortion optimization (RDO) model using the training data; processing the one or more video sequences using the rate-distortion optimization model.

EMBEDDING FRAME MASKS IN A VIDEO STREAM

Techniques are described for efficiently embedding frame masks in a video stream. In some solutions, a computer implemented method includes operations for encoding a frame of video data comprising an array of pixels to generate an encoded video frame and transmitting the encoded video frame. The array of pixels can include foreground pixels and background pixels. The foreground pixels can have respective original luma values which are bounded within a first luma range. In certain examples, encoding the frame of video data can include converting the original luma values of the foreground pixels to updated luma values which are bounded within a second luma range. The second luma range can be shifted and/or compressed from the first luma range.

Techniques for optimizing encoding tasks
11539966 · 2022-12-27 · ·

In various embodiments, a shot collation application causes multiple encoding instances to encode a source video sequence that includes at least two shot sequences. The shot collation application assigns a first shot sequence to a first chunk. Subsequently, the shot collation application determines that a second shot sequence does not meet a collation criterion with respect to the first chunk. Consequently, the shot collation application assigns the second shot sequence or a third shot sequence derived from the second shot sequence to a second chunk. The shot collation application causes a first encoding instance to independently encode each shot sequence assigned to the first chunk. Similarly, the shot collation application causes a second encoding instance to independently encode each shot sequence assigned to the second chunk. Finally, a chunk assembler combines the first encoded chunk and the second encoded chunk to generate an encoded video sequence.