H04N19/587

Reference subpicture scaling ratios for subpictures in video coding

A video decoder can be configured to determine that a first subpicture of a current picture has associated scaling parameters; receive the associated scaling parameters for the first subpicture of the current picture in response to determining that the first subpicture of the current picture has the associated scaling parameters; determine motion information, for a block of the first subpicture of the current picture, that identifies a subpicture of a reference picture; locate a prediction block for the block of the first subpicture of the current picture in the subpicture of the reference picture; and scale the prediction block based on the associated scaling parameters for the first subpicture of the current picture.

MACHINE LEARNING BASED VIDEO COMPRESSION

Systems and methods are disclosed for compressing a target video. A computer-implemented method may use a computer system that include one or more physical computer processors and non-transient electronic storage. The computer-implemented method may include: obtaining the target video, extracting one or more frames from the target video, and generating an estimated optical flow based on a displacement of pixels between the one or more frames. The one or more frames may include one or more of a key frame and a target frame.

MACHINE LEARNING BASED VIDEO COMPRESSION

Systems and methods are disclosed for compressing a target video. A computer-implemented method may use a computer system that include one or more physical computer processors and non-transient electronic storage. The computer-implemented method may include: obtaining the target video, extracting one or more frames from the target video, and generating an estimated optical flow based on a displacement of pixels between the one or more frames. The one or more frames may include one or more of a key frame and a target frame.

Efficient Motion-Compensated Spatiotemporal Sampling

In one embodiment, a computing system may access a video including a first frame and a second frame. The computing system may determine first sampling locations for the first frame and determine second sampling locations for the second frame by transforming the first sampling locations to the second frame according to an optical flow between the first frame and the second frame. The computing system may detect one or more invalid second sampling locations based on determining pixels in the first frame corresponding to the first sampling locations do not match pixels in the second frame corresponding to the second sampling locations. The computing system may reject the one or more invalid second sampling locations to determine third sampling locations for the second frame. The computing system may generate a sample of the video.

Directed interpolation and data post-processing

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images.

Directed interpolation and data post-processing

An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images.

Scalable video coding using derivation of subblock subdivision for prediction from base layer

Scalable video coding is rendered more efficient by deriving/selecting a subblock subdivision to be used for enhancement layer prediction, among a set of possible subblock subdivisions of an enhancement layer block by evaluating the spatial variation of the base layer coding parameters over the base layer signal. By this measure, less of the signalization overhead has to be spent on signaling this subblock subdivision within the enhancement layer data stream, if any. The subblock subdivision thus selected may be used in predictively coding/decoding the enhancement layer signal.

Scalable video coding using derivation of subblock subdivision for prediction from base layer

Scalable video coding is rendered more efficient by deriving/selecting a subblock subdivision to be used for enhancement layer prediction, among a set of possible subblock subdivisions of an enhancement layer block by evaluating the spatial variation of the base layer coding parameters over the base layer signal. By this measure, less of the signalization overhead has to be spent on signaling this subblock subdivision within the enhancement layer data stream, if any. The subblock subdivision thus selected may be used in predictively coding/decoding the enhancement layer signal.

Image processing method and image processing device
11636573 · 2023-04-25 · ·

An image processing method includes: downsizing a current frame and a reference frame; dividing the down-sized current frame and the down-sized reference frame into multiple first current blocks and multiple first reference blocks, respectively; performing a first motion estimation to the first current blocks and the first reference blocks to generate multiple first motion vectors; dividing the current picture and the reference picture into multiple second current blocks and multiple second reference blocks, respectively; performing a second motion estimation to the second current blocks and the second reference blocks to generate multiple second motion vectors; and generating a compensated frame between the current frame and the reference frame according to the second motion vectors. The second motion estimation includes: performing a 3D recursive search for each second current block; and adjusting multiple estimation parameters in the 3D recursive search according to the first motion vector.