Patent classifications
H04N19/543
Polynomial fitting for motion compensation and luminance reconstruction in texture synthesis
The present disclosure relates to encoding a decoding video employing texture coding. In particular, a texture region is identified within a video picture and a texture patch is determined for said region. Moreover, a set of parameters specifies luminance within the texture region (1001) by fitting the texture region samples to a two-dimensional polynomial function of the patch determined according to the set of parameters (1040); and/or motion within the texture region by fitting motion estimated between the texture region of the video picture and an adjacent picture to a two-dimensional polynomial The texture patch and the first set of parameters are then included into a bitstream which is output of the encoder and provided in this way to the decoder which reconstructs the texture based on the patch and the function applied to the patch.
NON-TRANSITORY COMPUTER-READABLE STORAGE MEDIUM, ENCODING PROCESSING METHOD, AND ENCODING PROCESSING APPARATUS
A non-transitory computer-readable storage medium storing an encoding processing program that causes a computer to execute a process, the process including acquiring image feature information of each of a plurality of image regions obtained by partitioning first image data and image feature information of each of the plurality of image regions obtained by partitioning second image data, determining, for each of the plurality of image regions, whether image feature information of the first image data and image feature information of the second image data have a correlation greater than or equal to a certain value, and coding a specified image region of the second image data which has the correlation, the coding being performed by using coding unit used for coding of an image region of the first image data whose position corresponds to the specified image region of the second image data.
NON-TRANSITORY COMPUTER-READABLE STORAGE MEDIUM, ENCODING PROCESSING METHOD, AND ENCODING PROCESSING APPARATUS
A non-transitory computer-readable storage medium storing an encoding processing program that causes a computer to execute a process, the process including acquiring image feature information of each of a plurality of image regions obtained by partitioning first image data and image feature information of each of the plurality of image regions obtained by partitioning second image data, determining, for each of the plurality of image regions, whether image feature information of the first image data and image feature information of the second image data have a correlation greater than or equal to a certain value, and coding a specified image region of the second image data which has the correlation, the coding being performed by using coding unit used for coding of an image region of the first image data whose position corresponds to the specified image region of the second image data.
System and method for object based parametric video coding
A video compression framework based on parametric object and background compression is proposed. At the encoder, an embodiment detects objects and segments frames into regions corresponding to the foreground object and the background. The object and the background are individually encoded using separate parametric coding techniques. While the object is encoded using the projection of coefficients to the orthonormal basis of the learnt subspace (used for appearance based object tracking), the background is characterized using an auto-regressive (AR) process model. An advantage of the proposed schemes is that the decoder structure allows for simultaneous reconstruction of object and background, thus making it amenable to the new multi-thread/multi-processor architectures.
System and method for object based parametric video coding
A video compression framework based on parametric object and background compression is proposed. At the encoder, an embodiment detects objects and segments frames into regions corresponding to the foreground object and the background. The object and the background are individually encoded using separate parametric coding techniques. While the object is encoded using the projection of coefficients to the orthonormal basis of the learnt subspace (used for appearance based object tracking), the background is characterized using an auto-regressive (AR) process model. An advantage of the proposed schemes is that the decoder structure allows for simultaneous reconstruction of object and background, thus making it amenable to the new multi-thread/multi-processor architectures.
METHOD AND APPARATUS FOR COMMUNICATING AND RECOVERING MOTION INFORMATION
This invention describes a method for communicating crude motion information using tracking metadata and recovering more accurate motion information from the received tracking metadata and partial video frame data; in particular, we use metadata to convey crude boundaries of objects in the scene and signal motion information for these objects. The proposed method leaves the task of identifying the exact boundaries of an object to the decoder/client. The proposed method is particularly appealing when metadata itself carries semantics that the client is interested in, such as tracking information in surveillance applications, because, in this case, metadata does not constitute an overhead.
The proposed method involves motion descriptions that can be used to predict the appearance of an object in any one frame from its appearance in any other frame that contains the object. That is, the motion information itself allows locations within an object to be invertibly mapped to locations within the same object in any other relevant frame. This is a departure from conventional motion coding schemes, which tightly-couple motion information to the prediction strategy. This property makes the proposed method particularly suitable for applications which require flexible access to the content.
METHOD AND APPARATUS FOR COMMUNICATING AND RECOVERING MOTION INFORMATION
This invention describes a method for communicating crude motion information using tracking metadata and recovering more accurate motion information from the received tracking metadata and partial video frame data; in particular, we use metadata to convey crude boundaries of objects in the scene and signal motion information for these objects. The proposed method leaves the task of identifying the exact boundaries of an object to the decoder/client. The proposed method is particularly appealing when metadata itself carries semantics that the client is interested in, such as tracking information in surveillance applications, because, in this case, metadata does not constitute an overhead.
The proposed method involves motion descriptions that can be used to predict the appearance of an object in any one frame from its appearance in any other frame that contains the object. That is, the motion information itself allows locations within an object to be invertibly mapped to locations within the same object in any other relevant frame. This is a departure from conventional motion coding schemes, which tightly-couple motion information to the prediction strategy. This property makes the proposed method particularly suitable for applications which require flexible access to the content.
METHODS AND APPARATUSES FOR PREDICTION REFINEMENT WITH OPTICAL FLOW, BI-DIRECTIONAL OPTICAL FLOW, AND DECODER-SIDE MOTION VECTOR REFINEMENT
A bit-depth representation method, apparatus, and a non-transitory computer-readable storage medium are provided. A decoder obtains a reference picture I associated with a video block within the video signal. The decoder obtains prediction samples of the video block from a reference block in the reference picture I. The decoder controls internal PROF parameters of a PROF derivation process by applying right-shifting to the internal PROF parameters based on a bit-shift value to achieve a preset precision. The decoder obtains prediction refinement values for samples in the video block based on the PROF derivation process being applied to the video block based on the prediction samples. The decoder obtains prediction samples of the video block based on the combination of the prediction samples and the prediction refinement values.
MOTION COMPENSATION METHOD AND MODULE, CHIP, ELECTRONIC DEVICE AND STORAGE MEDIA
The present disclosure relates to a motion compensation method and module, a chip, an electronic device, and a storage medium, to improve the problem of haloes easily appearing on the edges of moving objects.
BIDIRECTIONAL OPTICAL FLOW BASED VIDEO CODING AND DECODING
Devices, systems and methods for sample refinement and filtering method for video coding are described. In an exemplary aspect, a method for video processing includes modifying, for a conversion between a block of a video and a bitstream representation of the video, a refinement value for a prediction sample in the block by applying a clipping operation to refinement value. The refinement value is derived based on a gradient value of an optical flow coding process. An output of the clipping operation is within a range. The method also includes refining the prediction sample based on the refinement value and performing the conversion based on the refined prediction sample.