H04N11/04

Motion estimation apparatus and method for multiview video
09832479 · 2017-11-28 · ·

A motion estimation apparatus and method (carried out electronically) provides for encoding of multiview video, such as stereoscopic video, by providing motion estimation for pixels in a dependent eye view, using motion vector information from a colocated group of pixels in a base eye view and neighboring pixels to the colocated group of pixels in the base eye view. The method and apparatus encodes a group of pixels in a dependent eye view based on the estimated motion vector information. The method and apparatus may also include obtaining a frame of pixels that includes both base eye view pixels and dependent eye pixels so that, for example, frame compatible format packing can be employed. In one example, estimating the motion vector information for a block of pixels, for example, in a dependent eye view is based on a median value calculation of motion vectors for a block of pixels in a base eye view and motion vectors for neighboring blocks of pixels to the colocated group of pixels in the base eye view. An apparatus and method may include transmitting the encoded dependent eye view and base eye view information to another device and decoding the encoded dependent eye view and base eye view information for display.

Palette-based video coding

In an example, a method of coding video data includes determining, by a video coder and for a block of video data, a palette having a plurality of entries indicating a plurality of respective color values, wherein a first line of the block of video data includes a pixel located adjacent to an edge of the block of video data, and wherein a second line of the block of video data includes a pixel located adjacent to the edge of the block and adjacent to the pixel of the first line. In this example, the method also includes coding, in a scan order, index values that map pixels of the block to entries in the palette, wherein the pixel of the second line immediately follows the pixel of the first line in the scan order.

Video or image coding based on signaling of scaling list data
11503343 · 2022-11-15 · ·

According to the disclosure of the present document, scaling list data delivered in an adaptation parameter set (APS) may be signaled through a hierarchical structure, and the amount of data that needs to be signaled for video/image coding may be reduced and implementation may be facilitated by placing limits on the scaling list data delivered in the APS.

Video content analysis and/or processing using encoding logs
09813706 · 2017-11-07 · ·

A system to facilitate analyzing and/or processing video content using encoding logs is provided. The system includes a first analysis component, a second analysis component and a transcoding component. The first analysis component determines a first set of shot-change locations associated with a cut transition in a video sequence based on an encoding log associated with at least one previously encoded video frame. The second analysis component determines a second set of shot-change locations associated with a fading transition in the video sequence based on the encoding log associated with the at least one previously encoded video frame. The transcoding component processes the video sequence based on at least a portion of the first set of shot-change locations and the second set of shot-change locations.

General Constraints Information for Video Coding
20220353495 · 2022-11-03 ·

A method for implementing an adaptive color transform (ACT) during image/video encoding and decoding, comprises determining, for a conversion between a video comprising a block and a bitstream of the video, that a size of the block is greater than a maximum allowed size for an ACT mode, and performing, based on the determining, the conversion, wherein, in response to the size of the block being greater than the maximum allowed size for the ACT mode, the block is partitioned into multiple sub-blocks, and wherein each of the multiple sub-blocks share a same prediction mode, and the ACT mode is enabled at a sub-block level.

Sample adaptive offset (SAO) adjustment method and apparatus and SAO adjustment determination method and apparatus

A video encoding method and apparatus, and a video decoding method and apparatus for generating a reconstructed image having a minimized error between an original image and the reconstructed image. The video decoding method accompanied by a sample adaptive offset (SAO) adjustment, the method includes: obtaining 5 slice SAO parameters with respect to a current slice from a slice header of a received bitstream; obtaining luma SAO use information for a luma component of the current slice and chroma SAO use information for chroma components thereof from among the slice SAO parameters; determining whether to perform a SAO operation on the luma component of 10 the current slice based on the obtained luma SAO use information; and equally determining whether to perform the SAO adjustment on a first chroma component and a second chroma component of the current slice based on the obtained chroma SAO use information.

Fine-grained bit-rate control

A system implements rate control for encoding and decoding operations, for example, operations performed on slices of data such as image data. The system implements fine-grained bit rate control allowing for non-integer bit rates to be specified for the system. The non-integer values may allow the system to more accurately characterize a data rate of a communication link between a source and sink. The more accurate characterization may facilitate improved utilization of the communication link capacity.

Picture flushing and decoded picture buffer parameter inference for multi-layer bitstreams

An apparatus according to certain aspects includes a memory and a processor configured to: in response to determining that a reference layer intra random access point (IRAP) picture starts a new coded video sequence (CVS) at the current AU: determine whether the reference layer IRAP picture activates a new video parameter set (VPS) or defines a prediction boundary at the current AU; in response to determining that the reference layer IRAP picture activates a new VPS or defines a prediction boundary at the current AU, remove pictures in the decoded picture buffer (DPB) associated with each of a plurality of layers; and in response to determining that the reference layer IRAP picture does not activate a new VPS and does not define a prediction boundary at the current AU, remove only pictures in the DPB associated with the reference layer.

Image coding apparatus, image decoding apparatus, image processing system, image coding method, and image decoding method

An image coding apparatus is an image coding apparatus which encodes a plurality of to-be-displayed images forming video by using inter prediction. The image coding apparatus includes an obtaining unit that obtains a reference-specific image which is an image different from the plurality of to-be-displayed images or different from a plurality of reconfigured images obtained by reconfiguration of the plurality of to-be-displayed images and which is an image used as an image that is specific to reference for the inter prediction, and an encoding unit that encodes one or more to-be-displayed images among the plurality of to-be-displayed images by referring to the reference-specific image as a reference image for the inter prediction.

Picture header presence

A method for decoding a picture from a bitstream. In one embodiment, The method includes: receiving a slice header for a slice of the picture, wherein the slice header comprises a state syntax element; decoding a state value from the state syntax element in the slice header, wherein a) if the state value is not equal to a first value, then the state value indicates that i) the bitstream includes for the picture a picture header comprising a set of picture syntax elements and ii) the slice header does not comprise the set of picture syntax elements and b) if the state value is equal to the first value, then the state value indicates that i) the slice header contains said set of picture syntax elements and ii) the bitstream does not include a picture header for the picture; and using the set of picture syntax elements to decode the slice of the picture.