Patent classifications
H04N19/70
Method and Apparatus for Entropy Coding of Source Samples with Large Alphabet
A general entropy coding method for source symbols is disclosed. This method determines a prefix part and any suffix part for the current symbol. The method divides prefix of the source symbol into at least two parts by comparing a test value related to the prefix part against a threshold. If the test value is greater than or equal to the threshold, the method derives a first binary string by binarizing a first prefix part related to the prefix part using a first variable length code. If the test value related to the prefix part is less than the threshold, the method derives a second binary string by binarizing a second prefix part related to the prefix part using a second variable length code or a first fixed-length code. The method then encodes at least one of the first binary string and the second binary string using a CABAC mode.
TRANSMITTING DEVICE, TRANSMITTING METHOD, RECEIVING DEVICE, AND RECEIVING METHOD
The present invention enables a receiving side to easily recognize a high-quality format corresponding to encoded image data included in an extended video stream. Two video streams including a basic video stream including encoded image data of basic format image data, and an extended video stream including encoded image data of high-quality format image data of one type selected from a plurality of types are generated. A container of a predetermined format including the basic video stream and the extended video stream is transmitted. Information indicating a high-quality format corresponding to the encoded image data included in the extended video stream is inserted into the extended video stream and/or the container.
ADAPTIVE TILE DATA SIZE CODING FOR VIDEO AND IMAGE COMPRESSION
A method for encoding a video signal includes estimating a space requirement for encoding a tile of a video frame, writing a first value in a first value space of the bitstream, wherein the first value describes a size of a second value space, and defining the second value space in the bitstream, wherein the size of the second value space is based on an estimated space requirement. The method also includes writing encoded content in a content space of the bitstream, determining a size of the content space subsequent to writing encoded content in the content space, and writing a second value in the second value space of the bitstream, wherein the second value describes the size of the content space.
IMAGE PREDICTION ENCODING/DECODING SYSTEM
An encoding target region in an image can be partitioned into a plurality of prediction regions. Based on prediction information of a neighboring region neighboring a target region, the number of previously-encoded prediction regions in the target region, and previously-encoded prediction information of the target region, a candidate for motion information to be used in generation of a predicted signal of the target prediction region as a next prediction region is selected from previously-encoded motion information of regions neighboring the target prediction region. According to the number of candidates for motion information selected, merging block information to indicate generation of the predicted signal of the target prediction region using the selected candidate for motion information and motion information detected by prediction information estimation means, or either one of the merging block information or the motion information is encoded.
ENCODING, DECODING, AND REPRESENTING HIGH DYNAMIC RANGE IMAGES
Techniques are provided to encode and decode image data comprising a tone mapped (TM) image with HDR reconstruction data in the form of luminance ratios and color residual values. In an example embodiment, luminance ratio values and residual values in color channels of a color space are generated on an individual pixel basis based on a high dynamic range (HDR) image and a derivative tone-mapped (TM) image that comprises one or more color alterations that would not be recoverable from the TM image with a luminance ratio image. The TM image with HDR reconstruction data derived from the luminance ratio values and the color-channel residual values may be outputted in an image file to a downstream device, for example, for decoding, rendering, and/or storing. The image file may be decoded to generate a restored HDR image free of the color alterations.
SIGNALING COLOR VALUES FOR 3D LOOKUP TABLE FOR COLOR GAMUT SCALABILITY IN MULTI-LAYER VIDEO CODING
Techniques are described for signaling information used to generate three-dimensional (3D) color lookup tables for color gamut scalability in multi-layer video coding. A lower layer of video data may include color data in a first color gamut and a higher layer of the video data may include color data in a second color gamut. To generate inter-layer reference pictures, a video encoder and/or video decoder performs color prediction to convert the color data of a reference picture in the first color gamut to the second color gamut. The video coder may perform color prediction using a 3D lookup table. According to the techniques, a video encoder may encode partition information and/or color values of a 3D lookup table generated for color gamut scalability. A video decoder may decode the partition information and/or color values to generate the 3D lookup table in order to perform color gamut scalability.
SIGNALING COLOR VALUES FOR 3D LOOKUP TABLE FOR COLOR GAMUT SCALABILITY IN MULTI-LAYER VIDEO CODING
Techniques are described for signaling information used to generate three-dimensional (3D) color lookup tables for color gamut scalability in multi-layer video coding. A lower layer of video data may include color data in a first color gamut and a higher layer of the video data may include color data in a second color gamut. To generate inter-layer reference pictures, a video encoder and/or video decoder performs color prediction to convert the color data of a reference picture in the first color gamut to the second color gamut. The video coder may perform color prediction using a 3D lookup table. According to the techniques, a video encoder may encode partition information and/or color values of a 3D lookup table generated for color gamut scalability. A video decoder may decode the partition information and/or color values to generate the 3D lookup table in order to perform color gamut scalability.
METHOD FOR DEPTH LOOKUP TABLE SIGNALING
A method and apparatus for depth lookup table (DLT) signaling in a three-dimensional and multi-view coding system. The method identifies one or more pictures to be processed. If one or more pictures contain depth data, then the method determines the DLT associated with said one or more pictures, applies predictive coding to the DLT based on the previous DLT, includes syntax related to the DLT in the PPS, and includes first bit-depth information related to first depth samples of the DLT in the PPS. The first bit-depth information is consistent with second bit depth information signaled in a sequence level. The method further signals the PPS in a video bitstream for a sequence including said one or more pictures. A circuit is also provided that embodies circuitry configured to carry out the operations specified above.
METHOD FOR DEPTH LOOKUP TABLE SIGNALING
A method and apparatus for depth lookup table (DLT) signaling in a three-dimensional and multi-view coding system. The method identifies one or more pictures to be processed. If one or more pictures contain depth data, then the method determines the DLT associated with said one or more pictures, applies predictive coding to the DLT based on the previous DLT, includes syntax related to the DLT in the PPS, and includes first bit-depth information related to first depth samples of the DLT in the PPS. The first bit-depth information is consistent with second bit depth information signaled in a sequence level. The method further signals the PPS in a video bitstream for a sequence including said one or more pictures. A circuit is also provided that embodies circuitry configured to carry out the operations specified above.
SYNTAX STRUCTURES INDICATING COMPLETION OF CODED REGIONS
Syntax structures that indicate the completion of coded regions of pictures are described. For example, a syntax structure in an elementary bitstream indicates the completion of a coded region of a picture. The syntax structure can be a type of network abstraction layer unit, a type of supplemental enhancement information message or another syntax structure. For example, a media processing tool such as an encoder can detect completion of a coded region of a picture, then output, in a predefined order in an elementary bitstream, syntax structure(s) that contain the coded region as well as a different syntax structure that indicates the completion of the coded region. Another media processing tool such as a decoder can receive, in a predefined order in an elementary bitstream, syntax structure(s) that contain a coded region of a picture as well as a different syntax structure that indicates the completion of the coded region.