Patent classifications
H04N19/89
METHOD, DEVICE, STORAGE MEDIUM, AND COMPUTER PROGRAM PRODUCT FOR DETECTING IMAGE FRAME LOSS
An image frame loss detection method is performed by a computer device, including: acquiring first coded data respectively corresponding to a plurality of first image frames and a color signal corresponding to at least one second image frame; obtaining second coded data corresponding to at least one second image frame generated by a terminal device through image rendering of a color signal based on the coded data respectively corresponding to the plurality of first image frames; and comparing the first coded data respectively corresponding to the plurality of first image frames with the second coded data corresponding to the at least one second image frame to determine whether a frame loss occurs. The first coded data and the second coded data each include color-coded data respectively corresponding to M image blocks of a correspond image frame, and each of the M image blocks has a color in the image frame.
SYNTAX STRUCTURES INDICATING COMPLETION OF CODED REGIONS
Syntax structures that indicate the completion of coded regions of pictures are described. For example, a syntax structure in an elementary bitstream indicates the completion of a coded region of a picture. The syntax structure can be a type of network abstraction layer unit, a type of supplemental enhancement information message or another syntax structure. For example, a media processing tool such as an encoder can detect completion of a coded region of a picture, then output, in a predefined order in an elementary bitstream, syntax structure(s) that contain the coded region as well as a different syntax structure that indicates the completion of the coded region. Another media processing tool such as a decoder can receive, in a predefined order in an elementary bitstream, syntax structure(s) that contain a coded region of a picture as well as a different syntax structure that indicates the completion of the coded region.
DIRECTED INTERPOLATION AND DATA POST-PROCESSING
An encoding device evaluates a plurality of processing and/or post-processing algorithms and/or methods to be applied to a video stream, and signals a selected method, algorithm, class or category of methods/algorithms either in an encoded bitstream or as side information related to the encoded bitstream. A decoding device or post-processor utilizes the signaled algorithm or selects an algorithm/method based on the signaled method or algorithm. The selection is based, for example, on availability of the algorithm/method at the decoder/post-processor and/or cost of implementation. The video stream may comprise, for example, downsampled multiplexed stereoscopic images and the selected algorithm may include any of upconversion and/or error correction techniques that contribute to a restoration of the downsampled images.
IMAGE ENCODING DEVICE, IMAGE ENCODING METHOD AND STORAGE MEDIUM, IMAGE DECODING DEVICE, AND IMAGE DECODING METHOD AND STORAGE MEDIUM
An image encoding device includes a prediction unit configured to generate prediction errors being a difference between a predicted image obtained by prediction processing for an input image and the input image, a first transform unit configured to generate first transform coefficients by performing orthogonal transform on the prediction errors, a second transform unit configured to generate second transform coefficients by performing LFNST processing on the first transform coefficients, a quantization unit configured to generate quantization coefficients by performing quantization processing on the second transform coefficients, and an encoding unit configured to encode the quantization coefficients, wherein the encoding unit encodes information indicating whether a range of possible values at least taken by the second transform coefficients is to be a range determined based on a bit depth or a fixed range.
IMAGE ENCODING DEVICE, IMAGE ENCODING METHOD AND STORAGE MEDIUM, IMAGE DECODING DEVICE, AND IMAGE DECODING METHOD AND STORAGE MEDIUM
An image encoding device includes a prediction unit configured to generate prediction errors being a difference between a predicted image obtained by prediction processing for an input image and the input image, a first transform unit configured to generate first transform coefficients by performing orthogonal transform on the prediction errors, a second transform unit configured to generate second transform coefficients by performing LFNST processing on the first transform coefficients, a quantization unit configured to generate quantization coefficients by performing quantization processing on the second transform coefficients, and an encoding unit configured to encode the quantization coefficients, wherein the encoding unit encodes information indicating whether a range of possible values at least taken by the second transform coefficients is to be a range determined based on a bit depth or a fixed range.
Opportunistic progressive encoding
Methods, systems, and devices are described for communicating data from multiple data terminals to an aggregator terminal over a communication link having changing link conditions. In some embodiments, source data is received at multiple data terminals, each in communication with an aggregator terminal over a communication link. For example, during a live newscast, one mobile camera may receive live video of an event from a first position while another mobile camera receives live video of the event from a second position. For various reasons (e.g., as the cameras move) each communication link may experience independently changing link conditions. Each data terminal encodes the source data (or store source data for later encoding) as a function of its respective link conditions and transmits encoded source data over its respective communication link to the aggregator terminal.
Opportunistic progressive encoding
Methods, systems, and devices are described for communicating data from multiple data terminals to an aggregator terminal over a communication link having changing link conditions. In some embodiments, source data is received at multiple data terminals, each in communication with an aggregator terminal over a communication link. For example, during a live newscast, one mobile camera may receive live video of an event from a first position while another mobile camera receives live video of the event from a second position. For various reasons (e.g., as the cameras move) each communication link may experience independently changing link conditions. Each data terminal encodes the source data (or store source data for later encoding) as a function of its respective link conditions and transmits encoded source data over its respective communication link to the aggregator terminal.
VIDEO DECODING IMPLEMENTATIONS FOR A GRAPHICS PROCESSING UNIT
Video decoding innovations for multithreading implementations and graphics processor unit (“GPU”) implementations are described. For example, for multithreaded decoding, a decoder uses innovations in the areas of layered data structures, picture extent discovery, a picture command queue, and/or task scheduling for multithreading. Or, for a GPU implementation, a decoder uses innovations in the areas of inverse transforms, inverse quantization, fractional interpolation, intra prediction using waves, loop filtering using waves, memory usage and/or performance-adaptive loop filtering. Innovations are also described in the areas of error handling and recovery, determination of neighbor availability for operations such as context modeling and intra prediction, CABAC decoding, computation of collocated information for direct mode macroblocks in B slices, reduction of memory consumption, implementation of trick play modes, and picture dropping for quality adjustment.
VIDEO DECODING IMPLEMENTATIONS FOR A GRAPHICS PROCESSING UNIT
Video decoding innovations for multithreading implementations and graphics processor unit (“GPU”) implementations are described. For example, for multithreaded decoding, a decoder uses innovations in the areas of layered data structures, picture extent discovery, a picture command queue, and/or task scheduling for multithreading. Or, for a GPU implementation, a decoder uses innovations in the areas of inverse transforms, inverse quantization, fractional interpolation, intra prediction using waves, loop filtering using waves, memory usage and/or performance-adaptive loop filtering. Innovations are also described in the areas of error handling and recovery, determination of neighbor availability for operations such as context modeling and intra prediction, CABAC decoding, computation of collocated information for direct mode macroblocks in B slices, reduction of memory consumption, implementation of trick play modes, and picture dropping for quality adjustment.
Guaranteed Data Compression
A method of converting 10-bit pixel data (e.g. 10:10:10:2 data) into 8-bit pixel data involves converting the 10-bit values to 7-bits or 8-bits and generating error values for each of the converted values. Two of the 8-bit output channels comprise a combination of a converted 7-bit value and one of the bits from the fourth input channel. A third 8-bit output channel comprises the converted 8-bit value and the fourth 8-bit output channel comprises the error values. In various examples, the bits of the error values may be interleaved when they are packed into the fourth output channel.