Patent classifications
H04N19/102
Moving picture coding apparatus, moving picture decoding apparatus, and stream data
A signal separation unit (101) separates an input picture made up of component pictures of RGB, each of which has an equal number of pixels as the input picture, into three component pictures, and outputs the three component pictures. Each of coding units (102 to 104) codes one of the component pictures into an intra-picture prediction coded picture or an inter-picture prediction coded picture, and outputs a bit stream corresponding to the component picture. A bit stream multiplexing unit (105) multiplexes three bit streams outputted from the three coding units into one bit stream, and outputs the bit stream. Each of the coding units (102), (103) and (104) determines a prediction method for the component picture at the time of coding.
METHOD AND APPARATUS FOR GENERATING LOW BIT WIDTH HDR IMAGE, STORAGE MEDIUM, AND TERMINAL
A method and an apparatus for generating HDR image with low bit width, a storage medium, and a terminal are provided. The method includes: determining a to-be-processed HDR image with high bit width; splitting the to-be-processed HDR image with high bit width to obtain N frames of split images which include a first frame to an Nth frame; performing bit width reduction processing on the N frames of split images respectively, to obtain N frames of low-bit-width images; and synthesizing the N frames of low-bit-width images to obtain the HDR image with low bit width.
Point Cloud Encoding and Decoding Method and Apparatus, Computer-Readable Medium, and Electronic Device
The present subject matter relates to point cloud encoding and decoding methods and apparatuses. A point cloud file transmitted by a data source is received. The point cloud file comprises point cloud media tracks having the same point cloud content and point cloud media tracks having different frame rates. File encapsulation information of the point cloud media tracks are parsed to obtain frame rate indication information carried in the file encapsulation information. The frame rate indication information indicates frame rates of the point cloud media tracks. A point cloud media track with a designated frame rate is selected and decoded from the point cloud file according to the frame rate indication information carried in the file encapsulation information.
Methods and apparatuses for performing artificial intelligence encoding and artificial intelligence decoding on image
Provided is an artificial intelligence (AI) decoding apparatus includes: a memory storing one or more instructions; and a processor configured to execute the one or more instructions stored in the memory, the processor is configured to: obtain AI data related to AI down-scaling an original image to a first image; obtain image data corresponding to an encoding result on the first image; obtain a second image corresponding to the first image by performing a decoding on the image data; obtain deep neural network (DNN) setting information among a plurality of DNN setting information from the AI data; and obtain, by an up-scaling DNN, a third image by performing the AI up-scaling on the second image, the up-scaling DNN being configured with the obtained DNN setting information, wherein the plurality of DNN setting information comprises a parameter used in the up-scaling DNN, the parameter being obtained through joint training of the up-scaling DNN and a down-scaling DNN, and wherein the down-scaling DNN is used to obtain the first image from the original image.
SIGNALLING OF GENERAL CONSTRAINS FLAG
Methods, systems, apparatus for video processing are described. The processing may include encoding, decoding or transcoding. One example video processing method includes performing a conversion between a video and a bitstream of the video according to a rule, and wherein the rule specifies that a syntax element is included in a profile, tier, level information syntax structure to indicate whether one or more general constraint information syntax elements are included in a general constraint information syntax structure and/or whether the general constraint information syntax structure is included in the profile, tier, level information syntax structure.
Transport controlled video coding
Embodiments are generally directed to transport controlled video coding. An embodiment of an apparatus includes one or more processors to process data; a memory to store data, including data for video streaming; and a video processing mechanism including an encoder and a transport mechanism, wherein the video processing mechanism is to generate a prediction of channel throughput for a network channel, encode one or more bitstreams based on the prediction, including encoding a plurality of bitstreams including a first bitstream and a second bitstream if the prediction indicates an increase or decrease in channel throughput and encoding a single bitstream if the prediction indicates a stable channel throughput; and select a bitstream of the one or more bitstreams for a current frame.
Transport controlled video coding
Embodiments are generally directed to transport controlled video coding. An embodiment of an apparatus includes one or more processors to process data; a memory to store data, including data for video streaming; and a video processing mechanism including an encoder and a transport mechanism, wherein the video processing mechanism is to generate a prediction of channel throughput for a network channel, encode one or more bitstreams based on the prediction, including encoding a plurality of bitstreams including a first bitstream and a second bitstream if the prediction indicates an increase or decrease in channel throughput and encoding a single bitstream if the prediction indicates a stable channel throughput; and select a bitstream of the one or more bitstreams for a current frame.
Scene aware video content encoding
Scene aware video content encoding techniques can determine if video content is a given content type and is one of one or more given titles that include one or more given scenes. The one or more given scenes of the video content of the given type and given one of the titles can be encoded using corresponding scenes specific encoding parameter values, and the non-given scenes can be encoded using one or more general encoding parameter values. The one or more given titles can be selected based on a rate of streaming of various video content titles of the given type.
IMAGE ENCODING/DECODING METHOD AND DEVICE FOR SIGNALING DPB-RELATED INFORMATION AND PTL-RELATED INFORMATION, AND COMPUTER-READABLE RECORDING MEDIUM IN WHICH BITSTREAM IS STORED
An image encoding/decoding method and apparatus for signaling decoded picture buffer (DPB) related information and profile tier level (PTL) related information and a method of transmitting a bitstream are provided. The image decoding method according to the present disclosure may comprise obtaining first information indicating the number of one or more decoded picture buffer (DPB) parameter syntax structures in a video parameter set (VPS), obtaining the one or more DPB parameter syntax structures from the VPS based on the first information, obtaining second information between one or more multi-layer output layer sets (OLSs) and the one or more DPB parameter syntax structures from the VPS based on the first information, selecting a DPB parameter syntax structure that applies to a current OLS based on the second information, and processing the current OLS based on the selected DPB parameter syntax structure.
IMAGE ENCODING/DECODING METHOD AND DEVICE FOR SIGNALING DPB-RELATED INFORMATION AND PTL-RELATED INFORMATION, AND COMPUTER-READABLE RECORDING MEDIUM IN WHICH BITSTREAM IS STORED
An image encoding/decoding method and apparatus for signaling decoded picture buffer (DPB) related information and profile tier level (PTL) related information and a method of transmitting a bitstream are provided. The image decoding method according to the present disclosure may comprise obtaining first information indicating the number of one or more decoded picture buffer (DPB) parameter syntax structures in a video parameter set (VPS), obtaining the one or more DPB parameter syntax structures from the VPS based on the first information, obtaining second information between one or more multi-layer output layer sets (OLSs) and the one or more DPB parameter syntax structures from the VPS based on the first information, selecting a DPB parameter syntax structure that applies to a current OLS based on the second information, and processing the current OLS based on the selected DPB parameter syntax structure.