Patent classifications
H04N19/164
DEEP PALETTE PREDICTION
Example embodiments allow for training of encoders (e.g., artificial neural networks (ANNs)) to generate a color palette based on an input image. The color palette can then be used to generate, using the input image, a quantized, reduced color depth image that corresponds to the input image. Differences between a plurality of such input images and corresponding quantized images are used to train the encoder. Encoders trained in this manner are especially suited for generating color palettes used to convert images into different reduced color depth image file formats. Such an encoder also has benefits, with respect to memory use and computational time or cost, relative to the median-cut algorithm or other methods for producing reduced color depth color palettes for images.
METHOD AND APPARATUS FOR PROCESSING NON-SEQUENTIAL POINT CLOUD MEDIA, DEVICE, AND STORAGE MEDIUM
This application provides a method and apparatus for processing non-sequential point cloud media, a device, and a storage medium. The method includes: processing non-sequential point cloud data of a static object using a Geometry-based Point Cloud Compression (GPCC) coding scheme to obtain a GPCC bitstream; encapsulating the GPCC bitstream to generate an item of at least one GPCC region; encapsulating the item of the at least one GPCC region to generate at least one piece of non-sequential point cloud media of the static object; transmitting media presentation description (MPD) signaling of the at least one piece of non-sequential point cloud media; receiving a first request message transmitted by a video playback device; and transmitting first non-sequential point cloud media, the item of the GPCC region being used to represent a GPCC component of a three-dimensional (3D) spatial region corresponding to the GPCC region, and the non-sequential point cloud media including: an identifier of the static object, so that a user can purposefully request non-sequential point cloud media of a same static object a plurality of times, thereby improving the user experience.
METHOD AND APPARATUS FOR PROCESSING NON-SEQUENTIAL POINT CLOUD MEDIA, DEVICE, AND STORAGE MEDIUM
This application provides a method and apparatus for processing non-sequential point cloud media, a device, and a storage medium. The method includes: processing non-sequential point cloud data of a static object using a Geometry-based Point Cloud Compression (GPCC) coding scheme to obtain a GPCC bitstream; encapsulating the GPCC bitstream to generate an item of at least one GPCC region; encapsulating the item of the at least one GPCC region to generate at least one piece of non-sequential point cloud media of the static object; transmitting media presentation description (MPD) signaling of the at least one piece of non-sequential point cloud media; receiving a first request message transmitted by a video playback device; and transmitting first non-sequential point cloud media, the item of the GPCC region being used to represent a GPCC component of a three-dimensional (3D) spatial region corresponding to the GPCC region, and the non-sequential point cloud media including: an identifier of the static object, so that a user can purposefully request non-sequential point cloud media of a same static object a plurality of times, thereby improving the user experience.
A Method, An Apparatus and a Computer Program Product for Video Encoding and Video Decoding
The embodiments relate to a method including generating a bitstream defining a presentation including an omnidirectional visual media content; encoding into the bitstream a parameter to indicate viewport-control options for viewing the presentation, wherein the viewport-control options includes options controllable by a receiving device and options not-controllable by the receiving device and sending the bitstream to the receiver device; receiving one of the indicated viewport-control options from the receiver device as a response; streaming the presentation to the receiver device; when the response has included an indication on a viewport-control controllable by the receiving device, the method also includes receiving information on viewport definitions from the receiver device during streaming of the presentation and adapting the presentation accordingly; when the response has included an indication on a viewport-control not- controllable by the receiving device, the presentation is streamed to the receiver device according to the viewport-control specified in the response.
SMART SMALL FORM-FACTOR PLUGGABLE (SFP) TRANSCEIVER
Approaches for processing video in a smart small form-factor pluggable (SFP) transceiver. The smart SFP transceiver may dynamically select, from a plurality of codecs available to the smart SFP transceiver, an appropriate codec for use in processing the video prior to the video being transmitted over a link. The selection of the codec may be based, at least in part, on assessed environmental attributes. The smart SFP transceiver may then use the codec selected by the smart SFP transceiver to process the video, e.g., the video may be encoded, compressed, or timing information generated.
METHODS AND APPARATUS FOR PERFORMING REAL-TIME VVC DECODING
Apparatus and methods for implementing a real-time Versatile Video Coding (VVC) decoder use multiple threads to address the limitation with existing parallelization techniques and fully utilizes the available CPU computation resource without compromising on the coding efficiency. The proposed Multi-threaded (MT) framework uses CTU level parallel processing techniques without compromising on the memory bandwidth. Picture level parallel processing separates the sequence into temporal levels by considering the picture's referencing hierarchy. Embodiments are provided using various optimization techniques to achieve real-time VVC decoding on heterogenous platforms with multi-core CPUs, for those bitstreams generated using a VVC reference encoder with a default configuration.
Opportunistic progressive encoding
Methods, systems, and devices are described for communicating data from multiple data terminals to an aggregator terminal over a communication link having changing link conditions. In some embodiments, source data is received at multiple data terminals, each in communication with an aggregator terminal over a communication link. For example, during a live newscast, one mobile camera may receive live video of an event from a first position while another mobile camera receives live video of the event from a second position. For various reasons (e.g., as the cameras move) each communication link may experience independently changing link conditions. Each data terminal encodes the source data (or store source data for later encoding) as a function of its respective link conditions and transmits encoded source data over its respective communication link to the aggregator terminal.
Opportunistic progressive encoding
Methods, systems, and devices are described for communicating data from multiple data terminals to an aggregator terminal over a communication link having changing link conditions. In some embodiments, source data is received at multiple data terminals, each in communication with an aggregator terminal over a communication link. For example, during a live newscast, one mobile camera may receive live video of an event from a first position while another mobile camera receives live video of the event from a second position. For various reasons (e.g., as the cameras move) each communication link may experience independently changing link conditions. Each data terminal encodes the source data (or store source data for later encoding) as a function of its respective link conditions and transmits encoded source data over its respective communication link to the aggregator terminal.
Switchable chroma sampling for wireless display
Methods, apparatuses and systems may provide for a video transmitter that generates a primary bitstream based on a video signal, wherein the primary bitstream is encoded with subsampled chroma information, and detects a static condition with respect to the video signal. Additionally, a plurality of auxiliary bitstreams may be generated, in response to the static condition, based on the video signal. Each of the plurality of auxiliary bitstreams may be encoded with full resolution chroma information. In one example, a video receiver may detect that the auxiliary bitstreams are associated with the primary bitstream, decode the primary bitstream and the plurality of auxiliary bitstreams to obtain luma information and the full resolution chroma information, and multiplex the luma information with the full resolution chroma information.
METHODS FOR NON-REFERENCE VIDEO-QUALITY PREDICTION
A system for non-reference video-quality prediction includes a video-processing block to receive an input bitstream and to generate a first vector, and a neural network to provide a predicted-quality vector after being trained using training data. The training data includes the first vector and a second vector, and elements of the first vector include high-level features extracted from a high-level syntax processing of the input bitstream.