H04N19/20

System and method for content-layer based video compression
11582494 · 2023-02-14 · ·

Embodiments of the present invention disclose a method of content-layer based compression of a video being broadcasted over a network. The method may include: receiving a video stream comprising a plurality of video stream frames; identifying in at least some of the plurality of video stream frames at least two content-layers of predefined content-layers to yield corresponding at least two content-layer streams, wherein each of the at least two content-layer streams is associated with one of the at least two content-layers; and compressing each of the at least two content-layer video streams according to predetermined parameters of the content-layer associated with the respective content-layer video stream and according to available resources of the network to yield corresponding at least two compressed content-layer streams.

System and method for content-layer based video compression
11582494 · 2023-02-14 · ·

Embodiments of the present invention disclose a method of content-layer based compression of a video being broadcasted over a network. The method may include: receiving a video stream comprising a plurality of video stream frames; identifying in at least some of the plurality of video stream frames at least two content-layers of predefined content-layers to yield corresponding at least two content-layer streams, wherein each of the at least two content-layer streams is associated with one of the at least two content-layers; and compressing each of the at least two content-layer video streams according to predetermined parameters of the content-layer associated with the respective content-layer video stream and according to available resources of the network to yield corresponding at least two compressed content-layer streams.

Method and apparatus for point cloud coding
11582469 · 2023-02-14 · ·

An apparatus for point cloud decoding includes processing circuitry. The processing circuitry receives, from a coded bitstream for a point cloud, encoded occupancy codes for nodes in an octree structure for the point cloud. The nodes in the octree structure correspond to three dimensional (3D) partitions of a space of the point cloud. Sizes of the nodes are associated with sizes of the corresponding 3D partitions. Further, the processing circuitry decodes, from the encoded occupancy codes, occupancy codes for the nodes. At least a first occupancy code for a child node of a first node is decoded without waiting for a decoding of a second occupancy code for a second node having a same node size as the first node. Then, the processing circuitry reconstructs the octree structure based on the decoded occupancy codes for the nodes, and reconstructs the point cloud based on the octree structure.

Method and apparatus for point cloud coding
11582469 · 2023-02-14 · ·

An apparatus for point cloud decoding includes processing circuitry. The processing circuitry receives, from a coded bitstream for a point cloud, encoded occupancy codes for nodes in an octree structure for the point cloud. The nodes in the octree structure correspond to three dimensional (3D) partitions of a space of the point cloud. Sizes of the nodes are associated with sizes of the corresponding 3D partitions. Further, the processing circuitry decodes, from the encoded occupancy codes, occupancy codes for the nodes. At least a first occupancy code for a child node of a first node is decoded without waiting for a decoding of a second occupancy code for a second node having a same node size as the first node. Then, the processing circuitry reconstructs the octree structure based on the decoded occupancy codes for the nodes, and reconstructs the point cloud based on the octree structure.

Generative adversarial neural network assisted video reconstruction

A latent code defined in an input space is processed by the mapping neural network to produce an intermediate latent code defined in an intermediate latent space. The intermediate latent code may be used as appearance vector that is processed by the synthesis neural network to generate an image. The appearance vector is a compressed encoding of data, such as video frames including a person's face, audio, and other data. Captured images may be converted into appearance vectors at a local device and transmitted to a remote device using much less bandwidth compared with transmitting the captured images. A synthesis neural network at the remote device reconstructs the images for display.

Generative adversarial neural network assisted video reconstruction

A latent code defined in an input space is processed by the mapping neural network to produce an intermediate latent code defined in an intermediate latent space. The intermediate latent code may be used as appearance vector that is processed by the synthesis neural network to generate an image. The appearance vector is a compressed encoding of data, such as video frames including a person's face, audio, and other data. Captured images may be converted into appearance vectors at a local device and transmitted to a remote device using much less bandwidth compared with transmitting the captured images. A synthesis neural network at the remote device reconstructs the images for display.

Transferring data from autonomous vehicles
11580687 · 2023-02-14 · ·

A system includes at least one imaging sensor and a processor. The processor is configured to acquire, using the imaging sensor, detected data describing an environment of an autonomous vehicle. The processor is further configured to derive reference data, which describe the environment, from a predefined map, to compute difference data representing a difference between the detected data and the reference data, and to transfer the difference data. Other embodiments are also described.

Constraints for inter-layer referencing

A video coding method using inter-layer prediction or referencing is provided. A video decoder receives data from a bitstream carrying data for video pictures in a plurality of different layers. At least one of the plurality of layers comprises temporal sublayers that correspond to levels in a hierarchical temporal prediction structure. Each temporal sublayer is associated with a temporal identifier. The video decoder receives an inter-layer prediction constraint parameter constraining a maximum temporal sublayer used in inter-layer prediction. The video decoder reconstructs a first picture in a first layer by referencing data of a second picture in a second layer. A temporal identifier of the referenced data satisfies the received inter-layer prediction constraint parameter.

Method and apparatus for storage and signaling of sub-sample entry descriptions

Described are methods, apparatuses and computer program products for signaling and storing compressed point clouds. Sub-sample entries associated with sequences of sub-samples within sequences of samples may indicate whether sequences of sub-samples were encapsulated alone in a track, without other sub-samples or additional header data. Sub-sample entry types can be indexed at track-level sub-sample description boxes. Point cloud compression coded bitstream component types may be signaled by including respective point cloud unit header information in a codec-specific parameters-related field of track level sub-sample description boxes. Sub-sample information boxes may indicate sub-sample entry indices for respective sub-samples. A flag in such information boxes may indicate the presence of sub-sample description entry indexes. Description index boxes can contain sub-sample description entry indexes in the same container as sub-sample information boxes. Track fragment header boxes can include sub-sample description entry indices that apply to samples of a track fragment.

Method and apparatus for storage and signaling of sub-sample entry descriptions

Described are methods, apparatuses and computer program products for signaling and storing compressed point clouds. Sub-sample entries associated with sequences of sub-samples within sequences of samples may indicate whether sequences of sub-samples were encapsulated alone in a track, without other sub-samples or additional header data. Sub-sample entry types can be indexed at track-level sub-sample description boxes. Point cloud compression coded bitstream component types may be signaled by including respective point cloud unit header information in a codec-specific parameters-related field of track level sub-sample description boxes. Sub-sample information boxes may indicate sub-sample entry indices for respective sub-samples. A flag in such information boxes may indicate the presence of sub-sample description entry indexes. Description index boxes can contain sub-sample description entry indexes in the same container as sub-sample information boxes. Track fragment header boxes can include sub-sample description entry indices that apply to samples of a track fragment.