Patent classifications
H04N19/597
Three-dimensional data encoding method, three-dimensional data decoding method, three-dimensional data encoding device, three-dimensional data decoding device
A three-dimensional data encoding method includes: shifting point cloud data indicating three-dimensional positions in a three-dimensional space by a first shift amount; dividing the point cloud data into pieces of sub point cloud data; shifting each of the pieces of sub point cloud data by a second shift amount based on a position of one of the subspaces that includes the sub point cloud data, the pieces of sub point cloud data being included in the point cloud data shifted by the first shift amount; and encoding the pieces of sub point cloud data shifted, to generate a bitstream. The bitstream includes first shift information for calculating the first shift amount, and pieces of second shift information each for calculating a corresponding one of second shift amounts by which the pieces of sub point cloud data are shifted and each of which is the second shift amount.
Three-dimensional data encoding method, three-dimensional data decoding method, three-dimensional data encoding device, three-dimensional data decoding device
A three-dimensional data encoding method includes: shifting point cloud data indicating three-dimensional positions in a three-dimensional space by a first shift amount; dividing the point cloud data into pieces of sub point cloud data; shifting each of the pieces of sub point cloud data by a second shift amount based on a position of one of the subspaces that includes the sub point cloud data, the pieces of sub point cloud data being included in the point cloud data shifted by the first shift amount; and encoding the pieces of sub point cloud data shifted, to generate a bitstream. The bitstream includes first shift information for calculating the first shift amount, and pieces of second shift information each for calculating a corresponding one of second shift amounts by which the pieces of sub point cloud data are shifted and each of which is the second shift amount.
Method and apparatus for point cloud coding
An apparatus for point cloud decoding includes processing circuitry. The processing circuitry receives, from a coded bitstream for a point cloud, encoded occupancy codes for nodes in an octree structure for the point cloud. The nodes in the octree structure correspond to three dimensional (3D) partitions of a space of the point cloud. Sizes of the nodes are associated with sizes of the corresponding 3D partitions. Further, the processing circuitry decodes, from the encoded occupancy codes, occupancy codes for the nodes. At least a first occupancy code for a child node of a first node is decoded without waiting for a decoding of a second occupancy code for a second node having a same node size as the first node. Then, the processing circuitry reconstructs the octree structure based on the decoded occupancy codes for the nodes, and reconstructs the point cloud based on the octree structure.
Method for alignment across layers in coded video stream
A method, computer program, and computer system is provided for aligning across layers in a coded video stream. A video bitstream having multiple layers is decoded. One or more subpicture regions are identified from among the multiple layers of the decoded video bitstream, the subpicture regions including a background region and one or more foreground subpicture regions. An enhanced subpicture is decoded and displayed based on a determination that a foreground subpicture region is selected. The background region is decoded and displayed based on a determination that a foreground subpicture region was not selected.
Hash-based attribute prediction for point cloud coding
A method, computer program, and computer system is provided for point cloud coding. Data corresponding to a point cloud is received. Hash elements corresponding to attribute values associated with the received data is reconstructed. A size of a hash table may be decreased based on deleting one or more of the hash elements corresponding to non-border regions associated with the attribute values. The data corresponding to the point cloud is decoded based on the reconstructed hash elements.
Techniques and apparatus for generalized Trisoup geometry coding
There is included a method and apparatus comprising computer code configured to cause a processor or processors to perform obtaining a leaf node of geometry based point cloud compression (G-PCC) data, splitting the leaf node into a plurality of cuboids, deriving separate triangle soups for each of the cuboids, and coding a plurality of flags respectively for each of the edges of each of the cuboids, where the plurality of flags indicate whether vertices of the separate triangle soups are present on ones of the edges.
Techniques and apparatus for generalized Trisoup geometry coding
There is included a method and apparatus comprising computer code configured to cause a processor or processors to perform obtaining a leaf node of geometry based point cloud compression (G-PCC) data, splitting the leaf node into a plurality of cuboids, deriving separate triangle soups for each of the cuboids, and coding a plurality of flags respectively for each of the edges of each of the cuboids, where the plurality of flags indicate whether vertices of the separate triangle soups are present on ones of the edges.
Method and apparatus for inter-channel prediction and transform for point cloud attribute coding
A method of interframe point cloud attribute coding is performed by at least one processor and includes coding a first color attribute of a point of a point cloud to obtain a first reconstructed residual, coding a second color attribute of the point to obtain a second reconstructed residual, and determining a quantization index of the second reconstructed residual, based on the first reconstructed residual and the second reconstructed residual. The method further includes updating the second reconstructed residual, based on the quantization index and the first reconstructed residual.
Methods and apparatus for encoding, communicating and/or using images
Methods and apparatus for capturing, communicating and using image data to support virtual reality experiences are described. Images, e.g., frames, are captured at a high resolution but lower frame rate than is used for playback. Interpolation is applied to captured frames to generate interpolated frames. Captured frames, along with interpolated frame information, are communicated to the playback device. The combination of captured and interpolated frames correspond to a second frame playback rate which is higher than the image capture rate. Cameras operate at a high image resolution but slower frame rate than images could be captured with the same cameras at a lower resolution. Interpolation is performed prior to delivery to the user device with segments to be interpolated being selected based on motion and/or lens FOV information. A relatively small amount of interpolated frame data is communicated compared to captured frame data for efficient bandwidth use.
VOLUMETRIC VIDEO WITH AUXILIARY PATCHES
Methods and devices for encoding and decoding data representative of a 3D scene are disclosed. A set of first patches is generated from a first MVD content acquired from a first region of the 3D scene. A patch is a part of one of the views of the MVD content. A set of second patches is generated from a second MVD content acquired from a second region of the 3D scene. An atlas packing first and second patches is generated and associated with metadata indicating, for a patch of the atlas, whether the patch is a first or a second patch At the decoding side, first patches are used for rendering the viewport image and second patches are used for pre-processing or post-processing the viewport image.