Patent classifications
H04N19/20
Point cloud compression
A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. The encoder generates an occupancy map and may also encode the occupancy map as an image based representation. In some embodiments, a video encoder encodes image based representations of spatial information for the points of the point cloud, image based representations of attribute values for points of the point cloud, and an image based representation of an occupancy map for the spatial and attribute images.
Point cloud compression
A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. The encoder generates an occupancy map and may also encode the occupancy map as an image based representation. In some embodiments, a video encoder encodes image based representations of spatial information for the points of the point cloud, image based representations of attribute values for points of the point cloud, and an image based representation of an occupancy map for the spatial and attribute images.
Point cloud and mesh compression using image/video codecs
An decoding device, an encoding device and a method for point cloud encoding is disclosed. The method includes generating, from a three-dimensional point cloud, multiple two-dimensional frames, the two-dimensional frames including at least a first frame representing a geometry of points in the three-dimensional point cloud and a second frame representing texture of points in the three-dimensional point cloud. The method also includes generating an occupancy map indicating locations of pixels in the two-dimensional frames that represent points in the three-dimensional point cloud. The method further includes encoding the two-dimensional frames and the occupancy map to generate a compressed bitstream. The method also includes transmitting the compressed bitstream.
METHOD FOR IDENTIFYING STATIONARY REGIONS IN FRAMES OF A VIDEO SEQUENCE
A method for identifying stationary regions in frames of a video sequence comprises receiving an encoded version of the video sequence, wherein the encoded version of the video sequence includes an intra-coded frame followed by a plurality of inter-coded frames; reading coding-mode information in the inter-coded frames of the encoded version of the video sequence, wherein the coding-mode information is indicative of blocks of pixels in the inter-coded frames being skip-coded; finding, using the read coding-mode information, one or more blocks of pixels that each was skip-coded in a respective plurality of consecutive frames in the encoded version of the video sequence; and designating each found block of pixels as a stationary region in the respective plurality of consecutive frames.
METHOD FOR IDENTIFYING STATIONARY REGIONS IN FRAMES OF A VIDEO SEQUENCE
A method for identifying stationary regions in frames of a video sequence comprises receiving an encoded version of the video sequence, wherein the encoded version of the video sequence includes an intra-coded frame followed by a plurality of inter-coded frames; reading coding-mode information in the inter-coded frames of the encoded version of the video sequence, wherein the coding-mode information is indicative of blocks of pixels in the inter-coded frames being skip-coded; finding, using the read coding-mode information, one or more blocks of pixels that each was skip-coded in a respective plurality of consecutive frames in the encoded version of the video sequence; and designating each found block of pixels as a stationary region in the respective plurality of consecutive frames.
Automatic extraction of secondary video streams
A system and method to automatically generate a secondary video stream based on an incoming primary video stream. The method including performing video analytics on the primary video stream to generate one or more analysis results, detecting the first target of interest using the analysis results, automatically extracting a first secondary video stream that captures at least a portion of a first target of interest and has a field of view smaller than that of the primary video stream, tracking the first target of interest, displaying the first secondary video stream, detecting a second target of interest using the analysis results, automatically adapting the first secondary video stream from the primary video stream to capture a portion of the first and second targets of interest, tracking the second target of interest, and displaying the first secondary stream including the portion of the first and second targets of interest.
Automatic extraction of secondary video streams
A system and method to automatically generate a secondary video stream based on an incoming primary video stream. The method including performing video analytics on the primary video stream to generate one or more analysis results, detecting the first target of interest using the analysis results, automatically extracting a first secondary video stream that captures at least a portion of a first target of interest and has a field of view smaller than that of the primary video stream, tracking the first target of interest, displaying the first secondary video stream, detecting a second target of interest using the analysis results, automatically adapting the first secondary video stream from the primary video stream to capture a portion of the first and second targets of interest, tracking the second target of interest, and displaying the first secondary stream including the portion of the first and second targets of interest.
Source color volume information messaging
Methods are described to communicate source color volume information in a coded bitstream using SEI messaging. Such data include at least the minimum, maximum, and average luminance values in the source data plus optional data that may include the color volume x and y chromaticity coordinates for the input color primaries (e.g., red, green, and blue) of the source data, and the color x and y chromaticity coordinates for the color primaries corresponding to the minimum, average, and maximum luminance values in the source data. Messaging data signaling an active region in each picture may also be included.
Source color volume information messaging
Methods are described to communicate source color volume information in a coded bitstream using SEI messaging. Such data include at least the minimum, maximum, and average luminance values in the source data plus optional data that may include the color volume x and y chromaticity coordinates for the input color primaries (e.g., red, green, and blue) of the source data, and the color x and y chromaticity coordinates for the color primaries corresponding to the minimum, average, and maximum luminance values in the source data. Messaging data signaling an active region in each picture may also be included.
SPLIT RENDERING OF EXTENDED REALITY DATA OVER 5G NETWORKS
An example device for processing extended reality (XR) data includes a processors configured to: parse entry point data of an XR scene to extract information about one or more required virtual objects for the XR scene, the required virtual objects including a number of dynamic virtual objects equal to or greater than one, each of the dynamic virtual objects including at least one dynamic media component for which media data is to be retrieved; initialize a number of streaming sessions equal to or greater than the number of dynamic virtual objects using the entry point data; configure quality of service (QoS) and charging information for the streaming sessions; retrieve media data for the dynamic virtual objects via the streaming sessions; and send the retrieved media data to a rendering unit to render the XR scene to include the retrieved media data at corresponding locations within the XR scene.