Patent classifications
H04N21/8451
Method, apparatus and stream for volumetric video format
Methods and device for encoding/decoding data representative of a 3D scene. First data representative of texture of the 3D scene visible from a first viewpoint is encoded into first tracks. The first data is arranged in first tiles of a first frame. Second data representative of depth associated with points of the 3D scene is encoded into second tracks. The second data is arranged in second tiles of a second frame, the total number of second tiles being greater than the total number of first tiles. Instructions to extract at least a part of the first data and second data from at least a part of the at least a first track and at least a second track are further encoded into one or more third tracks.
EFFICIENT ENCODING OF DEPTH DATA ACROSS DEVICES
One embodiment provides a method comprising determining a spatial relationship between an augmented reality (AR) device and a camera-equipped device. The AR device is worn by a user. The camera-equipped device is positioned within proximity of the user. The method comprises determining, based on the spatial relationship, minimum and maximum boundaries of a user volume encompassing a head and one or more extremities of the user. The minimum and maximum boundaries represent a reduced depth range that is narrower than a full depth range of a camera frustum of the camera-equipped device. The method further comprises encoding, based on the minimum and maximum boundaries, a depth image frame captured via the camera-equipped device, and transmitting the encoded depth image frame over a remote network connection. A reconstructed 3D image based in part on the encoded depth image frame is rendered by a different AR device.
Information stream management
Aspects of the disclosure relate to management of information streams. The information streams can be delivered according to adaptive streaming mechanisms. In one aspect, a method of data stream management can comprise receiving a plurality of data streams having a specific bit rate and a segmentation signaling structure, comprising at least one segmentation signaling marker. The method may also comprise monitoring the segmentation signaling structure of at least one data stream and supplying, based on the monitoring, a metric indicative of compliance with a predetermined segmentation signaling structure.
Schedule-based uninterrupted buffering and streaming
Methods and systems are provided for schedule-based uninterrupted buffering and streaming. In an embodiment, at least one first media file can be received by a streaming server. The streaming server can create a first media file playlist including the at least one first media file. A schedule defining a first desired output time at which a client is to output the first media file playlist can be received by the streaming server. A schedule-based streaming buffer can be generated by the streaming server. The schedule-based streaming buffer can store the first media file playlist prior to the first desired output time. The schedule-based streaming buffer can be provided to the client by the streaming server such that the client can output the first media file playlist at the first desired output time.
Video Decoding Method, Video Coding Method, Apparatus, Device, and Storage Medium
A video encoding method, a video decoding method, an apparatus, a device, and a storage medium include parsing a first flag from the video bitstream when a video bitstream is allowed to be decoded by referencing a library picture corresponding to a library picture bitstream, using a value of a target parameter of the video bitstream as a value of a target parameter of the library picture bitstream referenced by the video bitstream when a value of the first flag is a first value, and reconstructing, based on the value of the target parameter of the library picture bitstream referenced by the video bitstream and the library picture bitstream referenced by the video bitstream to obtain the library picture corresponding to the library picture bitstream referenced by the video bitstream.
Broadcast streaming of panoramic video for interactive clients
For transmitting a panorama video having a resolution higher than a decoder can decode, at a transmitter side, encoded data coding different (groups of) spatial segments of a video picture of a video stream are packetized into separate substreams, to obtain a group of separate substreams. At a receiver side, from the group of separate substreams a proper subset of the separate substreams is extracted and combined to a data stream containing encoded data coding respectively a proper subset of the spatial segments or groups of subsequent spatial segments of the video picture of the video stream. Thus, a decoder may decode only a subregion of the video picture of the video stream, the subregion being defined by the spatial segments or groups of spatial segments coded in the encoded data contained in the data stream.
DUAL STREAM DYNAMIC GOP ACCESS BASED ON VIEWPORT CHANGE
A multi-view image stream encoded with primary and secondary image is accessed. Each primary image stream comprises groups of pictures (GOPs). Each secondary image stream comprises I-frames generated from a corresponding primary image stream. Viewpoint data collected in real time is received from a recipient decoding device to indicate that the viewer's viewpoint has changed from a specific time point. A camera is selected based on the viewer's changed viewpoint. It is determined whether the specific time point corresponds to a non-I-frame in a GOP of a primary image stream of the selected camera. If so, an I-frame from a secondary image stream corresponding to the primary image stream is transmitted to the recipient decoding device.
Method and apparatus for processing image service
A method, performed by a digital device, for processing an image service according to the present document comprises the steps of: receiving image information; decoding a first image on the basis of the image information; processing the decoded first image to be displayed on a first area of a display screen; and processing a second image to be displayed on a second area of the display screen.
Handling of coded video in sub-bitstream extraction process
Embodiments for video processing, including video coding, video decoding and video transcoding are described. One example method includes performing a conversion between a video having one or more video pictures in a video layer and a bitstream of the video according to a rule, wherein the rule specifies that, in a process of sub-bitstream extraction, removal of (i) a video coding layer (VCL) network abstraction layer (NAL) unit, (ii) filler data NAL units associated with the VCL NAL unit, and (iii) filler payload supplemental enhancement information (SEI) messages associated with the VCL NAL unit is performed regardless of an availability of an external means used to replace a parameter set that is removed during the sub-bitstream extraction.
Method for identification of random access point and picture types
An approach for reconstructing a current Network Abstraction Layer (NAL) unit for video decoding using at least one processor includes determining the current NAL unit to be an Intra Random Access Picture (IRAP) NAL unit; determining whether a previous NAL unit decoded immediately before the current NAL unit indicates an end of a coded video sequence (CVS); based on determining that the previous NAL unit indicates the end of the CVS, decoding the current NAL unit as one from among an Instantaneous Decoder Refresh (IDR) NAL unit or a Broken Link Access (BLA) NAL unit; and based on determining that the previous NAL unit does not indicate the end of the CVS, decoding the current NAL unit as a Clean Random Access (CRA) NAL unit, and reconstructing the decoded current NAL unit.