H04N19/20

SPLIT RENDERING OF EXTENDED REALITY DATA OVER 5G NETWORKS
20220369000 · 2022-11-17 ·

An example device for processing extended reality (XR) data includes a processors configured to: parse entry point data of an XR scene to extract information about one or more required virtual objects for the XR scene, the required virtual objects including a number of dynamic virtual objects equal to or greater than one, each of the dynamic virtual objects including at least one dynamic media component for which media data is to be retrieved; initialize a number of streaming sessions equal to or greater than the number of dynamic virtual objects using the entry point data; configure quality of service (QoS) and charging information for the streaming sessions; retrieve media data for the dynamic virtual objects via the streaming sessions; and send the retrieved media data to a rendering unit to render the XR scene to include the retrieved media data at corresponding locations within the XR scene.

Systems and methods for region-of-interest tone remapping

Systems and methods are described for providing viewers of adaptive bit rate (ABR) streaming video with the option to view alternative streams in which an alternative tone mapping is applied to one or more regions of interest. The availability of streams with alternative tone mappings may be identified in a media presentation description (MPD) in an MPEG-DASH system. In some embodiments, the streaming video is divided into slices, and alternative tone mappings are applied to regions of interest within the slices. When a server receives a request from a client device for alternative tone mappings of different regions, slices with the appropriate mapping may be assembled on demand and delivered to the requestor as a single video stream. Tone mappings may be used, for example, to highlight particular players in a sporting event.

Systems and methods for region-of-interest tone remapping

Systems and methods are described for providing viewers of adaptive bit rate (ABR) streaming video with the option to view alternative streams in which an alternative tone mapping is applied to one or more regions of interest. The availability of streams with alternative tone mappings may be identified in a media presentation description (MPD) in an MPEG-DASH system. In some embodiments, the streaming video is divided into slices, and alternative tone mappings are applied to regions of interest within the slices. When a server receives a request from a client device for alternative tone mappings of different regions, slices with the appropriate mapping may be assembled on demand and delivered to the requestor as a single video stream. Tone mappings may be used, for example, to highlight particular players in a sporting event.

Geometric encoding of data

Disclosed in some examples are methods, systems, devices, and machine-readable mediums which encode data into a geometric representation for more efficient and secure processing. For example, data may be converted from a binary representation to a geometric representation using an encoding dictionary. The encoding dictionary specifies one or more geometric shapes used in the encoding. The geometrically encoded data may comprise one or more identifiers that specify one or more of the shapes of the encoding dictionary that best match one or more detected features in an image corresponding to the data. In some examples, the geometrically encoded data may also comprise one or more transformations of the one or more shapes to reduce error in the geometric encoding.

SYSTEMS AND METHODS FOR REGION-OF-INTEREST TONE REMAPPING
20230039194 · 2023-02-09 ·

Systems and methods are described for providing viewers of streamed video content with the option to view alternative video content in which alternative tone maps are applied to respective regions of interest. In some embodiments, the streamed video content is divided into slices, and alternative tone maps are applied to respective regions of interest within the slices. When a server receives a request from a client for alternative tone mappings of different regions, slices with the appropriate mapping may be assembled on demand and delivered to the client in a single video stream. Tone mappings may be used, for example, to highlight particular objects, such as players in a sporting event captured in the streamed video content.

SYSTEMS AND METHODS FOR REGION-OF-INTEREST TONE REMAPPING
20230039194 · 2023-02-09 ·

Systems and methods are described for providing viewers of streamed video content with the option to view alternative video content in which alternative tone maps are applied to respective regions of interest. In some embodiments, the streamed video content is divided into slices, and alternative tone maps are applied to respective regions of interest within the slices. When a server receives a request from a client for alternative tone mappings of different regions, slices with the appropriate mapping may be assembled on demand and delivered to the client in a single video stream. Tone mappings may be used, for example, to highlight particular objects, such as players in a sporting event captured in the streamed video content.

Moving image analysis apparatus, system, and method

A moving image analysis apparatus includes at least one of a processor and a circuitry configured to perform operations including acquiring first data and second data used in processing, in which a moving image is compressed and encoded, for a first frame and a second frame, respectively, included in the moving image, detecting first feature data indicating a first feature of the moving image on the basis of the first frame and the first data and detecting second feature data indicating a second feature of the moving image on the basis of the second frame and the second data, and detecting an object included in the first frame on the basis of the first feature data and the second feature data.

Inference method, inference device and display
11494678 · 2022-11-08 · ·

An inference method, an inference device, and a display are provided. The method includes: receiving an input signal through a first inference device or a second inference device; performing a first inference operation according to the input signal through the first inference device to obtain first inference information; performing a second inference operation according to the input signal through the second inference device to obtain second inference information; and providing an output signal according to the input signal, the first inference information and the second inference information through the second inference device.

Viewport-based transcoding for immersive visual streams

In one example, a processing system including at least one processor may obtain a predicted viewport of a mobile computing device for an immersive visual stream, identify a first plurality of blocks of a frame of the immersive visual stream that are associated with the predicted viewport, encode the first plurality of blocks at a first encoding quality level, and encode a second plurality of blocks of the frame at a second encoding quality level, where the second encoding quality level is associated with a lesser visual quality as compared to the first encoding quality level and where the second plurality of blocks are outside of the predicted viewport. The processing system may then transmit the frame having the first plurality of blocks encoded at the first encoding quality level and the second plurality of blocks encoded at the second encoding quality level to the mobile computing device.

Viewport-based transcoding for immersive visual streams

In one example, a processing system including at least one processor may obtain a predicted viewport of a mobile computing device for an immersive visual stream, identify a first plurality of blocks of a frame of the immersive visual stream that are associated with the predicted viewport, encode the first plurality of blocks at a first encoding quality level, and encode a second plurality of blocks of the frame at a second encoding quality level, where the second encoding quality level is associated with a lesser visual quality as compared to the first encoding quality level and where the second plurality of blocks are outside of the predicted viewport. The processing system may then transmit the frame having the first plurality of blocks encoded at the first encoding quality level and the second plurality of blocks encoded at the second encoding quality level to the mobile computing device.