H04N21/8106

ADAPTIVE STREAMING OF MEDIA CONTENT WITH BITRATE SWITCHING
20230052666 · 2023-02-16 · ·

A method for adaptive streaming of media content with bitrate switching is described, wherein the media content comprising a plurality of consecutive media segments. The method comprising, at a media streaming server: transmitting a segment of the media content encoded in a first coding mode having a first bitrate; receiving an indication for a coding mode switch to a second coding mode having a second bitrate and in response transmitting a transition segment for transitioning between the first coding mode and the second coding mode; and transmitting another segment of the media content encoded in the second coding mode.

Audio improvement using closed caption data
11582532 · 2023-02-14 · ·

Methods and systems are described herein for improving audio for hearing impaired content consumers. An example method may comprise determining a content asset. Closed caption data associated with the content asset may be determined. At least a portion of the closed caption data may be determined based on a user setting associated with a hearing impairment. Compensating audio comprising a frequency translation associated with at least the portion of the closed caption data may be generated. The content asset may be caused to be output with audio content comprising the compensating audio and the original audio.

Vehicle-based media system with audio ad and visual content synchronization feature
11581969 · 2023-02-14 · ·

In one aspect, an example method to be performed by a vehicle-based media system includes (a) receiving audio content; (b) causing one or more speakers to output the received audio content; (c) using a microphone of the vehicle-based media system to capture the output audio content; (d) identifying reference audio content that has at least a threshold extent of similarity with the captured audio content; (e) identifying visual content based at least on the identified reference audio content; and (f) outputting, via a user interface of the vehicle-based media system, the identified visual content.

METHOD AND APPARATUS FOR PROCESSING VIRTUAL VIDEO LIVESTREAMING, STORAGE MEDIUM AND ELECTRONIC DEVICE
20230039789 · 2023-02-09 ·

A method includes: receiving text data and motion data of a virtual object, the motion data including a motion identifier of a specified motion and a start position identifier of a start position that the specified motion starts being in line with text in the text data; generating audio data and expression data of the virtual object according to the text data, and generating facial images of the virtual object according to the expression data; generating a background image sequence containing the specified motion according to the start position identifier and the motion identifier, the background image sequence including at least one background image; performing image fusion processing on the facial images and the at least one background image to obtain one or more live video frames; and synthesizing the live video frames with the audio data into a live video stream in real time.

Methods, systems, and media for providing dynamic media sessions with audio stream expansion features

Methods, systems, and media for providing dynamic media sessions with audio stream expansion features are provided. In some embodiments, the methods include: receiving an indication that audio content associated with a video content item is to be presented by a follower device synchronously with the audio content presented by the leader device; identifying candidate follower devices by determining whether devices connected to a local area network are capable of being designated as a follower device; causing a user interface to be presented that indicates each candidate follower device; receiving, via the user interface, a selection of one of the candidate follower devices; and transmitting, from the leader to the selected follower device, control instructions that cause the audio content associated with the video content item to be presented synchronously by the selected follower device with the video content item presented by the leader device.

Generating videos with a character indicating a region of an image
11595738 · 2023-02-28 · ·

Methods, systems, and computer-readable media for generating videos with characters indicating regions of images are provided. For example, an image containing a first region may be received. At least one characteristic of a character may be obtained. A script containing a first segment of the script may be received. The first segment of the script may be related to the first region of the image. The at least one characteristic of a character and the script may be used to generate a video of the character presenting the script and at least part of the image, where the character visually indicates the first region of the image while presenting the first segment of the script.

METADATA FOR DUCKING CONTROL

An audio encoding device and an audio decoding device are described herein. The audio encoding device may examine a set of audio channels/channel groups representing a piece of sound program content and produce a set of ducking values to associate with one of the channels/channel groups. During playback of the piece of sound program content, the ducking values may be applied to all other channels/channel groups. Application of these ducking values may cause (1) the reduction in dynamic range of ducked channels/channel groups and/or (2) movement of channels/channel groups in the sound field. This ducking may improve intelligibility of audio in the non-ducked channel/channel group. For instance, a narration channel/channel group may be more clearly heard by listeners through the use of selective ducking of other channels/channel groups during playback.

DECODER FOR DECODING A MEDIA SIGNAL AND ENCODER FOR ENCODING SECONDARY MEDIA DATA COMPRISING METADATA OR CONTROL DATA FOR PRIMARY MEDIA DATA
20180007398 · 2018-01-04 ·

An encoder for encoding secondary media data including metadata and control data for primary media data is shown, wherein the encoder is configured to encode the secondary media data using adding redundancy or bandlimiting and wherein the encoder is configured to output the encoded secondary media data as a stream of digital words. Therefore, the stream of digital words may be formed such that it is capable to resist a typical processing of a digital audio stream. Furthermore, processors for processing a digital audio stream are able to process the stream of digital words, since the stream of digital words may be designed as an audio-like or analog-like digital stream.

Video generation method and apparatus, electronic device, and computer readable medium

Disclosed are a video generation method and apparatus, an electronic device, and a computer readable medium. A specific embodiment of the method comprises: obtaining a video footage and an audio footage, the video footage comprising a picture footage; determining a music point of the audio footage, the music point being used for dividing the audio footage into a plurality of audio clips; using the video footage to generate a video clip for each music clip in the audio footage to obtain a plurality of video clips, corresponding music clips and video clips having the same duration; and splicing the plurality of video clips according to the time when music clips respectively corresponding to the plurality of video clips appear in the audio footage, and adding the audio footage as a video audio signal to obtain a composite video.

METHODS AND DEVICES FOR PROVIDING PERSONALIZED AUDIO TO A USER
20230007321 · 2023-01-05 · ·

The present application describes a method (400) for providing personalized audio to a user. The method (400) comprises receiving (401) a manifest file (140) for a media element from which audio is to be rendered, wherein the manifest file (140) comprises a description (141) for a plurality of different presentations (152) of audio content of the media element. In addition, the method (400) comprises selecting (402) a presentation (152) from the plurality of presentations (152) based on the manifest file (140). The method (400) further comprises receiving (403) a list of audio track objects comprised within the media element, and selecting (404) an audio track object from the list of audio track objects, in dependence of the selected presentation (152).