Patent classifications
H04N21/2368
Low latency wireless virtual reality systems and methods
Virtual Reality (VR) processing devices and methods are provided for transmitting user feedback information comprising at least one of user position information and user orientation information, receiving encoded audio-video (A/V) data, which is generated based on the transmitted user feedback information, separating the A/V data into video data and audio data corresponding to a portion of a next frame of a sequence of frames of the video data to be displayed, decoding the portion of a next frame of the video data and the corresponding audio data, providing the audio data for aural presentation and controlling the portion of the next frame of the video data to be displayed in synchronization with the corresponding audio data.
MANAGEMENT OF THE HTTP ADAPTIVE STREAMING OF AN ITEM OF DIGITAL CONTENT IN SCREEN SAVER MODE
A method for managing the HTTP adaptive streaming of an item of digital content within a multimedia stream player terminal in a local communication network is disclosed, the multimedia stream player terminal being associated with a terminal for rendering the item of digital content. On detection of an active state of a screen saver mode of the rendering terminal, such a management method implements a progressive reduction of a video rendering quality of the streamed item of digital content, and maintains an audio rendering quality of the item of content.
SYSTEMS AND METHODS FOR PROVIDING OPTIMIZED TIME SCALES AND ACCURATE PRESENTATION TIME STAMPS
The disclosed computer-implemented method includes determining, for multiple different media items, a current time scale at which the media items are encoded for distribution, where at least two of the media items are encoded at different frame rates. The method then includes identifying, for the media items, a unified time scale that provides a constant frame interval for each of the media items. The method also includes changing at least one of the media items from the current time scale to the identified unified time scale to provide a constant frame interval for the changed media item(s). Various other methods, systems, and computer-readable media are also disclosed.
METHOD AND DEVICE FOR GENERATING SPEECH VIDEO USING AUDIO SIGNAL
A device according to an embodiment has one or more processors and a memory storing one or more programs executable by the one or more processors. The device includes a first encoder configured to receive a person background image corresponding to a video part of a speech video of a person and extract an image feature vector from the person background image, a second encoder configured to receive a speech audio signal corresponding to an audio part of the speech video and extract a voice feature vector from the speech audio signal, a combiner configured to generate a combined vector by combining the image feature vector output from the first encoder and the voice feature vector output from the second encoder, and a decoder configured to reconstruct the speech video of the person using the combined vector as an input.
SELECTING SUPPLEMENTAL AUDIO SEGMENTS BASED ON VIDEO ANALYSIS
Aspects of the present application correspond to generation of supplemental content based on processing information associated with content to be rendered. More specifically, aspects of the present application correspond to the generation of audio track information, such as music tracks, that are created for playback during the presentation of video content. Illustratively, one or more frames of the video content are processed by machine learned algorithm(s) to generate processing results indicative of one or more attributes characterizing individual frames of video content. A selection system can then identify potential music track or other audio data in view of the processing results.
METHOD AND APPARATUS FOR PROCESSING AUDIO AND VIDEO, ELECTRONIC DEVICE AND STORAGE MEDIUM
A method and apparatus for processing an audio and video. The method includes: acquiring a target processing request including a target audio and video data stream; determining a target audio and video pipeline corresponding to the target processing request; the audio and video pipeline being constituted based on a plurality of functional components arranged in a chain structure, and the functional components being uniformly dispatched input data and recovered output data by a preset data stream dispatching module; and calling the target audio and video pipeline to continuously process the target audio and video data stream, and continuously outputting a processed audio and video data stream obtained after processing.
DISTRIBUTED NETWORK RECORDING SYSTEM WITH SINGLE USER CONTROL
A master recording session at a server computer corresponds to a video content stored in memory accessible by the server computer. A first device and a second device are provided access to the master recording session and the master recording session is updated responsive to receipt of an update from the first device, where the update reflects initiation of playback of the video content at a time stamp corresponding to the timeline of the video content and includes an audio input configuration for the second device. The update is provided to the second device and an audio recording is received from the second device corresponding to a portion of the video content from the time stamp, where the audio recording is recorded by the second device using the implemented audio input configuration for the second device.
DISTRIBUTED NETWORK RECORDING SYSTEM WITH SINGLE USER CONTROL
A master recording session at a server computer corresponds to a video content stored in memory accessible by the server computer. A first device and a second device are provided access to the master recording session and the master recording session is updated responsive to receipt of an update from the first device, where the update reflects initiation of playback of the video content at a time stamp corresponding to the timeline of the video content and includes an audio input configuration for the second device. The update is provided to the second device and an audio recording is received from the second device corresponding to a portion of the video content from the time stamp, where the audio recording is recorded by the second device using the implemented audio input configuration for the second device.
On-device multiplexing of streaming media content
Techniques are disclosed for multiplexing a dynamic bit-rate video stream with an audio stream received by a client device in a manner that allows the resulting multiplexed stream to be played back without disruption, despite dynamic changes in the bit rate of the video stream that may occur. A content server may stream both a video stream and an audio stream to a client device for playback. The client device may multiplex the video and audio streams prior to them being presented to a playback engine for decoding and playback to a user.
On-device multiplexing of streaming media content
Techniques are disclosed for multiplexing a dynamic bit-rate video stream with an audio stream received by a client device in a manner that allows the resulting multiplexed stream to be played back without disruption, despite dynamic changes in the bit rate of the video stream that may occur. A content server may stream both a video stream and an audio stream to a client device for playback. The client device may multiplex the video and audio streams prior to them being presented to a playback engine for decoding and playback to a user.