Patent classifications
H04N21/2368
Encoder, decoder and methods for encoding and decoding data segments representing a time-domain data stream
An apparatus for decoding data segments representing a time-domain data stream, a data segment being encoded in the time domain or in the frequency domain, a data segment being encoded in the frequency domain having successive blocks of data representing successive and overlapping blocks of time-domain data samples. The apparatus includes a time-domain decoder for decoding a data segment being encoded in the time domain and a processor for processing the data segment being encoded in the frequency domain and output data of the time-domain decoder to obtain overlapping time-domain data blocks. The apparatus further includes an overlap/add-combiner for combining the overlapping time-domain data blocks to obtain a decoded data segment of the time-domain data stream.
Encoder, decoder and methods for encoding and decoding data segments representing a time-domain data stream
An apparatus for decoding data segments representing a time-domain data stream, a data segment being encoded in the time domain or in the frequency domain, a data segment being encoded in the frequency domain having successive blocks of data representing successive and overlapping blocks of time-domain data samples. The apparatus includes a time-domain decoder for decoding a data segment being encoded in the time domain and a processor for processing the data segment being encoded in the frequency domain and output data of the time-domain decoder to obtain overlapping time-domain data blocks. The apparatus further includes an overlap/add-combiner for combining the overlapping time-domain data blocks to obtain a decoded data segment of the time-domain data stream.
ACTION SYNCHRONIZATION FOR TARGET OBJECT
A method for synchronizing an action of a target object with source audio is provided. Facial parameter conversion is performed on an audio parameter of the source audio at different time periods to obtain source parameter information of the source audio at the respective time periods. Parameter extraction is performed on a target video that includes the target object to obtain target parameter information of the target video. Image reconstruction is performed on the target object in the target video based on the source parameter information of the source audio and the target parameter information of the target video, to obtain a reconstructed image. Further, a synthetic video is generated based on the reconstructed image, the synthetic video including the target object, and the action of the target object being synchronized with the source audio.
ACTION SYNCHRONIZATION FOR TARGET OBJECT
A method for synchronizing an action of a target object with source audio is provided. Facial parameter conversion is performed on an audio parameter of the source audio at different time periods to obtain source parameter information of the source audio at the respective time periods. Parameter extraction is performed on a target video that includes the target object to obtain target parameter information of the target video. Image reconstruction is performed on the target object in the target video based on the source parameter information of the source audio and the target parameter information of the target video, to obtain a reconstructed image. Further, a synthetic video is generated based on the reconstructed image, the synthetic video including the target object, and the action of the target object being synchronized with the source audio.
Method and Apparatus For Selection of Content From A Stream of Data
A main stream contains successive content elements of video and/or audio information that encode video and/or audio information at a first data rate. A computation circuit (144) computes main fingerprints from the successive content elements. A reference stream is received having a second data rate lower than the first data rate. The reference stream defines a sequence of the reference fingerprints. A comparator unit (144) compares the main fingerprints with the reference fingerprints. The main stream is monitored for the presence of inserted content elements between original content elements, where the original content elements have main fingerprints that match successive reference fingerprints and the inserted content elements have main fingerprints that do not match reference fingerprints. Rendering of inserted content elements to be skipped. In an embodiment when more than one content element matches only one is rendered. In another embodiment matching is used to control zapping to or from the main stream. In another embodiment matching is used to control linking of separately received mark-up information such as subtitles to points in the main stream.
Method and Apparatus For Selection of Content From A Stream of Data
A main stream contains successive content elements of video and/or audio information that encode video and/or audio information at a first data rate. A computation circuit (144) computes main fingerprints from the successive content elements. A reference stream is received having a second data rate lower than the first data rate. The reference stream defines a sequence of the reference fingerprints. A comparator unit (144) compares the main fingerprints with the reference fingerprints. The main stream is monitored for the presence of inserted content elements between original content elements, where the original content elements have main fingerprints that match successive reference fingerprints and the inserted content elements have main fingerprints that do not match reference fingerprints. Rendering of inserted content elements to be skipped. In an embodiment when more than one content element matches only one is rendered. In another embodiment matching is used to control zapping to or from the main stream. In another embodiment matching is used to control linking of separately received mark-up information such as subtitles to points in the main stream.
Automated Dynamic Data Extraction, Distillation, and Enhancement
A dynamic data extraction, distillation, and enhancement system is disclosed that includes a dynamic extraction, distillation, and enhancement framework. The framework includes an allocator, extractor, and deconstructor stored in a non-transitory memory that, when executed by a processor, receive files in different formats from data sources, determine a native format of each file, identify and extract an embedded object from a file, deconstruct the file into components, assign each file to one of a plurality of streams based on the native format of the file, assign the embedded object to a stream based on a format of the embedded object, and assign a deconstructed component to a stream based on a format of the deconstructed component. The native format includes one of text, video, image, or audio. Each stream corresponds to one native format. The streams include a text stream, an audio stream, a video stream, and an image stream.
Automated Dynamic Data Extraction, Distillation, and Enhancement
A dynamic data extraction, distillation, and enhancement system is disclosed that includes a dynamic extraction, distillation, and enhancement framework. The framework includes an allocator, extractor, and deconstructor stored in a non-transitory memory that, when executed by a processor, receive files in different formats from data sources, determine a native format of each file, identify and extract an embedded object from a file, deconstruct the file into components, assign each file to one of a plurality of streams based on the native format of the file, assign the embedded object to a stream based on a format of the embedded object, and assign a deconstructed component to a stream based on a format of the deconstructed component. The native format includes one of text, video, image, or audio. Each stream corresponds to one native format. The streams include a text stream, an audio stream, a video stream, and an image stream.
LIVE STREAMING PROCESSING METHOD, APPARATUS, ELECTRONIC DEVICE, AND COMPUTER-READABLE STORAGE MEDIUM
This application provides a live streaming processing method performed by an electronic device. The method includes: displaying a live streaming room, the live stream room having a host account, a host sub-account and multiple viewer accounts, the host sub-account being used for assisting the host account of the live streaming room in operation; receiving real-time live streaming data of the live streaming room, and displaying a live streaming content on a live streaming room page according to the real-time live streaming data, the real-time live streaming data collected from the host account and the viewer accounts; and displaying, in response to a live streaming room operation of the host sub-account, an operation result of the live streaming room operation of the host sub-account in the live streaming room, wherein the operation result of the live streaming room operation updates the live streaming content on the live streaming room page.
Low latency wireless virtual reality systems and methods
Virtual Reality (VR) processing devices and methods are provided for transmitting user feedback information comprising at least one of user position information and user orientation information, receiving encoded audio-video (A/V) data, which is generated based on the transmitted user feedback information, separating the A/V data into video data and audio data corresponding to a portion of a next frame of a sequence of frames of the video data to be displayed, decoding the portion of a next frame of the video data and the corresponding audio data, providing the audio data for aural presentation and controlling the portion of the next frame of the video data to be displayed in synchronization with the corresponding audio data.