Patent classifications
G11B27/10
VIDEO PROCESSING DEVICE AND METHOD
A video processing apparatus includes a memory storing instructions, and at least one processor configured to execute the instructions to generate a plurality of feature information by analyzing a video signal comprising a plurality of images based on a first DNN, extract a first altitude component and a first planar component corresponding to a movement of an object in a video from the video signal based on a second DNN, extract a second planar component corresponding to a movement of a sound source in audio from a first audio signal based on a third DNN, generate a second altitude component based on the first altitude component, the first planar component, and the second planar component, output a second audio signal comprising the second altitude component based on the feature information, and synchronize the second audio signal with the video signal and output the synchronized second audio signal and video signal.
VIDEO PROCESSING DEVICE AND METHOD
A video processing apparatus includes a memory storing instructions, and at least one processor configured to execute the instructions to generate a plurality of feature information by analyzing a video signal comprising a plurality of images based on a first DNN, extract a first altitude component and a first planar component corresponding to a movement of an object in a video from the video signal based on a second DNN, extract a second planar component corresponding to a movement of a sound source in audio from a first audio signal based on a third DNN, generate a second altitude component based on the first altitude component, the first planar component, and the second planar component, output a second audio signal comprising the second altitude component based on the feature information, and synchronize the second audio signal with the video signal and output the synchronized second audio signal and video signal.
CONTEXT-BASED MEDIA INDEXING
Indexing of media sources available to media composition applications, such as video editing applications and digital audio workstations, is extended to sources of an arbitrary type, including non-file type sources. Examples of such sources include devices using baseband video and audio protocols, network ports and physical devices connected via USB, Thunderbolt, etc. The sources are discovered and indexed even when not previously known to the application with the result that any available bitstream becomes available to the application. Applications access the sources via feeders, which are addressable portals through which media essence is received. The indexing methods involve the instantiation of a hierarchy of plug-in software modules by a format-handling subsystem of the application or by a media indexing service servicing one or more applications.
CONTEXT-BASED MEDIA INDEXING
Indexing of media sources available to media composition applications, such as video editing applications and digital audio workstations, is extended to sources of an arbitrary type, including non-file type sources. Examples of such sources include devices using baseband video and audio protocols, network ports and physical devices connected via USB, Thunderbolt, etc. The sources are discovered and indexed even when not previously known to the application with the result that any available bitstream becomes available to the application. Applications access the sources via feeders, which are addressable portals through which media essence is received. The indexing methods involve the instantiation of a hierarchy of plug-in software modules by a format-handling subsystem of the application or by a media indexing service servicing one or more applications.
DYNAMIC VISUAL INTENSITY RENDERING
The present technology can provide a mechanism for adjusting a visual effect that is associated with an audio artifact at a given frequency bandwidth that is attenuated by speaker characteristics. The intensity of the visual effects that is adjusted can also be attributed to a change in volume settings of a processing device as well as an intensity of a multimedia skin in which the visual effect is encoded. The multimedia skin includes filters, transitions/animations, and/or image universal processing, that can be applied to any set of photos, videos, and/or songs, in order to create, in real-time, many variations of the same digital multimedia file, wherein each multimedia skin leads to a specific video rendering.
Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
A method is performed by an electronic device with a display and a touch-sensitive surface. The method includes: displaying a progress icon that indicates a current position within a first piece of content; displaying a multi-purpose content navigation icon; while providing the first piece of content: detecting a first contact at a first location that corresponds to the multi-purpose content navigation icon; while continuing to detect the contact at the first location, moving the current position within the first piece of content at a predefined scrubbing rate; and, in response to detecting movement of the contact that includes a first component of movement in a direction that corresponds to movement on the display parallel to the first predefined direction, moving the current position within the first piece of content at a variable scrubbing rate that varies monotonically as the first component of movement increases.
Image acquisition system and method
A method of capturing free viewpoint content at a location includes recording video on each of a plurality of portable video recording devices at the location; each portable video recording device detecting a wireless synchronisation signal transmitted at the location; and each portable video recording device periodically adding a timestamp to its respective recorded video; where the timestamp is responsive to the detected wireless synchronisation signal, thereby enabling synchronisation of a plurality of recorded videos responsive to the timestamps.
Jukebox with customizable avatar
A digital downloading jukebox system including a mechanism for delivering custom services to a recognized user is provided. For example, information specific to a recognized user may be stored and optionally may include a recognized user avatar representative of the recognized user. The user avatar may be an image, video, and/or animation, which may be displayed on and/or played through the jukebox. The user avatar may be associated with transactions associated with the user. For example, an avatar may be displayed when a playlist of the recognized user is played, when a message is sent, etc. In other examples, the avatar may introduce instances of media by playing an audio and/or video message, and the avatar may sing, dance, etc. while an instance of media is playing.
Method for sharing and searching playlists
A system that provides for the accessing and playing of media files having differing associated rights such as non-DRM media files, purchased and downloaded media files, subscription download files such as tethered downloads, and subscription streamed DRM files. The system also provides a method and user interface for sharing a media collection among computing devices in communication via a network. The system allows access and playback, from each computing device on a network, of all media files in a media collection, regardless of their associated rights.
METHOD AND APPARATUS FOR SYNCHRONIZING AUDIO AND TEXT, READABLE MEDIUM, AND ELECTRONIC DEVICE
This disclosure relates to a method and apparatus for synchronizing audio and text, a readable medium, and an electronic device, and relates to the technical field of electronic information, the method including: in response to a first page displayed on a presentation interface being updated to a second page, determining whether the second page includes multimedia content, wherein audio corresponding to text in the first page is played while the first page is displayed; if the second page includes the multimedia content and the multimedia content is in a mute playing state, controlling the audio to be played continuously; determining a first playing identification of the audio when the playing of the multimedia content is stopped, and determining corresponding target text in accordance with the first playing identification; and updating the second page to a third page on the presentation interface, the third page including the target text.