G10H1/368

Musical composition file generation and management system

A system and method to identify a digital representation of a first musical composition including a set of musical blocks. A set of parameters associated with source content are identified. In accordance with one or more rules, one or more of the set of musical blocks of the first musical composition are modified based on the set of parameters to generate a derivative musical composition. An audio file including the derivative musical composition is generated.

Method of combining data

A method of combining data, the method comprising: receiving video data, the video data corresponding to recorded video having a video duration determined by a user; selecting backing audio data, the backing audio data corresponding to backing audio having a predetermined duration; determining a difference between the predetermined duration and the video duration; and modifying the backing audio data by adjusting the predetermined duration based on the video duration to create an adjusted predetermined duration, the adjusted predetermined duration being such that the backing audio and recorded video may be simultaneously output in synchronisation.

IMAGE PROCESSING METHOD AND APPARATUS, AND HARDWARE APPARATUS
20210350545 · 2021-11-11 ·

The disclosure discloses an image processing method and apparatus, and a hardware apparatus. The image processing method includes the following steps: an audio is acquired, and then is preprocessed to obtain audio attribute data at each first time point of the audio; first audio attribute data corresponding to a current time point is acquired; and preset processing is performed on an image to be processed according to the first audio attribute data. According to the image processing method of the embodiments of the disclosure, when performing preset processing on the image to be processed according to the audio attribute data, the image processing can be completed only by setting the relationship between the audio attribute data and the image processing operation, which improves the flexibility and efficiency of image processing.

SYSTEM AND METHOD FOR PROVIDING A VIDEO WITH LYRICS OVERLAY FOR USE IN A SOCIAL MESSAGING ENVIRONMENT
20210343264 · 2021-11-04 ·

Some embodiments of the present disclosure provide a server system associated with a media-providing service. The server system receives, from a first client device, video content created by the first client device. The server system receives, from the first client device, an indication that the video content is to be associated with a song provided by the media-providing service. The server system provides, to a second client device, the video content in combination with the song. The server system provides, to the second client device, concurrently with the video content and the song, visual display of metadata about the song, including a name of the song.

METHOD FOR GENERATING ACTION ACCORDING TO AUDIO SIGNAL AND ELECTRONIC DEVICE

The disclosure provides a method for generating action according to an audio signal and an electronic device. The method includes: receiving an audio signal and extracting a high-level audio feature therefrom; extracting a latent audio feature from the high-level audio feature; in response to determining that the audio signal corresponds to a beat, obtaining a joint angle distribution matrix based on the latent audio feature; in response to determining that the audio signal corresponds to a music, obtaining a plurality of designated joint angles corresponding to a plurality of joint points based on the joint angle distribution matrix; and adjusting a joint angle of each of the joint points on the avatar according to the designated joint angles.

SYSTEM AND METHOD FOR PROVIDING A VIDEO WITH LYRICS OVERLAY FOR USE IN A SOCIAL MESSAGING ENVIRONMENT
20230335095 · 2023-10-19 ·

In accordance with an embodiment, described herein is a system and method for providing a live lyrics overlay in a social messaging environment. The system can utilize advances in three-dimensional mapping technology that allow social messaging services, to offer real time video lenses or overlays to their users, and extends this three-dimensional mapping technology to support for lyrics. During creation of a video with lyrics lens overlay, the lyrics corresponding to a selected song are retrieved from a lyrics source, and are displayed within the video. For example, with the lyrics lens, a user can record an image of themselves on live video, singing along to a song clip, with the lyrics of the song displayed as if they appear to be coming from their mouths. The created live lyrics content can also be shared with other users of a social messaging environment.

AUDIO-VISUAL EFFECTS SYSTEM FOR AUGMENTATION OF CAPTURED PERFORMANCE BASED ON CONTENT THEREOF

Visual effects schedules are applied to audiovisual performances with differing visual effects applied in correspondence with differing elements of musical structure. Segmentation techniques applied to one or more audio tracks (e.g., vocal or backing tracks) are used to compute some of the components of the musical structure. In some cases, applied visual effects schedules are mood-denominated and may be selected by a performer as a component of his or her visual expression or determined from an audiovisual performance using machine learning techniques.

Automated music composition and generation system employing virtual musical instrument libraries for producing notes contained in the digital pieces of automatically composed music
11776518 · 2023-10-03 · ·

An automated music composition and generation system including a system user interface for enabling system users to review and select one or more musical experience descriptors, as well as time and/or space parameters; and an automated music composition and generation engine, operably connected to the system user interface, for receiving, storing and processing musical experience descriptors and time and/or space parameters selected by the system user, so as to automatically compose and generate one or more digital pieces of music in response to the musical experience descriptors and time and/or space parameters selected by the system user. Each digital piece of composed and generated music contains a set of musical notes arranged and performed in the digital piece of music. The engine includes: a digital piece creation subsystem and a digital audio sample producing subsystem supported by virtual musical instrument libraries.

METHOD AND APPARATUS FOR MATCHING MUSIC WITH VIDEO, COMPUTER DEVICE, AND STORAGE MEDIUM
20230290382 · 2023-09-14 ·

This application relates to a method for matching music with a video performed by a computer device, and a storage medium. The method includes: determining a cut speed of a video; determining a long-time audio speed corresponding to each of a plurality of pieces of candidate music according to a high-scale point and a music duration of the candidate music; selecting matched music from the pieces of candidate music according to the cut speed and the corresponding long-time audio speeds; determining, according to a video duration of the video and a high-scale point corresponding to the matched music, a short-time audio speed corresponding to each music clip in the matched music; and determining a target music clip in the matched music according to the cut speed of the video and the corresponding short-time audio speed, and synthesizing the target music clip and the video to obtain a target video.

Automated generation of coordinated audiovisual work based on content captured from geographically distributed performers

Vocal audio of a user together with performance synchronized video is captured and coordinated with audiovisual contributions of other users to form composite duet-style or glee club-style or window-paned music video-style audiovisual performances. In some cases, the vocal performances of individual users are captured (together with performance synchronized video) on mobile devices, television-type display and/or set-top box equipment in the context of karaoke-style presentations of lyrics in correspondence with audible renderings of a backing track. Contributions of multiple vocalists are coordinated and mixed in a manner that selects for presentation, at any given time along a given performance timeline, performance synchronized video of one or more of the contributors. Selections are in accord with a visual progression that codes a sequence of visual layouts in correspondence with other coded aspects of a performance score such as pitch tracks, backing audio, lyrics, sections and/or vocal parts.