G10H2250/311

Computer-Implemented Method, System, and Non-Transitory Computer-Readable Storage Medium for Inferring Evaluation of Performance Information
20230009481 · 2023-01-12 ·

A computer-implemented method includes obtaining a trained model trained to store a relationship between first performance information and evaluation information. The first performance information includes a plurality of performance units. The evaluation information includes a plurality of pieces of evaluation information respectively associated with the plurality of performance units. The method also includes obtaining second performance information including an evaluation of each performance unit of the plurality of performance units. The method also includes processing the second performance information using the trained model to infer the evaluation of the each performance unit.

AUDIO TRANSPOSITION

An electronic device comprising circuitry configured to separate by audio source separation a first audio input signal into a first vocal signal and an accompaniment, and to transpose an audio output signal by a transposition value based on a pitch ratio, wherein the pitch ratio is based on comparing a first pitch range of the first vocal signal and a second pitch range of the second vocal signal.

HARMONY-AWARE HUMAN MOTION SYNTHESIS WITH MUSIC
20230005201 · 2023-01-05 ·

A method and device for harmony-aware audio-driven motion synthesis are provided. The method includes determining a plurality of testing meter units according to an input audio, each testing meter unit corresponding to an input audio sequence of the input audio, obtaining an auditory input corresponding to each testing meter unit, obtaining an initial pose of each testing meter unit as a visual input based on a visual motion sequence synthesized for a previous testing meter unit, and automatically generating a harmony-aware motion sequence corresponding to the input audio using a generator of a generative adversarial network (GAN) model. The GAN model is trained by incorporating a hybrid loss function. The hybrid loss function includes a multi-space pose loss, a harmony loss, and a GAN loss. The harmony loss is determined according to beat consistencies of audio-visual beat pairs.

Electronic musical instrument, electronic musical instrument control method, and storage medium

An electronic musical instrument includes at least one processor that, in accordance with a user operation on an operation unit, obtains lyric data and waveform data corresponding to a first tone color; inputs the obtained lyric data to a trained model so as to cause the trained model to output acoustic feature data in response thereto; generates waveform data corresponding to a singing voice of a singer and corresponding to a second tone color that is different from the first tone color, based on the acoustic feature data outputted from the trained model and the obtained waveform data corresponding to the first tone color; and outputs a singing voice based on the generated waveform data corresponding to the second tone color.

Systems and methods for embedding data in media content

An electronic device modifies a first media content item by superimposing a first set of data over a first accented musical event. The first accented musical event has a first audio profile. The first set of data has a second audio profile configured to be masked by the first audio profile during playback of the first media content item. The electronic device transmits, to a second electronic device, the modified first media content item.

METHODS, SYSTEMS, AND MEDIA FOR RIGHTS MANAGEMENT OF EMBEDDED SOUND RECORDINGS USING COMPOSITION CLUSTERING

Methods, systems, and media for determining and presenting information related to embedded sound recordings are provided. In some embodiments, the method comprises: receiving a content item; extracting a sound recording from the content item; generating a melody fingerprint of the extracted sound recording; determining whether the melody fingerprint of the extracted sound recording matches one of a plurality of clusters of similar sounding sound recordings in a reference database, wherein each cluster in the plurality of clusters of similar sounding sound recordings is associated with ownership information based on a plurality of ownership information associated with each of the sound recordings in the cluster; in response to determining that the melody fingerprint of the extracted sound recording matches a cluster of similar sounding sound recordings, retrieving ownership information associated with the cluster; mapping the ownership information to the sound recording extracted from the content item; and causing an action to be performed on the content item based on the mapped ownership information

Computer-Implemented Method, System, and Non-Transitory Computer-Readable Storage Medium for Inferring Audience's Evaluation of Performance Data
20220414472 · 2022-12-29 ·

A computer-implemented method includes obtaining a trained model trained to store a relationship between first performance data and first evaluation data. The first performance data indicates a performance performed by a performer. The first evaluation data indicates a first evaluation of the performance. The first evaluation has been made by an audience who has received the performance. The method also includes obtaining second performance data. The method also includes processing the second performance data using the trained model to make an inference of a second evaluation of the second performance data. The method also includes outputting second evaluation data indicating the inference of the second evaluation.

DEEP LEARNING SYSTEM FOR DETERMINING AUDIO RECOMMENDATIONS BASED ON VIDEO CONTENT
20220414381 · 2022-12-29 ·

Embodiments are disclosed for determining an answer to a query associated with a graphical representation of data. In particular, in one or more embodiments, the disclosed systems and methods comprise receiving an input including an unprocessed audio sequence and a request to perform an audio signal processing effect on the unprocessed audio sequence. The one or more embodiments further include analyzing, by a deep encoder, the unprocessed audio sequence to determine parameters for processing the unprocessed audio sequence. The one or more embodiments further include sending the unprocessed audio sequence and the parameters to one or more audio signal processing effects plugins to perform the requested audio signal processing effect using the parameters and outputting a processed audio sequence after processing of the unprocessed audio sequence using the parameters of the one or more audio signal processing effects plugins.

Mobile App riteTune to provide music instrument players instant feedback on note pitch and rhythms accuracy based on sheet music
20220415289 · 2022-12-29 ·

A tool is needed for music instrument learners to get feedbacks on the correctness of their performances of a particular piece of music. The invention disclosed here is such a tool that can provide music instrument players instant feedback on note pitch and rhythms accuracy based on sheet music. This is accomplished through audio signal processing, sheet music image processing, and conversion of both analogue images and audio signals into standard digital music representation so a comparison can be done and hence a feedback can be presented to the player. An advanced feature will allow users to save the data to the cloud and retrieve later for comparison of progress. It also will allow user to participate an online competition with other players of the same piece of music.

INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND INFORMATION PROCESSING PROGRAM
20220406283 · 2022-12-22 ·

An information processing apparatus according to the present disclosure includes: an acquisition unit that acquires music information; an extraction unit that extracts a plurality of types of feature amounts from the music information acquired by the acquisition unit; and a generation unit that generates information in which the plurality of types of feature amounts extracted by the extraction unit is associated with predetermined identification information as music feature information to be used as learning data in composition processing using machine learning.