G10L25/15

Speaker recognition with assessment of audio frame contribution

This application describes methods and apparatus for speaker recognition. An apparatus according to an embodiment has an analyzer (202) for analyzing each frame of a sequence of frames of audio data (A.sub.IN) which correspond to speech sounds uttered by a user to determine at least one characteristic of the speech sound of that frame. An assessment module (203) determines, for each frame of audio data, a contribution indicator of the extent to which the frame of audio data should be used for speaker recognition processing based on the determined characteristic of the speech sound. In this way frames which correspond to speech sounds that are of most use for speaker discrimination may be emphasized and/or frames which correspond to speech sounds that are of least use for speaker discrimination may be de-emphasized.

COGNITIVE FUNCTION EVALUATION DEVICE, COGNITIVE FUNCTION EVALUATION SYSTEM, AND COGNITIVE FUNCTION EVALUATION METHOD

A cognitive function evaluation device includes: an obtainment unit configured to obtain speech data indicating speech uttered by a subject; a calculation unit configured to extract a plurality of vowels from the speech data obtained by the obtainment unit, and calculate, for each of the plurality of vowels, a feature value based on a frequency and an amplitude of at least one formant obtained from a spectrum of the vowel; an evaluation unit configured to evaluate a cognitive function of the subject from the feature value calculated by the calculation unit; and an output unit configured to output an evaluation result of the evaluation unit.

COMMUNICATION SYSTEM FOR PROCESSING AUDIO INPUT WITH VISUAL DISPLAY
20200227066 · 2020-07-16 ·

A reference acoustic input is processed into a quantization representation such that the quantization representation comprises acoustic components determined from the reference acoustic input, wherein the acoustic components comprise amplitude, rhythm, and pitch frequency of the reference acoustic input. A visual representation is generated that simultaneously depicts the acoustic components comprising amplitude, rhythm, and pitch frequency of the reference acoustic input. A user spoken input may be received and similarly processed and displayed.

COMMUNICATION SYSTEM FOR PROCESSING AUDIO INPUT WITH VISUAL DISPLAY
20200227066 · 2020-07-16 ·

A reference acoustic input is processed into a quantization representation such that the quantization representation comprises acoustic components determined from the reference acoustic input, wherein the acoustic components comprise amplitude, rhythm, and pitch frequency of the reference acoustic input. A visual representation is generated that simultaneously depicts the acoustic components comprising amplitude, rhythm, and pitch frequency of the reference acoustic input. A user spoken input may be received and similarly processed and displayed.

CONCEPT FOR ENCODING AN AUDIO SIGNAL AND DECODING AN AUDIO SIGNAL USING DETERMINISTIC AND NOISE LIKE INFORMATION

An encoder for encoding an audio signal has: an analyzer configured for deriving prediction coefficients and a residual signal from an unvoiced frame of the audio signal; a gain parameter calculator configured for calculating a first gain parameter information for defining a first excitation signal related to a deterministic codebook and for calculating a second gain parameter information for defining a second excitation signal related to a noise-like signal for the unvoiced frame; and a bitstream former configured for forming an output signal based on an information related to a voiced signal frame, the first gain parameter information and the second gain parameter information.

CONCEPT FOR ENCODING AN AUDIO SIGNAL AND DECODING AN AUDIO SIGNAL USING DETERMINISTIC AND NOISE LIKE INFORMATION

An encoder for encoding an audio signal has: an analyzer configured for deriving prediction coefficients and a residual signal from an unvoiced frame of the audio signal; a gain parameter calculator configured for calculating a first gain parameter information for defining a first excitation signal related to a deterministic codebook and for calculating a second gain parameter information for defining a second excitation signal related to a noise-like signal for the unvoiced frame; and a bitstream former configured for forming an output signal based on an information related to a voiced signal frame, the first gain parameter information and the second gain parameter information.

Method of and system for spotting digital media objects and event markers using musical experience descriptors to characterize digital music to be automatically composed and generated by an automated music composition and generation engine
10672371 · 2020-06-02 · ·

An automated music composition and generation system and process for scoring a selected media object or event marker, with one or more pieces of digital music, by spotting the selected media object or event marker with musical experience descriptors selected and applied to the selected media object or event marker by the system user during a scoring process, and using said selected musical experience descriptors to drive an automated music composition and generation engine to automatically compose and generate the one or more pieces of digital music.

Method of and system for spotting digital media objects and event markers using musical experience descriptors to characterize digital music to be automatically composed and generated by an automated music composition and generation engine
10672371 · 2020-06-02 · ·

An automated music composition and generation system and process for scoring a selected media object or event marker, with one or more pieces of digital music, by spotting the selected media object or event marker with musical experience descriptors selected and applied to the selected media object or event marker by the system user during a scoring process, and using said selected musical experience descriptors to drive an automated music composition and generation engine to automatically compose and generate the one or more pieces of digital music.

METHOD OF AND SYSTEM FOR AUTOMATICALLY GENERATING MUSIC COMPOSITIONS AND PRODUCTIONS USING LYRICAL INPUT AND MUSIC EXPERIENCE DESCRIPTORS
20200168187 · 2020-05-28 · ·

An automated music composition and generation process within an automated music composition and generation system driven by lyrical musical experience descriptors. The process involves the system user accessing said automated music composition and generation system, employing an automated music composition and generation engine having a system user interface. The system user interface is used to select and provide musical experience descriptors, including lyrics, to the automated music composition and generation engine for processing by said automated music composition and generation engine. The system user initiates the automated music composition and generation engine to compose and generate music based on the musical experience descriptors and lyrics provided.

METHOD OF AND SYSTEM FOR AUTOMATICALLY GENERATING MUSIC COMPOSITIONS AND PRODUCTIONS USING LYRICAL INPUT AND MUSIC EXPERIENCE DESCRIPTORS
20200168187 · 2020-05-28 · ·

An automated music composition and generation process within an automated music composition and generation system driven by lyrical musical experience descriptors. The process involves the system user accessing said automated music composition and generation system, employing an automated music composition and generation engine having a system user interface. The system user interface is used to select and provide musical experience descriptors, including lyrics, to the automated music composition and generation engine for processing by said automated music composition and generation engine. The system user initiates the automated music composition and generation engine to compose and generate music based on the musical experience descriptors and lyrics provided.