G10L21/12

REAL-TIME GENERATION OF SPEECH ANIMATION
20220108510 · 2022-04-07 ·

To realistically animate a String (such as a sentence) a hierarchical search algorithm is provided to search for stored examples (Animation Snippets) of sub-strings of the String, in decreasing order of sub-string length, and concatenate retrieved sub-strings to complete the String of speech animation. In one embodiment, real-time generation of speech animation uses model visemes to predict the animation sequences at onsets of visemes and a look-up table based (data-driven) algorithm to predict the dynamics at transitions of visemes. Specifically posed Model Visemes may be blended with speech animation generated using another method at corresponding time points in the animation when the visemes are to be expressed. An Output Weighting Function is used to map Speech input and Expression input into Muscle-Based Descriptor weightings

REAL-TIME GENERATION OF SPEECH ANIMATION
20220108510 · 2022-04-07 ·

To realistically animate a String (such as a sentence) a hierarchical search algorithm is provided to search for stored examples (Animation Snippets) of sub-strings of the String, in decreasing order of sub-string length, and concatenate retrieved sub-strings to complete the String of speech animation. In one embodiment, real-time generation of speech animation uses model visemes to predict the animation sequences at onsets of visemes and a look-up table based (data-driven) algorithm to predict the dynamics at transitions of visemes. Specifically posed Model Visemes may be blended with speech animation generated using another method at corresponding time points in the animation when the visemes are to be expressed. An Output Weighting Function is used to map Speech input and Expression input into Muscle-Based Descriptor weightings

Displaying enhancement items associated with an audio recording

Systems, methods, and software are disclosed herein for displaying visual representations of entities associated with an audio recording. A waveform associated with an audio recording is presented in a user interface to an application. A set of entities associated with the audio recording are then identified. Responsive to identifying the set of entities, a set of enhancement items associated with the set of entities is presented. In response to a selection of a given one of the enhancement items, a visual representation of an associated one of the entities in the user interface to the application.

Displaying enhancement items associated with an audio recording

Systems, methods, and software are disclosed herein for displaying visual representations of entities associated with an audio recording. A waveform associated with an audio recording is presented in a user interface to an application. A set of entities associated with the audio recording are then identified. Responsive to identifying the set of entities, a set of enhancement items associated with the set of entities is presented. In response to a selection of a given one of the enhancement items, a visual representation of an associated one of the entities in the user interface to the application.

Processing Audio Information
20210249032 · 2021-08-12 ·

A method for capturing, recording, playing back, visually representing, storing and processing of audio signals, comprises converting the audio signal into a video that pairs the audio with a visual representation of the audio data where such visual representation may contain the waveform, relevant text, spectrogram, wavelet decomposition, or other transformation of the audio data in such a way that the viewer can identify which part of the visual representation is associated with the currently playing audio signal.

Processing Audio Information
20210249032 · 2021-08-12 ·

A method for capturing, recording, playing back, visually representing, storing and processing of audio signals, comprises converting the audio signal into a video that pairs the audio with a visual representation of the audio data where such visual representation may contain the waveform, relevant text, spectrogram, wavelet decomposition, or other transformation of the audio data in such a way that the viewer can identify which part of the visual representation is associated with the currently playing audio signal.

Audio Techniques for Music Content Generation
20210247954 · 2021-08-12 ·

Techniques are disclosed relating to implementing audio techniques for real-time audio generation. For example, a music generator system may generate new music content from playback music content based on different parameter representations of an audio signal. In some cases, an audio signal can be represented by both a graph of the signal (e.g., an audio signal graph) relative to time and a graph of the signal relative to beats (e.g., a signal graph). The signal graph is invariant to tempo, which allows for tempo invariant modification of audio parameters of the music content in addition to tempo variant modifications based on the audio signal graph.

Audio Techniques for Music Content Generation
20210247954 · 2021-08-12 ·

Techniques are disclosed relating to implementing audio techniques for real-time audio generation. For example, a music generator system may generate new music content from playback music content based on different parameter representations of an audio signal. In some cases, an audio signal can be represented by both a graph of the signal (e.g., an audio signal graph) relative to time and a graph of the signal relative to beats (e.g., a signal graph). The signal graph is invariant to tempo, which allows for tempo invariant modification of audio parameters of the music content in addition to tempo variant modifications based on the audio signal graph.

Listener-Defined Controls for Music Content Generation
20210247955 · 2021-08-12 ·

Techniques are disclosed relating to implementing user-created controls to modify music content. A music generator system may be configured to automatically generate output music content by selecting and combining audio tracks based on various parameters. Users may create their own control elements that the music generator system may train (e.g., using AI techniques) to generate output music content according to a user's intended functionality of a user-created control element.

Block-Chain Ledger Based Tracking of Generated Music Content
20210248213 · 2021-08-12 ·

Techniques are disclosed relating to tracking contributions to composed music content. In some embodiments, a computer system determines playback data for a music content mix, where the playback data indicates characteristics of playback of the music content mix and the music content mix includes a determined combination of multiple audio tracks. In some embodiments, the system records, in an electronic block-chain ledger data structure, information specifying individual playback data for one or more of the multiple audio tracks in the music content mix. The information specifying individual playback data for an individual audio track may include usage data for the individual audio track and signature information associated with the individual audio track.