Patent classifications
G10L17/14
AUTOCORRECTION OF PRONUNCIATIONS OF KEYWORDS IN AUDIO/VIDEOCONFERENCES
The present disclosure relates to automatically correcting mispronounced keywords during a conference session. More particularly, the present invention provides methods and systems for automatically correcting audio data generated from audio input having indications of mispronounced keywords during an audio/videoconferencing system. In some embodiments, the process of automatically correcting the audio data may require a re-encoding process of the audio data at the conference server. In alternative embodiments, the process may require updating the audio data at the receiver end of the conferencing system.
METHOD AND DEVICE FOR GENERATING SPEECH VIDEO USING AUDIO SIGNAL
A device according to an embodiment has one or more processors and a memory storing one or more programs executable by the one or more processors. The device includes a first encoder configured to receive a person background image corresponding to a video part of a speech video of a person and extract an image feature vector from the person background image, a second encoder configured to receive a speech audio signal corresponding to an audio part of the speech video and extract a voice feature vector from the speech audio signal, a combiner configured to generate a combined vector by combining the image feature vector output from the first encoder and the voice feature vector output from the second encoder, and a decoder configured to reconstruct the speech video of the person using the combined vector as an input.
METHOD AND DEVICE FOR GENERATING SPEECH VIDEO USING AUDIO SIGNAL
A device according to an embodiment has one or more processors and a memory storing one or more programs executable by the one or more processors. The device includes a first encoder configured to receive a person background image corresponding to a video part of a speech video of a person and extract an image feature vector from the person background image, a second encoder configured to receive a speech audio signal corresponding to an audio part of the speech video and extract a voice feature vector from the speech audio signal, a combiner configured to generate a combined vector by combining the image feature vector output from the first encoder and the voice feature vector output from the second encoder, and a decoder configured to reconstruct the speech video of the person using the combined vector as an input.
INTERACTION INFORMATION PROCESSING METHOD AND APPARATUS, DEVICE, AND MEDIUM
An interaction information processing method and apparatus, a device, and a medium are provided. The method includes: collecting voice data of at least one participating user in an interaction conducted by users on a real-time interactive interface; determining, based on the voice data, a source language type used by each of the at least one participating user; converting the voice data of the at least one participating user from the source language type to a target language type, to obtain translation data; and displaying the translation data on a target client device.
INTERACTION INFORMATION PROCESSING METHOD AND APPARATUS, DEVICE, AND MEDIUM
An interaction information processing method and apparatus, a device, and a medium are provided. The method includes: collecting voice data of at least one participating user in an interaction conducted by users on a real-time interactive interface; determining, based on the voice data, a source language type used by each of the at least one participating user; converting the voice data of the at least one participating user from the source language type to a target language type, to obtain translation data; and displaying the translation data on a target client device.
SYSTEMS AND METHODS TO ANALYZE AUDIO DATA TO IDENTIFY DIFFERENT SPEAKERS
A computing system may receive data representing dialog between persons, the data representing words spoken by at least first and second speakers, determine an intent of a speaker for a first portion of the data, the intent being indicative of an identity of the first or second speaker for the first portion of the data or another portion of the data different than the first portion, determine a name of the first or second speaker represented in the first portion of the data based at least in part on the determined intent, and output an indication of the determined name so that the indication identifies the first portion of the data or the another portion of the data with the first or second speaker.
SYSTEMS AND METHODS TO ANALYZE AUDIO DATA TO IDENTIFY DIFFERENT SPEAKERS
A computing system may receive data representing dialog between persons, the data representing words spoken by at least first and second speakers, determine an intent of a speaker for a first portion of the data, the intent being indicative of an identity of the first or second speaker for the first portion of the data or another portion of the data different than the first portion, determine a name of the first or second speaker represented in the first portion of the data based at least in part on the determined intent, and output an indication of the determined name so that the indication identifies the first portion of the data or the another portion of the data with the first or second speaker.
SPEECH CONTROL METHOD AND APPARATUS, ELECTRONIC DEVICE AND STORAGE MEDIUM
The disclosure provides a speech control method, an electronic device and a storage medium. The method includes: obtaining a speech to be processed; obtaining a speech feature vector by performing feature analysis on the speech to be processed; determining whether the speech to be processed belongs to a target type based on the speech feature vector; and in response to the speech to be processed belonging to the target type, performing wake-up control on a target device based on the speech to be processed.
SPEECH CONTROL METHOD AND APPARATUS, ELECTRONIC DEVICE AND STORAGE MEDIUM
The disclosure provides a speech control method, an electronic device and a storage medium. The method includes: obtaining a speech to be processed; obtaining a speech feature vector by performing feature analysis on the speech to be processed; determining whether the speech to be processed belongs to a target type based on the speech feature vector; and in response to the speech to be processed belonging to the target type, performing wake-up control on a target device based on the speech to be processed.
Voice monitoring system and voice monitoring method
A recording device records a video and an imaging time, and a voice. Based on the voice, a sound parameter calculator calculates a sound parameter for specifying magnitude of the voice in a monitoring area at the imaging time for each of pixels and for each of certain times. A sound parameter storage unit stores the sound parameter. A sound parameter display controller superimposes a voice heat map on a captured image of the monitoring area and displays the superimposed image on a monitor. At this time, the sound parameter display controller displays the voice heat map based on a cumulative time value of magnitude of the voice, according to designation of a time range.