Patent classifications
H04M2242/12
System and method for multilingual authentication access to communication system in controlled environment
A system and method for multilingual authentication access to a communication system in a controlled environment is disclosed. The system includes a communication subsystem that connects a communication between a first communication device and a second communication device, wherein one of the first communication device or the second communication device is used by the inmate of the correctional system. The system also includes a profile subsystem configured to determine profile information of a party associated with the first communication device. The system further includes an authentication subsystem that provides one or more challenge questions to the first communication device based on the profile information, receives at least one response to the one or more challenge questions from the first communication device, and authorizes connection of the communication between the first communication device and the second communication device based on the at least one response.
Auto-translation for multi user audio and video
The disclosed subject matter provides a system, computer readable storage medium, and a method providing an audio and textual transcript of a communication. A conferencing services may receive audio or audio visual signals from a plurality of different devices that receive voice communications from participants in a communication, such as a chat or teleconference. The audio signals representing voice (speech) communications input into respective different devices by the participants. A translation services server may receive over a separate communication channel the audio signals for translation into a second language. As managed by the translation services server, the audio signals may be converted into textual data. The textual data may be translated into text of different languages based the language preferences of the end user devices in the teleconference. The translated text may be further translated into audio signals.
TRANSFERRING INFORMATION FROM A SENDER TO A RECIPIENT DURING A TELEPHONE CALL UNDER NOISY ENVIRONMENT
A method for transferring information from a sender to a recipient during a telephone call between a sender communication apparatus and a recipient communication apparatus is provided. The method comprises detecting, by the recipient communication apparatus, that the recipient is subjected to a noisy environment; converting a voice message, originating from the senders voice and recorded by the sender communication apparatus, into a text message; and displaying, at a display of the recipient communication apparatus, the text message to the recipient.
PROMPT FEATURE TO LEAVE VOICEMAIL FOR APPROPRIATE ATTRIBUTE-BASED CALL BACK TO CUSTOMERS
Systems and methods are provided for attribute-based client callbacks. A client is prompted to leave a voice message. Attributes are extracted from the voice message and, based on the attributes, tokens created for the selection of an appropriate agent is connected to the client, such as having skills or attributes matching one or more tokens. A callback application server transmits prompts and receives requests for client callbacks. an interaction manager determines agent availability and arranges callback handling, and a session management server initiates callbacks to connect the selected agent with the client.
Methods, Systems, and Devices for Presenting an Audio Difficulties User Actuation Target in an Audio or Video Conference
A conferencing system terminal device includes a display, an audio output, a user interface, a communication device, and one or more processors. The one or more processors present an audio difficulties user actuation target upon the display during an audio or video conference occurring across a network and concurrently with a presentation of conference content. Actuation of the audio difficulties user actuation target indicates that audio content associated with the audio or video conference being delivered by the audio output is impaired.
Active transport based notifications
Techniques for triggering an action associated with a task items are provided. A task item associated with a triggering criterion is provided. The triggering criterion requires an occurrence of any of a plurality of communication events. The plurality of communication events includes a telephone call. In some implementations, at least two of the plurality of communication events are of different communication types. In some implementations, the different types are telephone calls, emails, and text messages. The occurrence of a first communication event of the plurality of communication events is detected. It is determined the triggering criterion is satisfied. In response to the determining, a notification associated with the task item is caused to be presented to a user of the electronic device.
PROCESSING SYSTEM, PROCESSING METHOD, AND NON-TRANSITORY STORAGE MEDIUM
The example embodiments provides a processing system (10) including: an acquisition unit (11) that acquires target speech data in which a target speech is recorded or a target feature value that indicates a feature of the target speech; an inference unit (12) that infers a language of the target speech, based on an inference model for inferring a language of a speech from speech data or a speech feature value and the target speech data or the target feature value; a result output unit (13) that outputs an inference result by the inference unit (12); a determination unit (14) that determines whether the inference result is correct; and a learning data output unit (15) that outputs the inference result determined to be correct by the determination unit (14) and the target speech data or the target feature value, as learning data for generating the inference model.
Systems and methods for providing and using translation-enabled multiparty communication sessions
Systems and methods for providing a translation-enabled multiparty communication session utilize a translation service that is configured to receive individual audio streams from each participant in the communication session, and which then generates a participant text stream for each participant based on their audio stream. This can be accomplished using speech-to-text techniques. Each participant's text stream is then translated into a different language to create a translated participant text stream. Individual participants in the communication session can then subscribe to the translated participant text streams to view a translation of what each participant is saying during the communication session.
PROCESSING SYSTEM, PROCESSING METHOD, AND NON-TRANSITORY STORAGE MEDIUM
The example embodiments provides a processing system (10) including: an acquisition unit (11) that acquires target speech data in which a target speech is recorded or a target feature value that indicates a feature of the target speech; an inference unit (12) that infers a language of the target speech, based on an inference model for inferring a language of a speech from speech data or a speech feature value and the target speech data or the target feature value; a result output unit (13) that outputs an inference result by the inference unit (12); a determination unit (14) that determines whether the inference result is correct; and a learning data output unit (15) that outputs the inference result determined to be correct by the determination unit (14) and the target speech data or the target feature value, as learning data for generating the inference model.
Processing system, processing method, and non-transitory storage medium
The example embodiments provides a processing system (10) including: an acquisition unit (11) that acquires target speech data in which a target speech is recorded or a target feature value that indicates a feature of the target speech; an inference unit (12) that infers a language of the target speech, based on an inference model for inferring a language of a speech from speech data or a speech feature value and the target speech data or the target feature value; a result output unit (13) that outputs an inference result by the inference unit (12); a determination unit (14) that determines whether the inference result is correct; and a learning data output unit (15) that outputs the inference result determined to be correct by the determination unit (14) and the target speech data or the target feature value, as learning data for generating the inference model.