Patent classifications
H04M2201/39
SYSTEM, METHOD AND APPARATUS FOR CONVERSATIONAL GUIDANCE
The present disclosure provides real-time, contextually appropriate behavioral guidance by utilizing machine learning models applied in real-time to call audio data. The systems and methods disclosed herein use a combination of acoustic signal processing and automatic speech recognition to convert raw call audio into features that are utilized in the machine learning models to create usable outputs to provide a user with behavioral guidance within a given context of a call or interaction with a customer in real-time.
Multi-call management system and method
A multi-call management system for allowing a user to interact with multiple calls by providing the user with multiple options for managing the calls, both inbound calls and outbound calls.
Secured switch for three-way communications
A tool for providing a secure communication channel during three-way communications. The tool initiates a three-way communication from a user device, wherein the three-way communication is a connection between a user, a third-party interpreter, and a service provider. The tool enables a passthrough function of a secured switch for transmitting non-privileged data during the three-way communication. The tool determines whether privileged data is to be transmitted during the three-way communication. Responsive to a determination that privileged data is to be transmitted during the three-way communication, the tool mutes an audio communication component. The tool activates a bypass function of the secured switch for transmitting the privileged data on an alternate audio communication component. The tool transmits the privileged data directly from the user to the service provider through the alternate audio communication component, wherein the alternate audio communication component supports unidirectional text-to-speech (TTS) audio communication.
ARTIFICIAL VENTRILOQUIST-LIKE CONTACT CENTER AGENTS
The need for efficient and effective communications is of key importance to contact centers. Agent communications with customers are designed to maximize results while minimizing resources, in particular the time required for human agents to be engaged with a particular customer. Often the impact of two agents on a communication can both improve customer satisfaction and better produce the intended result of the communication. However, two (or more) live agents is resource intensive. By providing a virtual agent controlled, entirely or in part, by a live agent, the customer may be presented with the appearance of two agents while requiring the human resources of a single agent.
REMOTE ACCESS CONTROL
A system that allows a user to access a secured area by confirming location of the user near the secured area and receipt of an access word by the system. A user in possession of a mobile device may be detected at the entrance to a secured location. The system can select a question asking for a passcode corresponding to certain access to the secured location. The question can be spoken to a user through a loudspeaker at the secured location. The user's spoken response can be processed by the system, for example using keyword spotting, to determine if the proper access word is included. If it is, the system can grant access to the secured area.
SELECTIVE PERFORMANCE OF AUTOMATED TELEPHONE CALLS TO REDUCE LATENCY AND/OR DURATION OF ASSISTANT INTERACTION
Implementations are directed to using an assistant to initiate automated telephone calls with entities. Some implementations identify an item of interest, identify a group of entities associated with the item, and initiate the calls with the entities. During a given call with a given entity, the assistant can request a status update regarding the item, and determine a temporal delay before initiating another call with the given entity to request a further status update regarding the item based on information received responsive to the request. Other implementations receive a request to perform an action on behalf of a user, identify a group of entities that can perform the action, and initiate a given call with a given entity. During the given call, the assistant can initiate an additional call with an additional entity, and generate notification(s), for the user, based on result(s) of the given call and/or the additional call.
Securing confidential information during a telecommunication session
Systems and methods disclosed herein securely provide confidential information associated with a user during an active voice call conducted using the user's mobile device. The confidential information may be provided by generating the information in spoken form and inserting the spoken information into an audio stream of the voice call. For example, a customer may be on a phone call with a customer agent. The customer agent may ask the customer for a credit card number in order to process a financial transaction. The customer may select the credit card number via an application executing on the customer's mobile phone. The application may cause the credit card number to be spoken out loud during the voice call. A voice sample of the spoken credit card number in inserted into the audio stream of the voice call and is therefore, audible to the customer and the customer agent without being overheard by a third party.
PHONE TREE TRAVERSAL SYSTEM AND METHOD
A phone tree traversal system includes an input device configured to receive a user call request, a memory, and one or more processors. The one or more processors analyze the user call request to identify an entity to call and an objective of the call, and obtain a map of a phone tree utilized by an automated call receiving system of the entity. The phone tree includes at least one node with multiple path segments that branch from the node, and each node includes a respective prompt. The one or more processors determine a route along the map to an endpoint of the phone tree associated with the objective. During the call, the one or more processors navigate the phone tree to reach the endpoint by submitting information in response to the respective prompt at each node along the route that is determined.
Technologies for incorporating an augmented voice communication into a communication routing configuration
A method for incorporating an augmented voice communication into a communication routing configuration of a contact center system according to an embodiment includes selecting a vocal avatar, wherein the vocal avatar includes phonetic characteristics having first values, receiving a text communication and input user parameters from an input user, generating the augmented voice communication based on the text communication and the input user parameters, wherein the augmented voice communication includes phonetic characteristics having second values, wherein the first values of the phonetic characteristics of the vocal avatar are different from the second values of the phonetic characteristics of the augmented voice communication, and incorporating the augmented voice communication into the communication routing configuration of the contact center system.
SWITCHING BETWEEN SPEECH RECOGNITION SYSTEMS
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining an availability of revoiced transcription units in a transcription system and in response to establishment of the communication session, selecting, based on the availability of revoiced transcription units, a revoiced transcription unit instead of a non-revoiced transcription unit to generate a transcript of the first audio data. The method may also include obtaining revoiced audio generated by a revoicing of the first audio data by a captioning assistant and generating a transcription of the revoiced audio using an automatic speech recognition system. The method may further include in response to selecting the revoiced transcription unit, directing the transcription of the revoiced audio to the second device as the transcript of the first audio data.