Patent classifications
H04M2203/252
VOICE MESSAGE DISPLAY
Aspects of this disclosure are directed to a method and a terminal device, and a computer-readable storage medium. The terminal device includes processing circuitry that obtains n voice messages from at least one user account. n is a positive integer. The processing circuitry displays a voice message presentation interface. The voice message presentation interface is configured to display virtual characters corresponding to the n voice messages in a virtual world. Locations of the n voice messages in the virtual world are based on whether the n voice messages are unplayed. The virtual characters can be based on respective message attributes of the n voice messages.
Dynamic rewardable activity value determination and allocation
A Sender Controlled Media (SCM) platform user is presented incentives for performing rewardable platform activities that result in accruing points in a user-specific account from which points can be redeemed by transfer or cryptocurrency conversion to an external user-designated wallet. A reward received for each instance of a rewardable platform activity is determined based on a range of platform context including a target average points payout, an aggregated total points payout over a distribution period, and a type of rewardable platform activity.
Securing confidential information during a telecommunication session
Systems and methods disclosed herein securely provide confidential information associated with a user during an active voice call conducted using the user's mobile device. The confidential information may be provided by generating the information in spoken form and inserting the spoken information into an audio stream of the voice call. For example, a customer may be on a phone call with a customer agent. The customer agent may ask the customer for a credit card number in order to process a financial transaction. The customer may select the credit card number via an application executing on the customer's mobile phone. The application may cause the credit card number to be spoken out loud during the voice call. A voice sample of the spoken credit card number in inserted into the audio stream of the voice call and is therefore, audible to the customer and the customer agent without being overheard by a third party.
Microphone State Display Method And Microphone State Display Apparatus
A microphone state display method includes receiving a mute-on or a mute-off operation by each of a plurality of microphones, displaying a state of a microphone that has received the mute-off operation as a first state on a display, when receiving the mute-on operation, in a case in which at least one microphone among the plurality of microphones is in a mute-off state, displaying a state of a microphone that has received the mute-on operation as a second state on a display, and, when receiving the mute-on operation in a case in which all of the plurality of microphones are in a mute-on state, displaying the state of the microphone that has received the mute-on operation as a third state on a display.
Voice message display method and apparatus in application, computer device, and computer-readable storage medium
Aspects of this disclosure are directed to a voice message display method and apparatus in an application, a computer device, and a computer-readable storage medium. The method can be performed by a terminal on which an application is installed and is capable of receiving a voice message. The method can include starting an application, and obtaining n voice messages published by at least one user account. The method can further include displaying a voice message presentation interface of the application, where the voice message presentation interface displays the voice message in a virtual world and the voice message are displayed by using a visible element in the virtual world as a carrier.
Determination and visual display of spoken menus for calls
Implementations relate to determination and visual display of spoken menus for calls. In some implementations, a computer-implemented method includes receiving audio data output in a call between a call device and a device associated with a target entity. The audio data includes speech indicating one or more selection options for a user of the call device to navigate through a call menu provided by the target entity in the call. Text is determined by programmatically analyzing the audio data, the text representing the speech. The selection options are determined based on programmatically analyzing at least one of the text or the audio data. At least a portion of the text is displayed by the call device during the call, as one or more visual options that correspond to the selection options. The visual options are each selectable via user input to cause corresponding navigation through the call menu.
Media channel management apparatus for network communications sessions
A system includes: a computerized appliance having a processor; and software executing on the computerized appliance from a non-transitory physical medium, the software providing: establishment of a communication session over a network between two or more end communications appliances; detection and recording of at least media channel capabilities on individual end communications appliances connected in the session; service of media channel toggle options to end communications appliances; detection of media channel toggle option selections made by users operating the end communications appliances; and implementation of the media channel toggle option or options selected for individual communications sessions.
AUTOMATICALLY IDENTIFYING SPEAKERS IN REAL-TIME THROUGH MEDIA PROCESSING WITH DIALOG UNDERSTANDING SUPPORTED BY AI TECHNIQUES
Automatically identifying speakers in real-time through media processing with dialog understanding. A plurality of audio streams may be received, an audio stream representing a speech of a participant speaking during an online meeting. A voice characteristic of a voice corresponding to the speech of the participant in the audio stream may be determined. The plurality of audio streams may be converted into text and a natural language processing may be performed to determine content context of the dialog. The natural language processing infers a name to associate with the voice in the audio stream based on the determined content context. A data structure linking the name with the voice may be created and stored in a knowledge base. A user interface associated with the online meeting application is triggered to present the name or identity of the speaker.
TRANSCRIPTION OF COMMUNICATIONS
A system may include a camera configured to obtain an image of a user, at least one processor, and at least one non-transitory computer-readable media communicatively coupled to the at least one processor. The non-transitory computer-readable media configured to store one or more instructions that when executed cause or direct the system to perform operations. The operations may include establish a communication session between the system and a device. The communication session may be configured such that the device provides audio for the system. The operations may further include compare the image to a particular user image associated with the system and select a first method of transcription generation from among two or more methods of transcription generation based on the comparison of the image to the particular user image. The operations may also include present, a transcription of the audio generated using the selected first method of transcription generation.
Information processing system, information processing terminal, and control method and control program of information processing terminal
An information processing system includes first and second terminals for acquiring a desired image captured actively using a speech communication partner's apparatus during speech communication. The first terminal includes a first microphone, a first transmitter that transmits voice data to the second terminal, and a first receiver that receives voice data from the second terminal. The second terminal includes a camera, a second microphone, a second transmitter that transmits the voice data to the first terminal, and a second receiver that receives the voice data from the first terminal. During speech communication, the first transmitter transmits, together with the voice data, an operation command for operating the camera. The second receiver receives the operation command and operates the camera. During speech communication with the first terminal, the second transmitter transmits, together with the voice data, the image data captured by the camera.