H04M2203/303

SYSTEM AND METHOD FOR AUTOMATED PROCESSING OF NATURAL LANGUAGE USING DEEP LEARNING MODEL ENCODING
20230025114 · 2023-01-26 ·

Automated systems and methods are provided for processing natural language, comprising obtaining first and second digitally-encoded speech representations, respectively corresponding to an agent script for and a voice recording of a telecommunication interaction; generating a similarity structure based on the speech representations, the similarity structure representing a degree of semantic similarity between the speech representations; matching markers in the first speech representation to markers in the second speech representation based on the similarity structure; and dividing the telecommunication interaction into a plurality of sections based on the matching.

Proximity based audio collaboration

A method includes: defining, by a computer device, an audio collaborative environment; defining, by the computer device, an access control of the audio collaborative environment, wherein the access control includes a geofence; receiving, by the computer device, a request from at least one user device to connect to the audio collaborative environment; determining, by the computer device, the at least one user device satisfies the access control; connecting, by the computer device, the at least one user device to an audio channel of the audio collaborative environment; recording, by the computer device, audio data transmitted on the audio channel by the at least one user device; storing, by the computer device, the audio data in a record; tagging, by the computer device, respective portions of the audio data in the record; and presenting one of the respective portions of the audio data to a user based on the tagging.

Automated Recording Highlights For Conferences

A transcript of a conference (e.g., a video conference, an audio conference, or a telephone call with two or more participants) is processed to extract a conference summary. Scores are determined for strings of the transcript that are used to select strings for inclusion in the conference summary. Determining the scores includes determining respective sentence vectors for strings. A sentence vector has elements corresponding to words in the transcript that are proportional to occurrences of the word in the string and inversely proportional to occurrences of the word in the transcript. A short video conference summary or a short audio conference summary is then generated using timestamps from the transcript associated with strings (e.g., sentences) that have been selected for inclusion in the conference summary. The short video or audio summary may be presented to users to enable efficient storage and transmission of conference information within a unified communications system.

SYSTEMS AND METHODS FOR DETECTING EMOTION FROM AUDIO FILES
20230076242 · 2023-03-09 ·

Disclosed embodiments may include a system that may receive an audio file comprising an interaction between a first user and a second user. The system may detect, using a deep neural network (DNN), moment(s) of interruption between the first and second users from the audio file. The system may extract, using the DNN, vocal feature(s) from the moment(s) of interruption. The system may determine, using a machine learning model (MLM) and based on the vocal feature(s), whether a threshold number of moments of the moment(s) of interruption corresponds to a first emotion type. When the threshold number of moments corresponds to the first emotion type, the system may transmit a first message comprising a first binary indication. When the threshold number of moments do not correspond to the first emotion type, the system may transmit a second message comprising a second binary indication.

Method and apparatus for summarization of dialogs

A method for summarizing dialogs may include obtaining an agent text stream and a customer text stream, segmenting the agent text stream and customer text stream into sentences, and labeling sentences associated with the segmented agent text stream and the segmented customer text stream. The method may further include extracting sentences from the agent text stream and the customer text stream based upon frequencies of appearance of words and terms of interest; generating an agent summary paragraph based on the extracted sentences from the agent text stream, and generating a customer summary paragraph based on the extracted sentences from the customer text stream. The method may identify keywords associated with each of the agent summary paragraph and the customer summary paragraph.

Real-time contact center speech analytics, with critical call alerts, deployed across multiple security zones
11652922 · 2023-05-16 · ·

The invention relates to systems/methods that enable real-time monitoring/processing of contact center communications to provide timely, actionable analytic insights and real-time critical call alerts, while simultaneously providing best-in-class protection of sensitive customer information.

SYSTEM AND METHOD FOR AUTOMATED ROOT CAUSE INVESTIGATION
20170372323 · 2017-12-28 ·

A system and method for investigating an issue may include defining a set of phrase categories; associating each of a set of recorded interactions with at least one phrase category; receiving a selection of a first time interval and a second time interval; selecting a reference phrase category; calculating, for at least some of the phrase categories in the set of phrase categories a correlation differential based on a trend of the phrase category and a trend of the reference phrase category and, if a correlation differential of a category is larger than a threshold then including the category in the reference category.

SYSTEMS AND METHODS OF COMMUNICATION SEGMENTS

Systems and methods include processing communication segments. A method may include obtaining at least one audio segment of an audio communication and distributing the at least one audio segment to a rules engine, analyzing, with the rules engine, the at least one audio segment for at least one acoustic trigger, determining at least one event that is triggered based on comparing the at least one acoustic trigger to one or more trigger rules, and communicating the at least one event to at least one end user through a web server.

SYSTEMS AND METHODS OF SEARCHING COMMUNICATIONS

Systems and methods for searching communications may include receiving a selection from a user of at least one of an acoustic characteristic, a language characteristic, or a category, accessing a database comprising a plurality of communications, the plurality of communications labeled with acoustic characteristics, language characteristics, and categories, restricting the plurality of communications to a subset of the plurality of communications including the user's selection of the at least one acoustic characteristic, language characteristic, or category, generating at least one of a word cloud or a tree for presenting a plurality of topics identified in the subset of the plurality of communications, and receiving a selection of one of the plurality of topics, the receiving causing a display of a further subset of the subset of the plurality of communications, the further subset having the selected topic.

SYSTEMS AND METHODS FOR IMPROVING AUDIO CONFERENCING SERVICES
20230169991 · 2023-06-01 ·

Systems and methods are disclosed herein for improving audio conferencing services. One aspect relates to processing audio content of a conference. A first audio signal is received from a first conference participant, and a start and an end of a first utterance by the first conference participant are detected from the first audio signal. A second audio signal is received from a second conference participant, and a start and an end of a second utterance by the second conference participant is detected from the second audio signal. The second conference participant is provided with at least a portion of the first utterance, wherein at least one of start time, start point, and duration is determined based at least in part on the start, end, or both, of the second utterance.