METHOD AND APPARATUS FOR AUTOMATED QUALITY MANAGEMENT OF COMMUNICATION RECORDS
20230007124 · 2023-01-05
Inventors
- Kathy Krucek (N. Barrington, IL, US)
- Filipe Plácido (Penela, PT)
- Nuno Eufrasio (Coimbra, PT)
- Rui Palma (Figueira-da-Foz, PT)
- Joao Salgado (Coimbra, PT)
- Ben Rigby (San Francisco, CA, US)
- Pedro Andrade (Coimbra, PT)
- Jason Fama (San Carlos, CA, US)
Cpc classification
G06F40/289
PHYSICS
G10L15/22
PHYSICS
G10L15/1815
PHYSICS
International classification
H04M3/51
ELECTRICITY
G06F40/289
PHYSICS
Abstract
Disclosed implementations use automated transcription and intent detection and an AI model to evaluate interactions between an agent and a customer within a call center environment. The evaluation flow used for manual evaluations is leveraged so that the evaluators can correct the AI evaluations when appropriate. Based on such corrections, the AI model can be retrained to accommodate specifics of the business and center—resulting in more confidence in the AI model over time.
Claims
1. A method for assessing communications between a user and an agent in a call center, the method comprising: extracting text from a plurality of communications between a call center user and a call center agent to thereby create a communication record; for each of the plurality of communications: assessing the corresponding text of a communication record by applying an AI assessment model to obtain an intent assessment of one or more aspects of the communication, wherein the AI assessment model is developed by processing a set of initial training data and supplemental training data to detect intents.
2. The method of claim 1, wherein the intent assessment includes a confidence score of the communication and further comprising flagging the communication record for manual quality management analysis and annotation if a confidence score of the intent assessment is below a threshold value.
3. The method of claim 2, wherein the intent assessment comprises multiple fields, each field having a value selected from a corresponding set of values and wherein the confidence level is based on a confidence sub-level determined for each value of each field.
4. The method of claim 3, wherein the fields and corresponding sets of values correspond to a human-readable form used for the manual annotation.
5. The method of claim 1, wherein the AI assessment model considers acceptable key words or phrases in each of a plurality of categories and the annotations include key words or phrases that are to be added to a category as acceptable.
6. The method of claim 1 where supplemental training data is added to the model based on reviewing manual corrections to previous assessments by the assessment model
7. The method of claim 6, wherein the supplemental data is based on manual quality analysis by a plurality of people and determining consensus between the people.
8. The method of claim 2, wherein the manual quality management analysis and annotation is accomplished by an agent in the call center.
9. The method of claim 2, wherein the manual quality management analysis and annotation includes a user interface displaying suggestions that have been marked for training from multiple models.
10. The method of claim 9 wherein the suggestions have been marked for training based on lack of confidence or explicit suggestion.
11. The method of claim 9, wherein the suggestions come from a review of an unsupervised clustering model.
12. A computer system for assessing communications between a user and an agent in a call center, the system comprising: at least one computer hardware processor; and at least one memory device operatively coupled to the at least one computer hardware processor and having instructions stored thereon which, when executed by the at least one computer hardware processor, cause the at least one computer hardware processor to carry out the method of: extracting text from a plurality of communications between a call center user and a call center agent to thereby create a communication record; for each of the plurality of communications: assessing the intent of corresponding text by applying an AI assessment model to obtain an intent assessment of the communication, wherein the AI assessment model is developed by processing a set of initial training data to detect intents.
13. The system of claim 12, wherein the intent assessment includes a confidence score of the communication and further comprising flagging the communication record for manual quality management analysis and annotation if a confidence level of the assessment is below a threshold score.
14. The system of claim 13 wherein each intent assessment comprises multiple fields, each field having a value selected from a corresponding set of values and wherein the confidence level is based on a confidence sub-level determined for each value of each field.
15. The system of claim 14, wherein the fields and corresponding sets of values correspond to a human-readable form used for the manual annotation.
16. The system of claim 12, wherein the AI assessment model considers acceptable key words or phrases in each of a plurality of categories and the annotations include key words or phrases that are to be added to a category as acceptable.
17. The system of claim 12 where supplemental training data is added to the model based on reviewing manual corrections to previous assessments by the assessment model
18. The system of claim 12, wherein the supplemental data is based on manual quality analysis by a plurality of people and determining consensus between the people.
19. The system of claim 13, wherein the manual quality management analysis and annotation is accomplished by an agent in the call center.
20. The system of claim 13, wherein the manual quality management analysis and annotation includes a user interface displaying suggestions that have been marked for training from multiple models.
21. The method of claim 20 wherein the suggestions have been marked for training based on lack of confidence or explicit suggestion.
22. The system of claim 20, wherein the suggestions come from a review of an unsupervised clustering model.
23. A method for assessing communications a contact center interaction, the method comprising: receiving communication records relating to an interaction in a contact center, wherein each communication record includes text strings extracted from the corresponding communication and wherein each call record has been designated by an AI assessment model trained to accomplish an assessment of one or more aspects of the communication records, wherein the AI assessment model is developed by processing a set of initial training data; for each communication record: displaying at least one of the text strings on a user interface in correspondence with at least one ai assessment; receiving, from a user, an assessment of the at least one text strings relating to the AI assessment; updating the communication record based on the assessment to create an updated communication record; and applying the updated communication record to the AI assessment model as supplemental training data.
24. The method of claim 15, wherein the supplemental data is based on manual quality analysis by a plurality of people and determining consensus between the people.
Description
BRIEF DESCRIPTION OF THE DRAWING
[0008] The foregoing summary, as well as the following detailed description of the invention, will be better understood when read in conjunction with the appended drawings. For the purpose of illustrating the invention, there are shown in the drawings various illustrative embodiments. It should be understood, however, that the invention is not limited to the precise arrangements and instrumentalities shown. In the drawings:
[0009]
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
DETAILED DESCRIPTION
[0016] Certain terminology is used in the following description for convenience only and is not limiting. Unless specifically set forth herein, the terms “a,” “an” and “the” are not limited to one element but instead should be read as meaning “at least one.” The terminology includes the words noted above, derivatives thereof and words of similar import.
[0017] Disclosed implementations overcome the above-identified disadvantages of the prior art by adapting contact center QM analysis to artificial intelligence systems. Disclosed implementations can leverage known methods of speech recognition and intent analysis to make corrections to inputs to be fed into an Artificial Intelligence (AI) model to be used for quality management scoring of communications. Matches with a low confidence score can be directed to a human for further review. Evaluation forms that are similar to forms used in conventional manual systems can be used. Retraining of the AI model is accomplished through individual corrections in an ongoing manner, as described below, as opposed to providing a new set of training data.
[0018] Disclosed implementations use automated transcription and intent detection and an AI model to evaluate every interaction, i.e. communication, (or alternatively a large percentage of interactions) between an agent and a customer. Disclosed implementations can leverage the evaluation flow used for manual evaluations so that the evaluators can correct the AI evaluations when appropriate. Based on such corrections, the AI model can be retrained to accommodate specifics of the business and center—resulting in more confidence in the AI model over time.
[0019]
[0020] As noted above, conventional QM forms are built by adding multiple choice questions where different choices are worth different point values. For example, a form for evaluating support interactions might start with a question where the quality of the greeting is evaluated. A good greeting where the agent introduced themselves and inquired about the problem might be worth 10 points and a poor greeting might be worth 0. There might be additional questions in the form relating to problem solving, displaying empathy, and closing. As noted above, forms can also be associated with one or more queues
[0021]
[0022] Form templates can be provided with the recommended best practice for sections, questions, and example utterances for each answer choice in order to maximize matching and increase confidence level. Customer users (admins) can edit the templates in accordance with their business needs. Additionally, users can specify a default answer choice which will be selected if none of the example utterances were detected with high confidence. In the example above, “no greeting given” might be a default answer choice, with 0 points, if a greeting is not detected. When an AI evaluation form created through UI 300 is saved, the example utterances are used to train AI model 232 (
[0023] When a voice interaction is completed, an audio recording of the interaction, created by recording module 222 (
[0024] Based on the positive or negative choices, a new evaluation of the corresponding interaction will be generated for the agent, by assessment module 230 of
[0025] Evaluations accomplished automatically by assessment module 230 are presented to the user on an evaluations page user UI 500 or results module 250 as shown in
[0026] Of course, other relevant data, such as Score (column 504), date of the interaction (column 506), queue associated with the interaction (column 508), and the like can be presented on evaluations page UI 500. Additionally, the average score, top skill, and bottom skill widgets (all results of calculations by assessment module 230 or results module 250) at the top of UI 500 could be based on taking the AI evaluations into account at a relatively low weighting (only 10% for example) as computer to forms completed manually by an evaluator employee. This weight may be configurable by the user.
[0027] When an AI form cannot be evaluated automatically and scored completely by the system (e.g., the intent/answer cannot be determined on one or more particular questions), then these evaluations will show in an AI Pending state in column 504 of
[0028] The UI can provide a single view into corrections from multiple systems that use intent detection enrichment. For example, incorrect classifications from a virtual agent or knowledge base search could also be reviewed on the UI. Real-time alerts can be provided based on real-time transcription and intent detection to notify a user immediately if an important question is being evaluated poorly by AI model 232. Emotion/crosstalk/silence checks can be added to the question choices on the forms in addition to example utterances. For example, for the AI model to detect Yes, it might have to both match the Yes intent via the example utterances and have a positive emotion based on word choice and tone.
[0029]
[0030] The reviewing user/trainer can be an agent. Corrections from multiple systems/models can be presented in the same UI view that can be used for each model. other elements of system architecture 100 (
[0031] The elements of the disclosed implementations can include computing devices including hardware processors and memories storing executable instructions to cause the processor to carry out the disclosed functionality. Numerous other general purpose or special purpose computing system environments or configurations may be used. Examples of well-known computing systems, environments, and/or configurations that may be suitable for use include, but are not limited to, personal computers, servers, handheld or laptop devices, multiprocessor systems, microprocessor-based systems, network personal computers (PCs), minicomputers, mainframe computers, embedded systems, distributed computing environments that include any of the above systems or devices, and the like. Computer-executable instructions, such as program modules, being executed by a computer may be used. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Distributed computing environments may be used where tasks are performed by remote processing devices that are linked through a communications network or other data transmission medium. In a distributed computing environment, program modules and other data may be located in both local and remote computer storage media including memory storage devices.
[0032] The computing devices can include a variety of tangible computer readable media. Computer readable media can be any available tangible media that can be accessed by device and includes both volatile and non-volatile media, removable and non-removable media. Tangible, non-transient computer storage media include volatile and non-volatile, and removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
[0033] The various data and code can be stored in electronic storage devices which may comprise non-transitory storage media that electronically stores information. The electronic storage media of the electronic storage may include one or both of system storage that is provided integrally (i.e., substantially non-removable) with the computing devices and/or removable storage that is removably connectable to the computing devices via, for example, a port (e.g., a USB port, a firewire port, etc.) or a drive (e.g., a disk drive, etc.). The electronic storage may include one or more of optically readable storage media (e.g., optical disks, etc.), magnetically readable storage media (e.g., magnetic tape, magnetic hard drive, floppy drive, etc.), electrical charge-based storage media (e.g., EEPROM, RAM, etc.), solid-state storage media (e.g., flash drive, etc.), and/or other electronically readable storage media.
[0034] Processor(s) of the computing devices may be configured to provide information processing capabilities and may include one or more of a digital processor, an analog processor, a digital circuit designed to process information, an analog circuit designed to process information, a state machine, and/or other mechanisms for electronically processing information. As used herein, the term “module” may refer to any component or set of components that perform the functionality attributed to the module. This may include one or more physical processors during execution of processor readable instructions, the processor readable instructions, circuitry, hardware, storage media, or any other components.
[0035] The contact center 150 of
[0036] It will be appreciated by those skilled in the art that changes could be made to the embodiments described above without departing from the broad inventive concept thereof. It is understood, therefore, that this invention is not limited to the particular embodiments disclosed, but it is intended to cover modifications within the spirit and scope of the present invention as defined by the appended claims.