Patent classifications
G06F40/117
EXTRACTION OF TASKS FROM DOCUMENTS USING WEAKLY SUPERVISION
This disclosure relates to extraction of tasks from documents based on a weakly supervised classification technique, wherein extraction of tasks is identification of mentions of tasks in a document. There are several prior arts addressing the problem of extraction of events, however due to crucial distinctions between events-tasks, task extraction stands as a separate problem. The disclosure explicitly defines specific characteristics of tasks, creates labelled data at a word-level based on a plurality of linguistic rules to train a word-level weakly supervised model for task extraction. The labelled data is created based on the plurality of linguistic rules for a non-negation aspect, a volitionality aspect, an expertise aspect and a plurality of generic aspects. Further the disclosure also includes a phrase expansion technique to capture the complete meaning expressed by the task instead of merely mentioning the task that may not capture the entire meaning of the sentence.
EXTRACTING ENGAGING QUESTIONS FROM A COMMUNICATION SESSION
Methods and systems provide for extracting engaging questions from a communication session. In one embodiment, the system connects to a communication session with a number of participants; receives a transcript of a conversation between the participants produced during the communication session; extracts, from the transcript, utterances including one or more sentences spoken by the participants; identifies a subset of the utterances spoken by a subset of the participants associated with a prespecified organization; extracts engaging questions within the subset of utterances, the engaging questions each including a question asked by the participant associated with the organization that is immediately answered in the following utterance by a participant not associated with the organization; and presents, for display at one or more client devices, data corresponding to the extracted engaging questions.
EXTRACTING ENGAGING QUESTIONS FROM A COMMUNICATION SESSION
Methods and systems provide for extracting engaging questions from a communication session. In one embodiment, the system connects to a communication session with a number of participants; receives a transcript of a conversation between the participants produced during the communication session; extracts, from the transcript, utterances including one or more sentences spoken by the participants; identifies a subset of the utterances spoken by a subset of the participants associated with a prespecified organization; extracts engaging questions within the subset of utterances, the engaging questions each including a question asked by the participant associated with the organization that is immediately answered in the following utterance by a participant not associated with the organization; and presents, for display at one or more client devices, data corresponding to the extracted engaging questions.
EXTRACTING FILLER WORDS AND PHRASES FROM A COMMUNICATION SESSION
Methods and systems provide for extracting filler words and phrases from a communication session. In one embodiment, the system receives a transcript of a conversation involving one or more participants produced during a communication session; extracts, from the transcript, utterances including one or more sentences spoken by the participants; identifies a subset of the utterances spoken by a subset of the participants associated with a prespecified organization; extracts filler phrases within the subset of utterances, the filler phrases each comprising one or more words representing disfluencies within a sentence, where extracting the filler phrases includes applying filler detection rules; and presents, for display at one or more client devices, data corresponding to the extracted filler phrases.
EXTRACTING FILLER WORDS AND PHRASES FROM A COMMUNICATION SESSION
Methods and systems provide for extracting filler words and phrases from a communication session. In one embodiment, the system receives a transcript of a conversation involving one or more participants produced during a communication session; extracts, from the transcript, utterances including one or more sentences spoken by the participants; identifies a subset of the utterances spoken by a subset of the participants associated with a prespecified organization; extracts filler phrases within the subset of utterances, the filler phrases each comprising one or more words representing disfluencies within a sentence, where extracting the filler phrases includes applying filler detection rules; and presents, for display at one or more client devices, data corresponding to the extracted filler phrases.
Synchronization and tagging of image and text data
A computing system accesses an image-based document and a text document having text extracted from the image-based document and provides a user interface displaying at least a portion of the image-based document. In response to selection of a text portion of the image-based document, the system determines an occurrence of the text portion within at least a portion of the image-based document and then applies a search model on the text document to identify the same occurrence of the text portion. Once matched, alignment data indicating a relationship between a selected tag and both the text portion of the image-based document and the text portion of the text document is stored.
Synchronization and tagging of image and text data
A computing system accesses an image-based document and a text document having text extracted from the image-based document and provides a user interface displaying at least a portion of the image-based document. In response to selection of a text portion of the image-based document, the system determines an occurrence of the text portion within at least a portion of the image-based document and then applies a search model on the text document to identify the same occurrence of the text portion. Once matched, alignment data indicating a relationship between a selected tag and both the text portion of the image-based document and the text portion of the text document is stored.
System and method of highlighting influential samples in sequential analysis
Attention weights in a hierarchical attention network indicate the relative importance of portions of a conversation between an individual at one terminal and a computer or a human agent at another terminal. Weighting the portions of the conversation after converting the conversation to a standard text format allows for a computer to graphically highlight, by color, font, or other indicator visible on a graphical user interface, which portions of a conversation led to an escalation of the interaction from an intelligent virtual assistant to a human customer service agent.
System and method of highlighting influential samples in sequential analysis
Attention weights in a hierarchical attention network indicate the relative importance of portions of a conversation between an individual at one terminal and a computer or a human agent at another terminal. Weighting the portions of the conversation after converting the conversation to a standard text format allows for a computer to graphically highlight, by color, font, or other indicator visible on a graphical user interface, which portions of a conversation led to an escalation of the interaction from an intelligent virtual assistant to a human customer service agent.
Form text extraction of key/value pairs
A computer-implemented method, apparatus and program product use the spatial locations of words identified in an unstructured document to both reconstruct lines in the unstructured document and vertically partition the unstructured document. Key/value pairs may then be generated from one or more of the reconstructed lines by using one or more words to one side of the vertical partition as keys and using one or more words to the other side of the vertical partition as values.