G10L17/14

Reducing bandwidth requirements of virtual collaboration sessions

A computer-implemented method, a computer system and a computer program product reduce bandwidth requirements of a virtual collaboration session. The method includes capturing session data from a virtual collaboration session. The session data is selected from a group consisting of video data, audio data, an image of a screen of a connected device and text data. The method also includes connecting to a live blog platform. The method further includes transmitting a text transcription of the virtual collaboration session to the live blog platform. The text transcription is generated by scanning the audio data using a speech-to-text algorithm. In addition, the method includes classifying a topic in the virtual collaboration session based on importance. Lastly, the method includes transmitting a multimedia file related to the topic to the live blog platform in response to the topic being classified as important. The multimedia file is extracted from the session data.

COMPUTER-BASED TECHNIQUES FOR VISUALLY NARRATING RECORDED MEETING CONTENT

In various embodiments, a meeting narration application generates visualizations of recorded meeting data. The meeting narration application generates a first visualization of a set of parameters based on a set of transcript sentences associated with the recorded meeting data. The meeting narration application displays the first visualization and a first expanded content visualization of a first transcript sentence included in the set of transcript sentences within a graphical user interface (GUI). Subsequently, the meeting narration application receives a user event associated with the first visualization via the GUI. The meeting narration application modifies a first parameter selection associated with the set of parameters based on the user event to generate a modified parameter selection. Based on the modified parameter selection, the meeting narration application displays a first compressed content visualization of the first transcript sentence within the GUI.

COMPUTER-BASED TECHNIQUES FOR VISUALLY NARRATING RECORDED MEETING CONTENT

In various embodiments, a meeting narration application generates visualizations of recorded meeting data. The meeting narration application generates a first visualization of a set of parameters based on a set of transcript sentences associated with the recorded meeting data. The meeting narration application displays the first visualization and a first expanded content visualization of a first transcript sentence included in the set of transcript sentences within a graphical user interface (GUI). Subsequently, the meeting narration application receives a user event associated with the first visualization via the GUI. The meeting narration application modifies a first parameter selection associated with the set of parameters based on the user event to generate a modified parameter selection. Based on the modified parameter selection, the meeting narration application displays a first compressed content visualization of the first transcript sentence within the GUI.

Speech recognition technology to improve retail store checkout
11557298 · 2023-01-17 · ·

Systems and methods for using speech recognition technologies to facilitate retail store checkout are disclosed. According to certain aspects, an electronic device may detect a user's speech and analyze the speech to identify a set of matching items that may correspond to items being purchased by a customer. The electronic device may display, via a user interface, the set of matching items as well as a code or identification associated with the set of matching items. The user may interface with a point of sale system to input a code for a desired item, and the point of sale system may add the desired item to an order and may facilitate a checkout for the order.

Speech recognition technology to improve retail store checkout
11557298 · 2023-01-17 · ·

Systems and methods for using speech recognition technologies to facilitate retail store checkout are disclosed. According to certain aspects, an electronic device may detect a user's speech and analyze the speech to identify a set of matching items that may correspond to items being purchased by a customer. The electronic device may display, via a user interface, the set of matching items as well as a code or identification associated with the set of matching items. The user may interface with a point of sale system to input a code for a desired item, and the point of sale system may add the desired item to an order and may facilitate a checkout for the order.

PROVIDING PROMPTS IN SPEECH RECOGNITION RESULTS IN REAL TIME
20230215441 · 2023-07-06 ·

The present disclosure provides methods and apparatuses for providing prompts in speech recognition results in real time. A current speech input in an audio stream for a target event may be obtained. A current utterance text corresponding to the current speech input may be identified. A prompt may be generated based at least on the current utterance text, the prompt comprising at least one predicted subsequent utterance text sequence. A speech recognition result for the current speech input may be provided, the speech recognition result comprising the current utterance text and the prompt.

PROVIDING PROMPTS IN SPEECH RECOGNITION RESULTS IN REAL TIME
20230215441 · 2023-07-06 ·

The present disclosure provides methods and apparatuses for providing prompts in speech recognition results in real time. A current speech input in an audio stream for a target event may be obtained. A current utterance text corresponding to the current speech input may be identified. A prompt may be generated based at least on the current utterance text, the prompt comprising at least one predicted subsequent utterance text sequence. A speech recognition result for the current speech input may be provided, the speech recognition result comprising the current utterance text and the prompt.

Intelligent Test Cases Generation Based on Voice Conversation

Aspects of the disclosure relate to generating test cases based on voice conversation. In some embodiments, a computing platform may receive voice data associated with an agile development meeting. Subsequently, the computing platform may identify, using a natural language processing engine, context of one or more requirements being discussed during the agile development meeting. Based on identifying the context of the one or more requirements being discussed during the agile development meeting, the computing platform may store context data into a database. Next, the computing platform may map the context data to a corresponding task item of a software development project. Thereafter, the computing platform may identify one or more test cases to be generated. Then, the computing platform may cause the identified test cases to be executed.

Intelligent Test Cases Generation Based on Voice Conversation

Aspects of the disclosure relate to generating test cases based on voice conversation. In some embodiments, a computing platform may receive voice data associated with an agile development meeting. Subsequently, the computing platform may identify, using a natural language processing engine, context of one or more requirements being discussed during the agile development meeting. Based on identifying the context of the one or more requirements being discussed during the agile development meeting, the computing platform may store context data into a database. Next, the computing platform may map the context data to a corresponding task item of a software development project. Thereafter, the computing platform may identify one or more test cases to be generated. Then, the computing platform may cause the identified test cases to be executed.

AUTOCORRECTION OF PRONUNCIATIONS OF KEYWORDS IN AUDIO/VIDEOCONFERENCES
20230005487 · 2023-01-05 ·

The present disclosure relates to automatically correcting mispronounced keywords during a conference session. More particularly, the present invention provides methods and systems for automatically correcting audio data generated from audio input having indications of mispronounced keywords during an audio/videoconferencing system. In some embodiments, the process of automatically correcting the audio data may require a re-encoding process of the audio data at the conference server. In alternative embodiments, the process may require updating the audio data at the receiver end of the conferencing system.