G10L15/005

Context-Based Text Suggestion
20220391584 · 2022-12-08 ·

Generating text suggestions based on context can leverage sources associated with the context to generate more accurate and informed text suggestions. For example, the context can be a user situation, such as the user is attending a meeting. Obtaining text from sources associated with the user situation can generate a corpus of text that can be leveraged for generating the context-based text suggestions.

CONTENT TRANSLATION USER INTERFACES
20220391603 · 2022-12-08 ·

The present disclosure generally relates to user interfaces for translating content and displaying translated content.

Automated audio mapping using an artificial neural network

According to one implementation, an automated audio mapping system includes a computing platform having a hardware processor and a system memory storing an audio mapping software code including an artificial neural network (ANN) trained to identify multiple different audio content types. The hardware processor is configured to execute the audio mapping software code to receive content including multiple audio tracks, and to identify, without using the ANN, a first music track and a second music track of the multiple audio tracks. The hardware processor is further configured to execute the audio mapping software code to identify, using the ANN, the audio content type of each of the multiple audio tracks except the first music track and the second music track, and to output a mapped content file including the multiple audio tracks each assigned to a respective one predetermined audio channel based on its identified audio content type.

ELECTRONIC CONTROL DEVICE FOR AN AVIONICS SYSTEM FOR IMPLEMENTING A CRITICAL AVIONICS FUNCTION, METHOD AND COMPUTER PROGRAM THEREFOR
20220380024 · 2022-12-01 ·

An electronic control device of an avionics system for implementation of a critical avionics function, comprising: a module for receiving a voice instruction signal; a speech recognition module configured to transform the voice signal into a textual transcript; a processing module configured to associate the textual transcript with at least one action to be performed; a monitoring system comprising: a control module configured to check whether the textual transcript and/or the action to be performed is consistent if and only if: a) the textual transcript and/or the action to be performed is consistent with the expected syntax, b) the textual transcript and/or the action to be performed is consistent with the expected lexical field, and c) the textual transcript and/or the action to be performed is consistent with the current context, a module for generating an associated command only if no inconsistencies are detected.

USAGE OF VOICE RECOGNITION CONFIDENCE LEVELS IN A PASSENGER INTERFACE
20220383870 · 2022-12-01 ·

A voice recognition system for an elevator system including: one or more microphones configured to capture a voice command from an individual and convert the voice command into an audio signal; a command arbitrator including one or more speech interpretation systems, the command arbitrator being configured to analyze the audio signal and determine an interpreted command for the elevator system from the audio signal using the one or more speech interpretation systems, wherein the interpreted command includes a confidence measure associated with the interpreted command, and wherein the confidence measure is an indicator depicting how confident the command arbitrator is that the interpreted command matches the voice command from the individual.

Method of interactive foreign language learning by voice talking each other using voice recognition function and TTS function
20220383772 · 2022-12-01 ·

Disclosed is an interactive foreign language learning method that enables foreign language learning through conversation between a learner and a terminal having a screen, such as a smartphone, a tablet computer, a notebook computer, a personal computer (PC), and the like, based on a video containing foreign language sound expressions, such as movies, dramas, news, and the like, through speech transmission using a speech recognition function and a TTS function of the terminal. In the interactive foreign language learning method, upon determining that speech input by a learner in a speech waiting state for speech recognition matches a previously stored voice command, the terminal performs operation corresponding to the voice command and enters the speech waiting state again, and upon determining that the speech input by the learner does not match the voice command, the terminal allows the learner to perform foreign language learning in learning modes according to learner selection, such as a learning mode in which the learner speaks after the terminal, a conversation mode in which the terminal and a user alternately speaks a sentence, and an intermediate learning mode, while changing the learning modes in response to a voice command of the learner, whereby the learner can perform interactive foreign language learning through speech transmission between the terminal and the learner while minimizing screen touch, and can have actual conversation with other learners performing foreign language learning using the same application program implementing the learning method.

SYSTEM AND METHOD FOR FEDERATED, CONTEXT-SENSITIVE, ACOUSTIC MODEL REFINEMENT
20220383859 · 2022-12-01 ·

A system and method for federated, context-sensitive, acoustic model refinement comprising a federated language model server and a plurality of edge devices. The federated language model server may comprise one or more machine learning models trained and developed centrally on the server, and distribute these one or more machine learning models to edge devices wherein they may be operated locally on the edge devices. The edge devices may gather or generate context data that can be used by a speech recognition engine, and the local language models contained therein, to develop adaptive, context-sensitive, user-specific language models. Periodically, the federated language model server may select a subset of edge devices from which to receive uploaded local model parameters, that may be aggregated to perform central model updates wherein the updated model parameters may then be sent back to edge devices in order to update the local model parameters.

DISPLAY SYSTEM, DISPLAY DEVICE, AND CONTROL METHOD FOR DISPLAY DEVICE
20220382513 · 2022-12-01 ·

A display system includes a microphone, a voice processing device, and a display device including at least one processor. The microphone collects a voice corresponding to a command and generates voice data representing the voice. The voice processing device analyzes the voice data to generate a language identifier indicating a type of a language of the voice and command data representing the command, and outputs the language identifier and the command data. The at least one processor executes displaying a user interface screen describing information using a display language, which is one language of a plurality of types of languages, receiving the language identifier and the command data outputted from the voice processing device, comparing the type indicated by the language identifier with the type of the display language, and changing the display language to the language of the type indicated by the language identifier when the type indicated by the language identifier and the type of the display language differ from each other.

Using multiple languages during speech to text input

A method and apparatus for correcting a wrongly-translated word in a device employing speech recognition is provided herein. During operation, a device will use a second language to correct a wrongly-translated word that was wrongly translated using a first language. More particularly, after speech recognition is performed using the first language, when a user selects text to be corrected, the user will utter the speech again using the second language that differs from the first language. Both the first and the second language can be used by the device to determine a best translation of the speech.

METHOD AND APPARATUS FOR GENERATING INTERACTION RECORD, AND DEVICE AND MEDIUM
20220375460 · 2022-11-24 ·

A method and apparatus for generating an interaction record, and a device and a medium are provided. The method includes: firstly, from a multimedia data stream, collecting behavior data, represented by the multimedia data stream, of a user, wherein the behavior data includes voice information and/or operation information; and then, on the basis of the behavior data, generating interaction record data corresponding to the behavior data. According to the technical solution, by means of collecting voice information and/or operation information from a multimedia data stream, and generating interaction record data on the basis of the voice information and the operation information, an interacting user can determine interaction information by using the interaction record data, and the interaction efficiency of the interacting user is improved, thereby also improving the user experience.