Patent classifications
H04M1/2475
SEMIAUTOMATED RELAY METHOD AND APPARATUS
A method and system to transcribe communications the method comprising the steps of obtaining an audio message originating at a first device during a voice communication session between the first device and a second device, providing the audio message to a first speech recognition system to generate a first transcript of the audio message, directing the first transcript to the second device, in response to obtaining an indication that indicates a quality of the first transcript is below a quality threshold, using a second speech recognition system to generate a second transcript based on the audio message while continuing to provide the audio data to the first speech recognition system to generate the first transcript and, in response to occurrence of an event that indicates the second transcript is to be directed to the second device, directing the second transcript to the second device instead of directing the first transcript.
Automated real time interpreter service
Aspects of the present disclosure relate to systems and methods that aid users with hearing and/or speech impediments to have a conversation with a remote phone number without human assistance. In one aspect, an application residing on a user's device, such as a smartphone, tablet computer, laptop, etc., may be used to initiate a phone call to a recipient. Upon initiating the phone call locally, a service residing on the server may receive a request to initiate a connection to the recipient. Once the recipient answers, the user may converse with the recipient by providing text input to their local app. The text input may be transmitted to the service. The service may use a text to speech converter to translate the received text to speech that can be delivered to the recipient.
SEMIAUTOMATED RELAY METHOD AND APPARATUS
A call captioning system for captioning a hearing user's (HU's) voice signal during an ongoing call with an assisted user (AU) includes: an AU communication device with a display screen and a caption service activation feature, and a first processor programmed to, during an ongoing call, receive the HU's voice signal. Prior to activating the caption service via the activation feature, the processor uses an automated speech recognition (ASR) engine to generate HU voice signal captions, detect errors in the HU voice signal captions, use the errors to train the ASR software to the HU's voice signal to increase accuracy of the HU captions generated by the ASR engine; and store the trained ASR engine for subsequent use. Upon activating the caption service during the ongoing call, the processor uses the trained ASR engine to generate HU voice signal captions and present them to the AU via the display screen.
Method, device, and system for audio data processing
A method and apparatus that filters audio data received from a speaking person that includes a specific filter for that speaker. The audio characteristics of the speaker's voice may be collected and the specific filter may be formed to reduce noise while also enhancing voice quality. For instance, if a speaker's voice does not contain specific frequencies, then a filter may cancel the noise at such frequencies to ease noise cancellation and reduce processing sound spectrum for cleaning that is not needed. Additionally, the strength frequencies of a speaker's voice may be identified from the collected audio characteristics and those spectrums can be filtered with finer granularity to provide a speaker specific filter that enhances the voice quality of the speaker's voice data that is transmitted or output by a communication device. The audio data may also be output based upon a user's predefined hearing spectrum.
Transcription presentation
According to one or more aspects of the present disclosure, operations related to providing transcriptions may include obtaining a first transcription of first audio obtained by a first device during a communication session conducted between the first device and a second device. The operations may further include providing the first transcription for presentation of the first transcription by a display device during the communication session. In addition, the operations may include providing, in response to a transcription quality indication, a second transcription of second audio obtained by the second device during the communication session for presentation of the second transcription by the display device during the communication session.
Transcription presentation of communication sessions
A system is provided that includes a first network interface for a first network type and a second network interface for a second network type that is different from the first network type. The system also includes at least one processor configured to cause the system to perform operations. The operations may include obtaining, from the first network interface, audio from a communication session with a remote device established over the first network and obtaining an indication of a communication device available to participate in the communication session and direct audio obtained from the communication session to a remote transcription system. The operations may also include directing the audio to the second network interface for transmission to the communication device, obtaining transcript data from the remote transcription system based on the audio, and directing the transcript data to the second network interface for transmission to the communication device.
911 call assistance for assisted device user
A system and method for use with an assisted user's communication device that includes a display. The method includes providing a user communication device including a processor that is programmed to perform the step of, upon disconnection of an emergency call to a hearing user in which a captioning service has provided text transcription of voice messages from the hearing user, maintaining a link to the captioning service for a predetermined period of time. The processor also is programmed to perform the step of recognizing the call as an emergency call upon placement of the call, and the captioning service is automatically initiated upon the processor performing that recognition.
Semiautomated relay method and apparatus
A method and system for assessing captioning quality of a call assistant (CA) that transcribes a hearing user's (HU's) voice signal during a call with an assisted user (AU) into caption text, the method comprising the steps of receiving an HU voice signal at a relay, providing the HU voice signal to a first CA to transcribe into first caption text, providing the first caption text to a second CA for correction, the text corrected by the second CA being a true text caption and comparing the first caption text to the true text caption to identify a quality factor associated with the first caption text.
Semiautomated relay method and apparatus
A captioning system comprising a processor and a memory having stored thereon software such that, when the software is executed by the one or more processors, the system generates text captions from speech data, including at least the following, receiving, from a hearing user's (HU's) device, an HU's speech data, generating, at the one or more hardware processors, first text captions from the speech data using a speech recognition algorithm, automatically determining, at the one or more processors, whether the generated first text captions meet a first accuracy threshold and when the first text captions meet the first accuracy threshold, sending the first text captions to an assisted user's (AU's) device for display, when the first text captions do not meet the first accuracy threshold, generating, at the one or more processors, second text captions from the speech data based on user input to the speech recognition algorithm from a call assistant and sending the second text captions to the AU's device for display.
SEMIAUTOMATED RELAY METHOD AND APPARATUS
A communication method to facilitate communication between first and second users that use first and second devices during a communication session, respectively, includes the steps of: obtaining a first voice signal from the first device during the communication session, using a first transcription process to generate a first caption text associated with the first voice signal, and automatically assessing an accuracy value indicating accuracy of the first caption text. When the accuracy value is below a threshold level, the method also includes obtaining a second voice signal during the communication session, using a second transcription process to generate a second caption text associated with the second voice signal, and presenting the second caption text via a display on a device that is used by the second user.