Patent classifications
G10L2021/065
SYSTEMS AND METHODS FOR VARIABLY PACED REAL-TIME TRANSLATION BETWEEN THE WRITTEN AND SPOKEN FORMS OF A WORD
An enunciation system (ES) enables users to gain acquaintance, understanding, and mastery of the relationship between letters and sounds in the context of an alphabetic writing system. The ES enables the user to experience the action of sounding out a word, before their own phonics knowledge enables them to sound out the word independently; its continuous, unbroken speech output or input avoids the common confusions that ensue from analyzing words by breaking them up into discrete sounds; its user-controlled pacing allows the user to slow down enunciation at specific points of difficulty within the word; its real-time touch control allows the written word to be played like a musical instrument, with expressive and aesthetic possibilities; and its highlighting of the letter cluster that is responsible for the recognized phoneme enunciated by the user as it occurs allows the user to more easily associated the letters with the sounds.
Natural language processing based sign language generation
In some examples, natural language processing based sign language generation may include ascertaining a speech video that is selected by a user, and determining, based on application of natural language processing to contents of the speech video, a plurality of sentences included in the speech video. For each sentence of the plurality of sentences identified in the speech video, a sign language sentence type, a sign language sentence structure, and a sentiment may be determined. For each sign language sentence structure and based on a corresponding sentiment, a sign video may be determined. Based on the sign video determined for each sentence of the plurality of sentences identified in the speech video, a combined sign video may be generated.
Methods and systems for providing images for facilitating communication
Aspects of the disclosure include a computer-implemented method for interacting with a user. Identity information for a user can be received. The identity information can be analyzed to identify the user. User information for an identified user can be retrieved, the user information indicating that a voice interacting with the identified user is to be translated into image data to help the identified user communicate with the voice. Translated image data translating voice instructions by the voice can be retrieved. The translated image data can be displayed to the identified user.
Semiautomated relay method and apparatus
A hearing user's device for communicating with a hearing impaired assisted user using an assisted user's device that includes a speaker and a display screen for broadcasting a hearing user's voice signal and presenting captioned text associated with the hearing user's voice signal to the assisted user, respectively, the hearing user's device comprising a microphone for receiving the hearing user's voice signal as spoken by the hearing user, a display screen and a processor linked to the microphone and the display screen, the processor transmitting the hearing user's voice signal to the assisted users device, the processor presenting a quality indication of the captioned text presented to the assisted user via the assisted users device for consideration by the hearing user.
Haptic communication system using cutaneous actuators for simulation of continuous human touch
A haptic communication device includes an array of cutaneous actuators to generate haptic sensations corresponding to actuator signals received by the array. The haptic sensations include at least a first haptic sensation and a second haptic sensation. The array includes at least a first cutaneous actuator to begin generating the first haptic sensation at a first location on a body of a user at a first time. A second cutaneous actuator begins generating the second haptic sensation at a second location on the body of the user at a second time later than the first time.
METHODS AND SYSTEMS FOR PROVIDING IMAGES FOR FACILITATING COMMUNICATION
Aspects of the disclosure include a computer-implemented method for interacting with a user. Identity information for a user can be received. The identity information can be analyzed to identify the user. User information for an identified user can be retrieved, the user information indicating that a voice interacting with the identified user is to be translated into image data to help the identified user communicate with the voice. Translated image data translating voice instructions by the voice can be retrieved. The translated image data can be displayed to the identified user.
Machine communication system using haptic symbol set
A haptic device comprises a signal generator that is configured to receive an input word that is a unit of a language. The signal generator converts the input word into one or more phonemes of the input word. The signal generator further converts the one or more phonemes into a sequence of actuator signals. The sequence of actuator signals is formed from a concatenation of sub-sequences of actuator signals. Each phoneme corresponding to a unique sub-sequence of actuator signals. The haptic device further comprises a two dimensional array of cutaneous actuators configured to receive the sequence of actuator signals from the signal generator, each of the actuator signals mapped to a cutaneous actuator of the two dimensional array of cutaneous actuators.
AUDIO IMPROVEMENT USING CLOSED CAPTION DATA
Methods and systems are described herein for improving audio for hearing impaired content consumers. An example method may comprise determining a content asset. Closed caption data associated with the content asset may be determined. At least a portion of the closed caption data may be determined based on a user setting associated with a hearing impairment. Compensating audio comprising a frequency translation associated with at least the portion of the closed caption data may be generated. The content asset may be caused to be output with audio content comprising the compensating audio and the original audio.
Device for enhancement of language processing in autism spectrum disorders through modifying the auditory stream including an acoustic stimulus to reduce an acoustic detail characteristic while preserving a lexicality of the acoustics stimulus
Methods and devices can enhance language processing in an autism spectrum disorder (ASD) individual through auditory manipulation of an auditory stream. The auditory stream is received and includes an acoustic stimulus perceptually representing an object. An acoustic manipulation parameter for a predetermined acoustic detail characteristic is selected. The predetermined acoustic detail characteristic is associated with the ASD individual and is based on a measured language processing capability of the ASD individual. The auditory stream is modified based on the selected parameter, to reduce the predetermined acoustic detail characteristic while preserving a lexicality of the stimulus, such that the reduced acoustic detail characteristic enhances perception of the object by the ASD individual even when the stimulus includes two or more acoustically distinct stimuli each perceptually representing the object. The modified auditory stream is output to the ASD individual via at least one loudspeaker.
GENERATING VISUAL CLOSED CAPTION FOR SIGN LANGUAGE
Embodiments describe an approach for generating a sign language translation of an audio portion of a video. Embodiments receive a request for a sign language translation for a selected video and extract audio from the selected video. Additionally, embodiments, convert the extracted audio into text, identify contextual sounds in the audio, and convert the text and the contextual sounds into sign language content. Furthermore, embodiments, generate a sign language video based on the sign language content, and display the sign language video in a separate display window on the selected video.