Patent classifications
G10L21/16
Augmenting speech segmentation and recognition using head-mounted vibration and/or motion sensors
Example methods and systems use multiple sensors to determine whether a speaker is speaking. Audio data in an audio-channel speech band detected by a microphone can be received. Vibration data in a vibration-channel speech band representative of vibrations detected by a sensor other than the microphone can be received. The microphone and the sensor can be associated with a head-mountable device (HMD). It is determined whether the audio data is causally related to the vibration data. If the audio data and the vibration data are causally related, an indication can be generated that the audio data contains HMD-wearer speech. Causally related audio and vibration data can be used to increase accuracy of text transcription of the HMD-wearer speech. If the audio data and the vibration data are not causally related, an indication can be generated that the audio data does not contain HMD-wearer speech.
Augmenting speech segmentation and recognition using head-mounted vibration and/or motion sensors
Example methods and systems use multiple sensors to determine whether a speaker is speaking. Audio data in an audio-channel speech band detected by a microphone can be received. Vibration data in a vibration-channel speech band representative of vibrations detected by a sensor other than the microphone can be received. The microphone and the sensor can be associated with a head-mountable device (HMD). It is determined whether the audio data is causally related to the vibration data. If the audio data and the vibration data are causally related, an indication can be generated that the audio data contains HMD-wearer speech. Causally related audio and vibration data can be used to increase accuracy of text transcription of the HMD-wearer speech. If the audio data and the vibration data are not causally related, an indication can be generated that the audio data does not contain HMD-wearer speech.
Interspecies language with enabling technology and training protocols
The Interspecies Language (ISL) is a method for communicating that includes a visual language, as tonal language, a training protocol that integrates them, and an enabling apparatus. The ISL visual language communicates through a display that creates a “picture” of a sentence on a touchscreen. This visual display embodies the ISL's syntax and structures how the user employs it to build sentences. Sentences are constructed by dragging icons, which are pictures or abstract images representing objects or concepts, into the structured areas of the visual array. The ISL tonal language communicates through a structured series of tones that are grouped so that tones within a grouping are members of a class of objects or concepts. The tones are sounded when an icon is placed into the visual array in the course of building a sentence. The ISL is deployed through an apparatus that includes a touchscreen, computer, chute, and industrial controls. With its clear structure for the animal's response, the ISL provides a method for meaningful, two-way communication between animals and humans or between humans.
Interspecies language with enabling technology and training protocols
The Interspecies Language (ISL) is a method for communicating that includes a visual language, as tonal language, a training protocol that integrates them, and an enabling apparatus. The ISL visual language communicates through a display that creates a “picture” of a sentence on a touchscreen. This visual display embodies the ISL's syntax and structures how the user employs it to build sentences. Sentences are constructed by dragging icons, which are pictures or abstract images representing objects or concepts, into the structured areas of the visual array. The ISL tonal language communicates through a structured series of tones that are grouped so that tones within a grouping are members of a class of objects or concepts. The tones are sounded when an icon is placed into the visual array in the course of building a sentence. The ISL is deployed through an apparatus that includes a touchscreen, computer, chute, and industrial controls. With its clear structure for the animal's response, the ISL provides a method for meaningful, two-way communication between animals and humans or between humans.
INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD, AND NON-TRANSIENT COMPUTER- READABLE STORAGE MEDIUM STORING PROGRAM
An information processing device includes a processor, the processor being configured to: acquire sound data collected in a predetermined facility; extract voice data generated by speech of a person in the predetermined facility from the sound data; and evaluate a status of customers in the predetermined facility based on the voice data.
INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD, AND NON-TRANSIENT COMPUTER- READABLE STORAGE MEDIUM STORING PROGRAM
An information processing device includes a processor, the processor being configured to: acquire sound data collected in a predetermined facility; extract voice data generated by speech of a person in the predetermined facility from the sound data; and evaluate a status of customers in the predetermined facility based on the voice data.
VOICE TRIGGER FOR A DIGITAL ASSISTANT
A method for operating a voice trigger is provided. In some implementations, the method is performed at an electronic device including one or more processors and memory storing instructions for execution by the one or more processors. The method includes receiving a sound input. The sound input may correspond to a spoken word or phrase, or a portion thereof. The method includes determining whether at least a portion of the sound input corresponds to a predetermined type of sound, such as a human voice. The method includes, upon a determination that at least a portion of the sound input corresponds to the predetermined type, determining whether the sound input includes predetermined content, such as a predetermined trigger word or phrase. The method also includes, upon a determination that the sound input includes the predetermined content, initiating a speech-based service, such as a voice-based digital assistant.
VOICE TRIGGER FOR A DIGITAL ASSISTANT
A method for operating a voice trigger is provided. In some implementations, the method is performed at an electronic device including one or more processors and memory storing instructions for execution by the one or more processors. The method includes receiving a sound input. The sound input may correspond to a spoken word or phrase, or a portion thereof. The method includes determining whether at least a portion of the sound input corresponds to a predetermined type of sound, such as a human voice. The method includes, upon a determination that at least a portion of the sound input corresponds to the predetermined type, determining whether the sound input includes predetermined content, such as a predetermined trigger word or phrase. The method also includes, upon a determination that the sound input includes the predetermined content, initiating a speech-based service, such as a voice-based digital assistant.
MUSIC LEARNING APPARATUS AND MUSIC LEARNING METHOD USING TACTILE SENSATION
A tactile music learning apparatus converts sound data of a user's voice corresponding to original music into first tactile data including tactile information, generates a synchronized tactile pattern by synchronizing the first tactile data with second tactile data including tactile information corresponding to sound data of the original music, and transfers the synchronized tactile pattern to a tactile reproducing apparatus to allow the tactile reproducing apparatus to reproduce the synchronized tactile pattern.
MUSIC LEARNING APPARATUS AND MUSIC LEARNING METHOD USING TACTILE SENSATION
A tactile music learning apparatus converts sound data of a user's voice corresponding to original music into first tactile data including tactile information, generates a synchronized tactile pattern by synchronizing the first tactile data with second tactile data including tactile information corresponding to sound data of the original music, and transfers the synchronized tactile pattern to a tactile reproducing apparatus to allow the tactile reproducing apparatus to reproduce the synchronized tactile pattern.