Patent classifications
H04S2400/11
LOW LATENCY, LOW POWER MULTI-CHANNEL AUDIO PROCESSING
An electronic eyewear device includes a display and a speaker system adapted to present augment reality objects and associated sounds in a scene being viewed by the user. A processor receives one or more audio tracks respectively associated with one or more augmented reality objects, encodes the audio tracks into an aggregated audio track including the audio tracks, a header for each audio track that uniquely identifies each respective audio track, and an aggregate header that identifies the number of tracks in the aggregated audio track. The processor transfers the aggregated audio track to an audio processor that uses the header for each audio track and the aggregate header to separate the audio tracks from the aggregated audio track. The audio processor processes the audio tracks independently in parallel and provides the audio tracks to the speaker system for presentation with the augmented reality objects.
SOUND EFFECT SIMULATION BY CREATING VIRTUAL REALITY OBSTACLE
According to one embodiment, a method, computer system, and computer program product for modulating external sounds to reflect the acoustic effects of virtual objects in a mixed-reality environment is provided. The present invention may include creating a knowledge corpus, recording a sound effect occurring externally to a mixed-reality environment experienced by a user operating the mixed-reality device; identifying one or more objects within the mixed-reality environment, including at least one virtual object; modulating the sound effect based on the knowledge corpus to simulate one or more acoustic effects of the one or more objects within the MR environment; and playing the modulated sound effect to the user.
Method and System for Detecting Sound Event Liveness Using a Microphone Array
A method performed by an electronic device in a room. The method performs an enrollment process in which a spatial profile of a location of an artificial sound source is created and performs an identification process that determines whether a sound event within the room is produced by the artificial sound source by 1) capturing the sound event using a microphone array and 2) determining a likelihood that the sound event occurred at the location of the artificial sound source.
Systems, Methods, and Graphical User Interfaces for Selecting Audio Output Modes of Wearable Audio Output Devices
A computer system displays an audio settings user interface that includes a first user interface element that is activatable to change a current audio output mode of a set of one or more audio output devices. The computer system receives a second set of one or more inputs including an input directed to the first user interface element. In response, the computer system transitions the set of one or more audio output devices from a first audio output mode to a different second audio output mode. In the first audio output mode, audio is output based on a first frame of reference that is a three-dimensional physical environment surrounding the set of one or more audio output devices. In the second audio output mode, audio is output based on a different second frame of reference that is fixed relative to the set of one or more audio output devices.
Audio processing in adaptive intermediate spatial format
Systems, methods, and computer program products of audio processing based on Adaptive Intermediate Spatial Format (AISF) are described. The AISF is an extension to ISF that allows spatial resolution around an ISF ring to be adjusted dynamically with respect to content of incoming audio objects. An AISF encoder device adaptively warps each ISF ring during ISF encoding to adjust angular distance between objects, resulting in increase in uniformity of energy distribution around the ISF ring. At an AISF decoder device, matrices that decode sound positions to the output speaker take into account the warping that was performed at the AISF encoder device to reproduce the true positions of sound sources.
Spatializing audio data based on analysis of incoming audio data
A system for enabling spatializing audio data is provided. The system analyzes audio data to identify when to generate spatialized audio data. The system can receive incoming audio data including a plurality of channel-based audio signals as well as object-based audio. The system performs an analysis of the audio data and/or metadata associated with the audio data to determine when to generate the spatialized audio data. The system can identify one or more categories associated with the audio data (e.g., stereo, mono, game effect, . . . ) and use the category to determine whether to spatialize the audio data or not spatialize the audio data.
Wearable electronic device that displays a boundary of a three-dimensional zone
A wearable electronic device (WED) includes one or more sensors and cameras that determine a location of a physical object in a zone where the user is located and that track movement of an electronic device that moves to define a boundary of the zone. The WED includes a processor that generates binaural sound and a display that displays a virtual image of the boundary of the zone and a visual warning that notifies the user of the physical object.
Interaural time difference crossfader for binaural audio rendering
Examples of the disclosure describe systems and methods for presenting an audio signal to a user of a wearable head device. According to an example method, a first input audio signal is received, the first input audio signal corresponding to a source location in a virtual environment presented to the user via the wearable head device. The first input audio signal is processed to generate a left output audio signal and a right output audio signal. The left output audio signal is presented to the left ear of the user via a left speaker associated with the wearable head device. The right output audio signal is presented to the right ear of the user via a right speaker associated with the wearable head device. Processing the first input audio signal comprises applying a delay process to the first input audio signal to generate a left audio signal and a right audio signal; adjusting a gain of the left audio signal; adjusting a gain of the right audio signal; applying a first head-related transfer function (HRTF) to the left audio signal to generate the left output audio signal; and applying a second HRTF to the right audio signal to generate the right output audio signal. Applying the delay process to the first input audio signal comprises applying an interaural time delay (ITD) to the first input audio signal, the ITD determined based on the source location.
Acoustic output device and buttons thereof
The present disclosure relates to an acoustic output device including an earphone core, a controller, a Bluetooth module, and a button module. The earphone core may include at least one low-frequency acoustic driver configured to output sounds from at least two first guiding holes and at least one high-frequency acoustic driver configured to output sounds from at least two second guiding holes. The controller may be configured to direct the at least one low-frequency acoustic driver to output the sounds in a first frequency range and direct the at least one high-frequency acoustic driver to output the sounds in a second frequency range. The Bluetooth module may be configured to connect the acoustic output device with at least one terminal device. The button module may be configured to implement an interaction between a user of the acoustic output device and the acoustic output device.
Bidirectional propagation of sound
The description relates to rendering directional sound. One implementation includes receiving directional impulse responses corresponding to a scene. The directional impulse responses can correspond to multiple sound source locations and a listener location in the scene. The implementation can also include encoding the directional impulse responses to obtain encoded departure direction parameters for individual sound source locations. The implementation can also include outputting the encoded departure direction parameters, the encoded departure direction parameters providing sound departure directions from the individual sound source locations for rendering of sound.