H04S7/00

WORLD LOCK SPATIAL AUDIO PROCESSING
20230046341 · 2023-02-16 ·

A method for providing a world-locked experience to a user of a headset in an immersive reality application includes receiving, from an immersive reality application, a first audio waveform from a first acoustic source to provide to a user of a headset, determining a direction of arrival for the first acoustic source relative to the headset, and providing, to a speaker in the headset, an audio signal including the first audio waveform and intended for an ear of the user of the headset, wherein the audio signal includes a time delay and an amplitude for the first audio waveform based on the direction of arrival for the first acoustic source relative to the user of the headset. A non-transitory, computer-readable medium storing instructions which, when executed by a processor, cause a system to perform the above method, and the system, are also provided.

AUDIO OUTPUT CONFIGURATION FOR MOVING DEVICES
20230048755 · 2023-02-16 ·

Described herein is a system for recalibrating an audio configuration for mobile or moving devices. The system may configure a multi-device output group to generate synchronous output audio using multiple devices. For example, the output group may include a first device generating a first portion of output audio corresponding to a first channel and a second device generating a second portion of the output audio corresponding to a second channel. If the second device detects motion and/or movement indicating a change in its location, the system may recalibrate the output group to continue generating the output audio without the second device. For example, the first device or a new device can generate the second portion of the output audio instead of the second device. When the second device returns, the system can recalibrate the output group to include the second device again.

DEVICES AND METHODS FOR AUDITORY REHABILITATION FOR INTERAURAL ASYMMETRY
20230049597 · 2023-02-16 ·

A device, system and related methods to provide assessment and treatment of amblyaudia through standardized methods that do not require advanced training or a booth with loudspeakers for the operator to administer. The ARIA stimuli protocols for both assessment and treatment, encoded in or to be used by a software program or application, are transferred to a stand-alone set of specialized noise-cancelling headphones attached or connected to, wired or wirelessly, a software platform on an electronic computing device. or integrated with the headphones. The program administers assessment tests to individuals through the noise-cancelling earphones. The device enables someone with minimal instructions to administer automatically or semi-automatically both assessment and treatment protocols, generate results, make interpretations, store data, and produce reports. The device or system may be loaded with standard protocols for English-speaking individuals, as well as dichotic speech material in any language.

Stereo Sound Pickup Method and Apparatus, Terminal Device, and Computer-Readable Storage Medium
20230048860 · 2023-02-16 ·

A stereo sound pickup method and apparatus, a terminal device, and a computer-readable storage medium. The method includes configuring a terminal device to record a video, wherein the terminal device comprises a plurality of microphones, configuring the plurality of microphones to capture a sound, and forming a stereo beam based on the captured sound. The stereo beam is related to a video recording scenario of the terminal device, and the video recording scenario includes a posture of the terminal device and usage of a camera, the posture includes that the terminal device is in a landscape mode or a portrait mode, and the usage of the camera includes that a rear-facing camera is used or a front-facing camera is used.

CAMERA-VIEW ACOUSTIC FENCE
20230053202 · 2023-02-16 ·

Determining the angle of sound relative to the centerline of a microphone array. The angle of the centerline of a camera field-of-view (FoV) and the angle of the camera FoV is determined. Knowing the angle from the centerline of the microphone array of the particular sound and then the angle of the centerline of the camera FoV and angles of the camera FoV allows a determination if the sound is inside the FoV of the camera. If so, the microphones are unmuted. If not, the microphones are muted. As the camera zooms or pans, the changes in camera FoV and centerline angle are computed and used with the sound angle, so that the muting and unmuting occurs automatically as the camera zoom and pan angle change.

XR RENDERING FOR 3D AUDIO CONTENT AND AUDIO CODEC
20230051841 · 2023-02-16 ·

A device includes a memory configured to store instructions and also includes one or more processors configured to execute the instructions to obtain audio data corresponding to a sound source and metadata indicative of a direction of the sound source. The one or more processors are configured to execute the instructions to obtain direction data indicating a viewing direction associated with a user of a playback device. The one or more processors are configured to execute the instructions to determine a resolution setting based on a similarity between the viewing direction and the direction of the sound source. The one or more processors are also configured to execute the instructions to process the audio data based on the resolution setting to generate processed audio data.

Audio Processing Methods and Systems for a Multizone Augmented Reality Space

An illustrative audio processing system identifies an experience location with which an augmented reality presentation device is associated. The experience location is included within a multizone augmented reality space that is presented by the augmented reality presentation device. The audio processing system determines that the experience location is within both a first sound zone and a second sound zone of the multizone augmented reality space, and, based on the determining that the experience location is within both the first and second sound zones, generates a binaural audio stream for presentation by the augmented reality presentation device. The binaural audio stream includes an environmental audio component implemented by a mix of a first environmental audio stream associated with the first sound zone and a second environmental audio stream associated with the second sound zone. Corresponding methods and systems are also disclosed.

SIMULTANEOUS DECONVOLUTION OF LOUDSPEAKER-ROOM IMPULSE RESPONSES WITH LINEARLY-OPTIMAL TECHNIQUES
20230052010 · 2023-02-16 ·

One embodiment provides a method comprising determining stimuli for simultaneously exciting a plurality of speakers within a spatial area. The method further comprises simultaneously exciting the plurality of speakers by providing the stimuli to the plurality of speakers at the same time for reproduction. The method further comprises recording, during the reproduction, one or more measurements of sound arriving at one or more microphones within the spatial area. The method further comprises simultaneously deconvolving a plurality of impulse responses of the plurality of speakers based on the stimuli and the one or more measurements.

MULTI-TRACK AUDIO IN A SECURITY SYSTEM
20230049433 · 2023-02-16 ·

A method, system, server and device are disclosed. According to one or more embodiments, a server is provided. A first audio track is received which includes first audio originating from a premises client at a premises location. A second audio track is received which includes second audio originating from a remote client. A first pan angle is determined for the first audio track and a second pan angle is determined for the second audio track. The second pan angle is different from the first pan angle. A stereo composite track is generated based on the first pan angle and the second pan angle, where the stereo composite track includes the first audio track and the second audio track.

AUDIO BEAM STEERING, TRACKING AND AUDIO EFFECTS FOR AR/VR APPLICATIONS

A method for audio beam steering, tracking, and audio effects for an immersive reality application is provided. The method includes receiving, from an immersive reality application, a first audio waveform from a first acoustic source to provide to a user of a headset, identifying a perceived direction for the first acoustic source relative to the headset based on a location of the first acoustic source, and providing, to a first speaker in a client device, an audio signal including the first audio waveform, wherein the audio signal includes a time delay and an amplitude of the first audio waveform based on the perceived direction. A non-transitory, computer-readable medium storing instructions which, when executed by a processor, cause a system to perform the above method, and the system, are also provided.