H04S2400/01

Splitting frequency-domain processing between multiple DSP cores

An audio processing system may split frequency-domain processing between multiple DSP cores. Processing multi-channel audio data—e.g., from devices with multiple speakers—may require more computing power than available on a single DSP core. Such processing typically occurs in the frequency domain; DSP cores, however, typically communicate via ports configured for transferring data in the time-domain. Converting frequency-domain data into the time domain for transfer requires additional resources and introduces lag. Furthermore, transferring frequency-domain data may result in scheduling issues due to a mismatch between buffer size, bit rate, and the size of the frequency-domain data chunks transferred. However, the buffer size and bit rate may be artificially configured to transfer a chunk of frequency-domain data corresponding to a delay in the communication mechanism used by the DSP cores. In this manner, frequency-domain data can be transferred with a proper periodicity.

Apparatus and Method for Reproducing a Spatially Extended Sound Source or Apparatus and Method for Generating a Description for a Spatially Extended Sound Source Using Anchoring Information
20220377489 · 2022-11-24 ·

An apparatus for reproducing a spatially extended sound source having a defined position or orientation and geometry in a space has an interface for receiving a listener position. The apparatus having a projector for calculating a projection of a two- or three-dimensional hull associated with the sound source onto a projection plane using the listener position, information on the geometry of the sound source, and on the position of the sound source; a sound position calculator for calculating positions of at least two sound sources for the spatially extended sound source using the projection plane; and a renderer for rendering the at least two sound sources at the positions to obtain a reproduction of the sound source having two or more output signals, configured to use different sound signals for the different positions.

METHODS, APPARATUS AND SYSTEMS FOR REPRESENTATION, ENCODING, AND DECODING OF DISCRETE DIRECTIVITY DATA

The present disclosure relates to a method of processing audio content including directivity information for at least one sound source, the directivity information comprising a first set of first directivity unit vectors representing directivity directions and associated first directivity gains. The disclosure further relates to corresponding methods of encoding and decoding audio content including directivity information for at least one sound source.

LOW LATENCY, LOW POWER MULTI-CHANNEL AUDIO PROCESSING

An electronic eyewear device includes a display and a speaker system adapted to present augment reality objects and associated sounds in a scene being viewed by the user. A processor receives one or more audio tracks respectively associated with one or more augmented reality objects, encodes the audio tracks into an aggregated audio track including the audio tracks, a header for each audio track that uniquely identifies each respective audio track, and an aggregate header that identifies the number of tracks in the aggregated audio track. The processor transfers the aggregated audio track to an audio processor that uses the header for each audio track and the aggregate header to separate the audio tracks from the aggregated audio track. The audio processor processes the audio tracks independently in parallel and provides the audio tracks to the speaker system for presentation with the augmented reality objects.

Systems, Methods, and Graphical User Interfaces for Selecting Audio Output Modes of Wearable Audio Output Devices
20220374197 · 2022-11-24 ·

A computer system displays an audio settings user interface that includes a first user interface element that is activatable to change a current audio output mode of a set of one or more audio output devices. The computer system receives a second set of one or more inputs including an input directed to the first user interface element. In response, the computer system transitions the set of one or more audio output devices from a first audio output mode to a different second audio output mode. In the first audio output mode, audio is output based on a first frame of reference that is a three-dimensional physical environment surrounding the set of one or more audio output devices. In the second audio output mode, audio is output based on a different second frame of reference that is fixed relative to the set of one or more audio output devices.

Apparatus and method for encoding or decoding a multi-channel signal

An apparatus for encoding a multi-channel signal having at least three channels includes an iteration processor, a channel encoder and an output interface. The iteration processor is configured to calculate inter-channel correlation values between each pair of the at least three channels, for selecting a pair including a highest value or including a value above a threshold, and for processing the selected pair using a multi-channel processing operation to derive first multi-channel parameters for the selected pair and to derive first processed channels. The iteration processor is configured to perform the calculating, the selecting and the processing using at least one of the processed channels to derive second multi-channel parameters and second processed channels. The channel encoder is configured to encode channels resulting from an iteration processing to obtain encoded channels. The output interface is configured to generate an encoded multi-channel signal including the encoded channels and the first and second multi-channel parameters.

Wearable electronic device that displays a boundary of a three-dimensional zone
11510022 · 2022-11-22 ·

A wearable electronic device (WED) includes one or more sensors and cameras that determine a location of a physical object in a zone where the user is located and that track movement of an electronic device that moves to define a boundary of the zone. The WED includes a processor that generates binaural sound and a display that displays a virtual image of the boundary of the zone and a visual warning that notifies the user of the physical object.

Interaural time difference crossfader for binaural audio rendering

Examples of the disclosure describe systems and methods for presenting an audio signal to a user of a wearable head device. According to an example method, a first input audio signal is received, the first input audio signal corresponding to a source location in a virtual environment presented to the user via the wearable head device. The first input audio signal is processed to generate a left output audio signal and a right output audio signal. The left output audio signal is presented to the left ear of the user via a left speaker associated with the wearable head device. The right output audio signal is presented to the right ear of the user via a right speaker associated with the wearable head device. Processing the first input audio signal comprises applying a delay process to the first input audio signal to generate a left audio signal and a right audio signal; adjusting a gain of the left audio signal; adjusting a gain of the right audio signal; applying a first head-related transfer function (HRTF) to the left audio signal to generate the left output audio signal; and applying a second HRTF to the right audio signal to generate the right output audio signal. Applying the delay process to the first input audio signal comprises applying an interaural time delay (ITD) to the first input audio signal, the ITD determined based on the source location.

APPARATUS AND METHOD FOR PROCESSING MULTI-CHANNEL AUDIO SIGNAL

Disclosed is an apparatus and method for processing a multichannel audio signal. A multichannel audio signal processing method may include: generating an N-channel audio signal of N channels by down-mixing an M-channel audio signal of M channels; and generating a stereo audio signal by performing binaural rendering of the N-channel audio signal.

Spatial Audio Representation and Rendering
20220369061 · 2022-11-17 ·

An apparatus including circuitry configured to: obtain a spatial audio signal including at least one audio signal and spatial metadata associated with the at least one audio signal; obtain at least one data set related to binaural rendering; obtain at least one pre-defined data set related to binaural rendering; and generate a binaural audio signal based on a combination of at least part of the at least one data set and the at least one pre-defined data set, and the spatial audio signal.