H04S3/008

Generating sound zones using variable span filters

The invention provides a method for generating output filters to a plurality of loudspeakers at respective positions for playback of a plurality of different input signals in respective spatially different sound zones by means of a processor system. The method comprising computing spatio-temporal correlation matrices in response to spatial information, e.g. measured transfer functions, and in response to desired sound pressures in the plurality of sound zones. Joint eigenvalue decomposition of the spatial correlation matrices are then computed, or at least an approximation thereof, to arrive at eigenvectors accordingly. Next, variable span filters a reformed from a linear combination of the eigenvectors in response to a desired trade-off between acoustic contrast and acoustic errors in the sound zones. Finally, output filter for each of the plurality of loudspeakers, for each of the plurality of input signals, in accordance with the variable span filters. The method is applicable also for optimization in one zone, e.g. for room equalization.

Splitting frequency-domain processing between multiple DSP cores

An audio processing system may split frequency-domain processing between multiple DSP cores. Processing multi-channel audio data—e.g., from devices with multiple speakers—may require more computing power than available on a single DSP core. Such processing typically occurs in the frequency domain; DSP cores, however, typically communicate via ports configured for transferring data in the time-domain. Converting frequency-domain data into the time domain for transfer requires additional resources and introduces lag. Furthermore, transferring frequency-domain data may result in scheduling issues due to a mismatch between buffer size, bit rate, and the size of the frequency-domain data chunks transferred. However, the buffer size and bit rate may be artificially configured to transfer a chunk of frequency-domain data corresponding to a delay in the communication mechanism used by the DSP cores. In this manner, frequency-domain data can be transferred with a proper periodicity.

METHODS, APPARATUS AND SYSTEMS FOR REPRESENTATION, ENCODING, AND DECODING OF DISCRETE DIRECTIVITY DATA

The present disclosure relates to a method of processing audio content including directivity information for at least one sound source, the directivity information comprising a first set of first directivity unit vectors representing directivity directions and associated first directivity gains. The disclosure further relates to corresponding methods of encoding and decoding audio content including directivity information for at least one sound source.

LOW LATENCY, LOW POWER MULTI-CHANNEL AUDIO PROCESSING

An electronic eyewear device includes a display and a speaker system adapted to present augment reality objects and associated sounds in a scene being viewed by the user. A processor receives one or more audio tracks respectively associated with one or more augmented reality objects, encodes the audio tracks into an aggregated audio track including the audio tracks, a header for each audio track that uniquely identifies each respective audio track, and an aggregate header that identifies the number of tracks in the aggregated audio track. The processor transfers the aggregated audio track to an audio processor that uses the header for each audio track and the aggregate header to separate the audio tracks from the aggregated audio track. The audio processor processes the audio tracks independently in parallel and provides the audio tracks to the speaker system for presentation with the augmented reality objects.

DIRECTIONAL SOUND TRANSMISSION METHOD, ELECTRONIC DEVICE AND READABLE STORAGE MEDIUM
20220377487 · 2022-11-24 ·

A directional sound transmission method executable by an electronic device is disclosed. A camera of the electronic device is activated. A divided area within a detectable range of the camera is recognized. A first detection operation is performed on first character information of a first person in the detectable range. A first three-dimensional (3D) coordinate information of a face of the first person is recognized through the camera, and the first 3D coordinate information is obtained. A first ultrasonic transducer transmitter corresponding to the first person is activated and the sound of the electronic device is sent to the first person.

Apparatus and method for encoding or decoding a multi-channel signal

An apparatus for encoding a multi-channel signal having at least three channels includes an iteration processor, a channel encoder and an output interface. The iteration processor is configured to calculate inter-channel correlation values between each pair of the at least three channels, for selecting a pair including a highest value or including a value above a threshold, and for processing the selected pair using a multi-channel processing operation to derive first multi-channel parameters for the selected pair and to derive first processed channels. The iteration processor is configured to perform the calculating, the selecting and the processing using at least one of the processed channels to derive second multi-channel parameters and second processed channels. The channel encoder is configured to encode channels resulting from an iteration processing to obtain encoded channels. The output interface is configured to generate an encoded multi-channel signal including the encoded channels and the first and second multi-channel parameters.

Spatializing audio data based on analysis of incoming audio data

A system for enabling spatializing audio data is provided. The system analyzes audio data to identify when to generate spatialized audio data. The system can receive incoming audio data including a plurality of channel-based audio signals as well as object-based audio. The system performs an analysis of the audio data and/or metadata associated with the audio data to determine when to generate the spatialized audio data. The system can identify one or more categories associated with the audio data (e.g., stereo, mono, game effect, . . . ) and use the category to determine whether to spatialize the audio data or not spatialize the audio data.

Wearable electronic device that displays a boundary of a three-dimensional zone
11510022 · 2022-11-22 ·

A wearable electronic device (WED) includes one or more sensors and cameras that determine a location of a physical object in a zone where the user is located and that track movement of an electronic device that moves to define a boundary of the zone. The WED includes a processor that generates binaural sound and a display that displays a virtual image of the boundary of the zone and a visual warning that notifies the user of the physical object.

Interaural time difference crossfader for binaural audio rendering

Examples of the disclosure describe systems and methods for presenting an audio signal to a user of a wearable head device. According to an example method, a first input audio signal is received, the first input audio signal corresponding to a source location in a virtual environment presented to the user via the wearable head device. The first input audio signal is processed to generate a left output audio signal and a right output audio signal. The left output audio signal is presented to the left ear of the user via a left speaker associated with the wearable head device. The right output audio signal is presented to the right ear of the user via a right speaker associated with the wearable head device. Processing the first input audio signal comprises applying a delay process to the first input audio signal to generate a left audio signal and a right audio signal; adjusting a gain of the left audio signal; adjusting a gain of the right audio signal; applying a first head-related transfer function (HRTF) to the left audio signal to generate the left output audio signal; and applying a second HRTF to the right audio signal to generate the right output audio signal. Applying the delay process to the first input audio signal comprises applying an interaural time delay (ITD) to the first input audio signal, the ITD determined based on the source location.

APPARATUS AND METHOD FOR PROCESSING MULTI-CHANNEL AUDIO SIGNAL

Disclosed is an apparatus and method for processing a multichannel audio signal. A multichannel audio signal processing method may include: generating an N-channel audio signal of N channels by down-mixing an M-channel audio signal of M channels; and generating a stereo audio signal by performing binaural rendering of the N-channel audio signal.