Patent classifications
H04S3/00
Electronic system for producing a coordinated output using wireless localization of multiple portable electronic devices
Device localization (e.g., ultra-wideband device localization) may be used to provide coordinated outputs and/or receive coordinated inputs using multiple devices. Providing coordinated outputs may include providing partial outputs using multiple devices, modifying an output of a device based on its position and/or orientation relative to another device, and the like. In some cases, each device of a set of multiple devices may provide a partial output, which combines with partial outputs of the remaining devices to produce a coordinated output.
QUANTIZATION OF SPATIAL AUDIO DIRECTION PARAMETERS
A method for spatial audio signal encoding comprising: obtaining a plurality of audio direction parameters, wherein each parameter comprises an elevation value and an azimuth value and wherein each parameter has an ordered position; deriving for each of the plurality of audio direction parameters a corresponding derived audio direction parameter (SP) comprising an elevation and an azimuth value, corresponding derived audio direction parameters (SP) being arranged in a manner determined by a spatial utilization defined by the elevation values and the azimuth values of the plurality of audio direction parameters; rotating each derived audio direction parameter (SP) by the azimuth value (φ.sub.0) of an audio direction parameter in the first position of the plurality of audio direction parameters and quantizing the rotation to determine for each a corresponding quantized rotated derived audio direction parameter; changing the ordered position of an audio direction parameter to a further position coinciding with a position of a rotated derived audio direction parameter when the azimuth value of the audio direction parameter is closest to the azimuth value of the further rotated derived audio direction parameter compared to the azimuth values of other rotated derived audio direction parameters, followed by determining for each of the plurality audio direction parameters a difference between each audio direction parameter and their corresponding quantized rotated derived audio direction parameter; and quantizing a difference for each of the plurality of audio direction parameters, wherein a difference quantization resolution for each of the plurality of audio direction parameters is defined based on a spatial extent of the audio direction parameters.
SPATIAL AUDIO MONAURALIZATION VIA DATA EXCHANGE
A device includes a memory configure to store instructions and one or more processors configured to execute the instructions to obtain spatial audio data at a first audio output device. The one or more processors are further configured to perform data exchange, between the first audio output device and a second audio output device, of exchange data based on the spatial audio data. The one or more processors are also configured to generate first monaural audio output at the first audio output device based on the spatial audio data.
APPARATUS AND METHOD FOR PROCESSING MULTI-CHANNEL AUDIO SIGNAL
An apparatus for processing audio includes at least one processor configured to obtain a down-mixed audio signal from a bitstream, to obtain down-mixing-related information from the bitstream, to de-mix the down-mixing-related information by using down-mixing-related information, and to reconstruct an audio signal including at least one frame based on the de-mixed audio signal. The down-mixing-related information is information generated in units of frames by using an audio scene type.
SYSTEM AND METHOD FOR AUTOMATICALLY TUNING DIGITAL SIGNAL PROCESSING CONFIGURATIONS FOR AN AUDIO SYSTEM
Embodiments include a processing device communicatively coupled to a plurality of audio devices comprising at least one microphone and at least one speaker, and to a digital signal processing (DSP) component having a plurality of audio input channels for receiving audio signals captured by the at least one microphone, the processing device being configured to identify one or more of the audio devices based on a unique identifier associated with each of said one or more audio devices; obtain device information from each identified audio device; and adjust one or more settings of the DSP component based on the device information. A computer-implemented method of automatically configuring an audio conferencing system, comprising a digital signal processing (DSP) component and a plurality of audio devices including at least one speaker and at least one microphone, is also provided.
Providing a multi-channel and a multi-zone audio environment
A multi-channel and multi-zone audio environment is provided. Various inventions are disclosed that allow playback devices on one or more networks to provide an effective multi-channel and a multi-zone audio environment using timing information. According to one example, timing information is used to coordinate playback devices connected over a low-latency network to provide audio along with a video display. In another example, timing information is used to coordinate playback devices connected over a mesh network to provide audio in one or more zones or zone groups.
Generating sound zones using variable span filters
The invention provides a method for generating output filters to a plurality of loudspeakers at respective positions for playback of a plurality of different input signals in respective spatially different sound zones by means of a processor system. The method comprising computing spatio-temporal correlation matrices in response to spatial information, e.g. measured transfer functions, and in response to desired sound pressures in the plurality of sound zones. Joint eigenvalue decomposition of the spatial correlation matrices are then computed, or at least an approximation thereof, to arrive at eigenvectors accordingly. Next, variable span filters a reformed from a linear combination of the eigenvectors in response to a desired trade-off between acoustic contrast and acoustic errors in the sound zones. Finally, output filter for each of the plurality of loudspeakers, for each of the plurality of input signals, in accordance with the variable span filters. The method is applicable also for optimization in one zone, e.g. for room equalization.
Splitting frequency-domain processing between multiple DSP cores
An audio processing system may split frequency-domain processing between multiple DSP cores. Processing multi-channel audio data—e.g., from devices with multiple speakers—may require more computing power than available on a single DSP core. Such processing typically occurs in the frequency domain; DSP cores, however, typically communicate via ports configured for transferring data in the time-domain. Converting frequency-domain data into the time domain for transfer requires additional resources and introduces lag. Furthermore, transferring frequency-domain data may result in scheduling issues due to a mismatch between buffer size, bit rate, and the size of the frequency-domain data chunks transferred. However, the buffer size and bit rate may be artificially configured to transfer a chunk of frequency-domain data corresponding to a delay in the communication mechanism used by the DSP cores. In this manner, frequency-domain data can be transferred with a proper periodicity.
METHODS, APPARATUS AND SYSTEMS FOR REPRESENTATION, ENCODING, AND DECODING OF DISCRETE DIRECTIVITY DATA
The present disclosure relates to a method of processing audio content including directivity information for at least one sound source, the directivity information comprising a first set of first directivity unit vectors representing directivity directions and associated first directivity gains. The disclosure further relates to corresponding methods of encoding and decoding audio content including directivity information for at least one sound source.
LOW LATENCY, LOW POWER MULTI-CHANNEL AUDIO PROCESSING
An electronic eyewear device includes a display and a speaker system adapted to present augment reality objects and associated sounds in a scene being viewed by the user. A processor receives one or more audio tracks respectively associated with one or more augmented reality objects, encodes the audio tracks into an aggregated audio track including the audio tracks, a header for each audio track that uniquely identifies each respective audio track, and an aggregate header that identifies the number of tracks in the aggregated audio track. The processor transfers the aggregated audio track to an audio processor that uses the header for each audio track and the aggregate header to separate the audio tracks from the aggregated audio track. The audio processor processes the audio tracks independently in parallel and provides the audio tracks to the speaker system for presentation with the augmented reality objects.