H04S2400/15

Audio processing apparatus and method therefor

An audio processing apparatus comprises a receiver (705) which receives audio data including audio components and render configuration data including audio transducer position data for a set of audio transducers (703). A renderer (707) generating audio transducer signals for the set of audio transducers from the audio data. The renderer (7010) is capable of rendering audio components in accordance with a plurality of rendering modes. A render controller (709) selects the rendering modes for the renderer (707) from the plurality of rendering modes based on the audio transducer position data. The renderer (707) can employ different rendering modes for different subsets of the set of audio transducers the render controller (709) can independently select rendering modes for each of the different subsets of the set of audio transducers (703). The render controller (709) can select the rendering mode for a first audio transducer of the set of audio transducers (703) in response to a position of the first audio transducer relative to a predetermined position for the audio transducer. The approach may provide improved adaptation, e.g. to scenarios where most speakers are at desired positions whereas a subset deviate from the desired position(s).

Systems and methods for processing audio signals based on user device parameters
11503421 · 2022-11-15 · ·

In various applications, the system provides a method for processing audio signals, including: receiving a request for audio content; receiving an identifier encoded in a personal audio device comprising a transducer for playing audio; retrieving at least one parameter associated with the identifier; and processing the audio content using at least the request, the identifier and the at least one parameter, wherein the processing is customized for the personal audio device based on the at least one parameter associated with the identifier. In various applications the parameter is, one or more of, associated with a specification of the personal audio device, acoustic metrics of the transducer, relates to control of equalization, relates to permission to enable proprietary sonic processing for enhanced acoustic reception of streaming content, relates to acoustic metrics of the transducer and wherein the identifier is associated with permission to enable proprietary sonic processing for enhanced acoustic reception of streaming content, and/or is stored in a chip on the personal audio device, is retrieved from a server in a network, among other things. In various applications, the personal audio device comprises ear buds and the identifier is stored in a non-volatile memory of the ear buds.

Encoded audio metadata-based equalization
11501789 · 2022-11-15 · ·

A system for producing an encoded digital audio recording has an audio encoder that encodes a digital audio recording having a number of audio channels or audio objects. An equalization (EQ) value generator produces a sequence of EQ values which define EQ filtering that is to be applied when decoding the encoded digital audio recording, wherein the EQ filtering is to be applied to a group of one or more of the audio channels or audio objects of the recording independent of any downmix. A bitstream multiplexer combines the encoded digital audio recording with the sequence of EQ values, the latter as metadata associated with the encoded digital audio recording. Other embodiments are also described including a system for decoding the encoded audio recording.

Mixed reality system for context-aware virtual object rendering

A computer-implemented method in conjunction with mixed reality gear (e.g., a headset) includes imaging a real scene encompassing a user wearing a mixed reality output apparatus. The method includes determining data describing a real context of the real scene, based on the imaging; for example, identifying or classifying objects, lighting, sound or persons in the scene. The method includes selecting a set of content including content enabling rendering of at least one virtual object from a content library, based on the data describing a real context, using various selection algorithms. The method includes rendering the virtual object in the mixed reality session by the mixed reality output apparatus, optionally based on the data describing a real context (“context parameters”). An apparatus is configured to perform the method using hardware, firmware, and/or software.

System and method for realistic rotation of stereo or binaural audio
11589181 · 2023-02-21 ·

A system for rotating sound or selective listening to sound provides for the ability of the apparent direction of sound sources in a listening environment to remain in consistent orientations in space despite rotations of the microphones used to capture the sound and despite rotations of the head of the listener, even when wearing headphones. Modules are provided in the system to distinguish the sound sources and their apparent directions, as well as to optionally rotate the sound sources in response to detected rotations of the listener's head and/or detected rotations of the microphones.

Differential spatial rendering of audio sources
11589184 · 2023-02-21 · ·

Methods and systems for intuitive spatial audio rendering with improved intelligibility are disclosed. By establishing a virtual association between an audio source and a location in the listener's virtual audio space, a spatial audio rendering system can generate spatial audio signals that create a natural and immersive audio field for a listener. The system can receive the virtual location of the source as a parameter and map the source audio signal to a source-specific multi-channel audio signal. In addition, the spatial audio rendering system can be interactive and dynamically modify the rendering of the spatial audio in response to a user's active control or tracked movement.

SYSTEM FOR AND METHOD OF GENERATING AN AUDIO IMAGE
20230050329 · 2023-02-16 ·

A system for and a method of generating an audio image for use in rendering audio. The method comprises accessing an audio stream; accessing positional information, the positional information comprising a first position, a second position and a third position; and generating an audio image. In some embodiments, generating the audio image comprises generating, based on the audio stream, a first virtual wave front to be perceived by a listener as emanating from the first position; generating, based on the audio stream, a second virtual wave front to be perceived by the listener as emanating from the second position; and generating, based on the audio stream, a third virtual wave front to be perceived by the listener as emanating from the third position.

Determining corrections to be applied to a multichannel audio signal, associated coding and decoding
20220358937 · 2022-11-10 ·

A method and device for determining a set of corrections to be made to a multichannel sound signal, in which the set of corrections is determined on the basis of an item of information representative of a spatial image of an original multichannel signal and an item of information representative of a spatial image of the original multichannel signal that has been coded and then decoded.

SIGNAL PROCESSING DEVICE, METHOD, AND PROGRAM
20220360930 · 2022-11-10 ·

The present technology relates to a signal processing device, a method, and a program that make it possible for a user to obtain a higher realistic feeling. The signal processing device includes: an audio generation unit that generates a sound source signal according to a type of a sound source on the basis of a recorded signal obtained by sound collection by a microphone attached to a moving object; a correction information generation unit that generates position correction information indicating a distance between the microphone and the sound source; and a position information generation unit that generates sound source position information indicating a position of the sound source in a target space on the basis of microphone position information indicating a position of the microphone in the target space and the position correction information. The present technology can be applied to a recording/transmission/reproduction system.

AUDIO ZOOM
20220360891 · 2022-11-10 ·

A device includes one or more processors configured to execute instructions to determine a first phase based on a first audio signal of first audio signals and to determine a second phase based on a second audio signal of second audio signals. The one or more processors are also configured to execute the instructions to apply spatial filtering to selected audio signals of the first audio signals and the second audio signals to generate an enhanced audio signal. The one or more processors are further configured to execute the instructions to generate a first output signal including combining a magnitude of the enhanced audio signal with the first phase and to generate a second output signal including combining the magnitude of the enhanced audio signal with the second phase. The first output signal and the second output signal correspond to an audio zoomed signal.