H04S3/008

Associated spatial audio playback

An apparatus including at least one processor and at least one memory including computer code for one or more programs, the at least one memory and the computer code configured, with the at least one processor, to cause the apparatus at least to: generate content lock information for a content lock, wherein the content lock information enables control of audio signal processing associated with audio signals related to one or more audio sources based on a position and/or orientation input.

Inter-channel phase difference parameter encoding method and apparatus

This application discloses an IPD parameter encoding method, including: obtaining a reference parameter used to determine an IPD parameter encoding scheme of a current frame of a multi-channel signal; determining the IPD parameter encoding scheme of the current frame based on the reference parameter, where the determined IPD parameter encoding scheme of the current frame is one of at least two preset IPD parameter encoding schemes; and processing an IPD parameter of the current frame based on the determined IPD parameter encoding scheme of the current frame. The technical solutions provided in this application can improve encoding quality of the multi-channel signal.

Spatial audio for interactive audio environments

Systems and methods of presenting an output audio signal to a listener located at a first location in a virtual environment are disclosed. According to embodiments of a method, an input audio signal is received. For each sound source of a plurality of sound sources in the virtual environment, a respective first intermediate audio signal corresponding to the input audio signal is determined, based on a location of the respective sound source in the virtual environment, and the respective first intermediate audio signal is associated with a first bus. For each of the sound sources of the plurality of sound sources in the virtual environment, a respective second intermediate audio signal is determined. The respective second intermediate audio signal corresponds to a reverberation of the input audio signal in the virtual environment. The respective second intermediate audio signal is determined based on a location of the respective sound source, and further based on an acoustic property of the virtual environment. The respective second intermediate audio signal is associated with a second bus. The output audio signal is presented to the listener via the first bus and the second bus.

SYSTEMS, METHODS AND APPARATUS FOR CONVERSION FROM CHANNEL-BASED AUDIO TO OBJECT-BASED AUDIO

Embodiments are disclosed for channel-based audio (CBA) (e.g., 22.2-ch audio) to object-based audio (OBA) conversion. The conversion includes converting CBA metadata to object audio metadata (OAMD) and reordering the CBA channels based on channel shuffle information derived in accordance with channel ordering constraints of the OAMD. The OBA with reordered channels is rendered in a playback device using the OAMD or in a source device, such as a set-top box or audio/video recorder. In an embodiment, the CBA metadata includes signaling that indicates a specific OAMD representation to be used in the conversion of the metadata. In an embodiment, pre-computed OAMD is transmitted in a native audio bitstream (e.g., AAC) for transmission (e.g., over HDMI) or for rendering in a source device. In an embodiment, pre-computed OAMD is transmitted in a transport layer bitstream (e.g., ISO BMFF, MPEG4 audio bitstream) to a playback device or source device.

LOUDSPEAKER SYSTEM AND CONTROL

A system for providing an audio interface at a mobile device is provided. The mobile device includes an interface programmed detect a loudspeaker system. The mobile device presents, via a user interface, a display screen to receive user input of sweet-spot commands. The mobile devices send sweet-spot parameters to the loudspeaker system in response to the sweet-spot commands.

In-Car Adaptive Sound Quality Output Method, Device, Storage Medium and Car Audio System
20230239652 · 2023-07-27 ·

The application relates to a sound quality output method. The method includes: obtaining the current volume level input by the user; determine the audio signal gain difference between the current volume level and the baseline volume level according to the current volume level and the preset baseline volume level, determining an equal loudness curve difference between the current volume level and the baseline volume level according to the audio signal gain difference, and the current volume level and the baseline volume level, determining the audio quality response curve corresponding to the current volume level according to the equal loudness curve difference and the pre-stored audio quality response curve corresponding to the baseline volume level, determining the output signal parameters of multiple audio output channels according to the audio quality response curve corresponding to the current volume level. The application optimizes the sound quality effect and enhances acoustic experience.

COLORLESS GENERATION OF ELEVATION PERCEPTUAL CUES USING ALL-PASS FILTER NETWORKS
20230025801 · 2023-01-26 ·

A system includes one or more computing devices that encode spatial perceptual cues into a monaural channel to generate a plurality of output channels. A computing device determines a target amplitude response for the mid and side channels of the plurality of output channels, defining a spatial perceptual associated with one or more frequency-dependent phase shifts. The computing device determines a transfer function of a single-input, multi-output allpass filter based on the target amplitude response and determines coefficients of the allpass filter based on the transfer function, and processes the monaural channel with the coefficients of the allpass filter to generate the plurality of channels having the encoded spatial perceptual cues. The allpass filter is configured to be colorless with respect to the individual output channels, allowing for the placement of spatial cues into the audio stream to be decoupled from the overall coloration of the audio.

Techniques for reducing latency in a wireless home theater environment
11709652 · 2023-07-25 · ·

A first playback device can include a wireless network interface, an audio input interface, one or more processors, and data storage. The data storage stores instructions that, when executed by the processors, cause the first playback device to determine a first radio frequency (RF) energy level associated with RF signal communications from a second playback device to the first playback device. The first playback device modifies a threshold RF energy level for holding off transmissions by the first playback device based on the first RF energy level. The first playback device receives multi-channel audio content via the audio input interface and detects an ambient RF energy level. Based on the ambient RF energy level and the threshold RF energy level, data that represents a channel of the multi-channel audio content is communicated by the first playback device to the second playback device for playback by the second playback device in synchrony with playback of one or more other channels of the multi-channel audio content by the first playback device.

ARRANGEMENT FOR DISTRIBUTING HEAD RELATED TRANSFER FUNCTION FILTERS
20230232179 · 2023-07-20 ·

Arrangement for distributing head related transfer function filters. In the arrangement a user device sends a request for a head related transfer function filter to the service being used. The service verifies if the user of the device has a subscription for a head related transfer function filters in the service being used and retrieves a filter as a response to a positive verification result. The service may filter audio channels and transmit filtered audio further. In an alternative embodiment the service transmits the filter to the user device for filtering the audio.

Methods and apparatus for rendering audio objects

Multiple virtual source locations may be defined for a volume within which audio objects can move. A set-up process for rendering audio data may involve receiving reproduction speaker location data and pre-computing gain values for each of the virtual sources according to the reproduction speaker location data and each virtual source location. The gain values may be stored and used during “run time,” during which audio reproduction data are rendered for the speakers of the reproduction environment. During run time, for each audio object, contributions from virtual source locations within an area or volume defined by the audio object position data and the audio object size data may be computed. A set of gain values for each output channel of the reproduction environment may be computed based, at least in part, on the computed contributions. Each output channel may correspond to at least one reproduction speaker of the reproduction environment.