H04S2400/11

QUANTIZATION OF SPATIAL AUDIO DIRECTION PARAMETERS
20220386056 · 2022-12-01 ·

A method for spatial audio signal encoding comprising: obtaining a plurality of audio direction parameters, wherein each parameter comprises an elevation value and an azimuth value and wherein each parameter has an ordered position; deriving for each of the plurality of audio direction parameters a corresponding derived audio direction parameter (SP) comprising an elevation and an azimuth value, corresponding derived audio direction parameters (SP) being arranged in a manner determined by a spatial utilization defined by the elevation values and the azimuth values of the plurality of audio direction parameters; rotating each derived audio direction parameter (SP) by the azimuth value (φ.sub.0) of an audio direction parameter in the first position of the plurality of audio direction parameters and quantizing the rotation to determine for each a corresponding quantized rotated derived audio direction parameter; changing the ordered position of an audio direction parameter to a further position coinciding with a position of a rotated derived audio direction parameter when the azimuth value of the audio direction parameter is closest to the azimuth value of the further rotated derived audio direction parameter compared to the azimuth values of other rotated derived audio direction parameters, followed by determining for each of the plurality audio direction parameters a difference between each audio direction parameter and their corresponding quantized rotated derived audio direction parameter; and quantizing a difference for each of the plurality of audio direction parameters, wherein a difference quantization resolution for each of the plurality of audio direction parameters is defined based on a spatial extent of the audio direction parameters.

SIGNALLING OF AUDIO EFFECT METADATA IN A BITSTREAM

Methods, systems, computer-readable media, and apparatuses for manipulating a soundfield are presented. Some configurations include receiving a bitstream that comprises metadata and a soundfield description; parsing the metadata to obtain an effect identifier and at least one effect parameter value; and applying, to the soundfield description, an effect identified by the effect identifier. The applying may include using the at least one effect parameter value to apply the identified effect to the soundfield description.

TRAINING DATA EXTENSION APPARATUS, TRAINING DATA EXTENSION METHOD, AND PROGRAM

An input of a first observation signal corresponding to an incoming signal from a first direction is received, an angular rotation operation of the first observation signal is performed to obtain a second observation signal corresponding to an incoming signal from a second direction that is different from the first direction and the second observation signal is added to a set of training data.

EXTRACTION OF AN AUDIO OBJECT
20220383894 · 2022-12-01 · ·

A method for extracting at least one audio object from at least two audio input signals, each of which contains the audio object. The second audio input signal is syncronized with the first audio input signal while obtaining a synchronized second audio input signal. The audio object is extracted by applying at least one trained model to the first audio signal and to the synchronized second audio input signal. The audio object is outputted. Further, the step of synchronizing the second audio input signal with the first audio input signal includes the steps of: generating audio signals; analytically calculating a correlation between the audio signals; optimizing the correlation vector; and determining the synchronized second audio input signal using the optimized correlation vector.

IMMERSIVE AUDIO PLATFORM

Disclosed herein are systems and methods for presenting audio content in mixed reality environments. A method may include receiving a first input from an application program; in response to receiving the first input, receiving, via a first service, an encoded audio stream; generating, via the first service, a decoded audio stream based on the encoded audio stream; receiving, via a second service, the decoded audio stream; receiving a second input from one or more sensors of a wearable head device; receiving, via the second service, a third input from the application program, wherein the third input corresponds to a position of one or more virtual speakers; generating, via the second service, a spatialized audio stream based on the decoded audio stream, the second input, and the third input; presenting, via one or more speakers of the wearable head device, the spatialized audio stream.

AUDIO PROCESSING METHOD AND APPARATUS, READABLE MEDIUM, AND ELECTRONIC DEVICE
20220386061 · 2022-12-01 ·

Provided are an audio processing method and apparatus, a readable medium, and an electronic device. The method includes: acquiring an original image captured by a terminal; determining a three-dimensional relative position of a target object relative to the terminal as a first three-dimensional relative position according to the original image; and performing three-dimensional effect processing on a target sound according to the first three-dimensional relative position to enable a sound source position of the target sound in audio obtained after the three-dimensional effect processing and the first three-dimensional relative position to conform to a positional relationship between the target object and a sound effect object corresponding to the target object, where the target sound is an effect sound corresponding to the sound effect object.

SPATIAL AUDIO MONAURALIZATION VIA DATA EXCHANGE
20220386054 · 2022-12-01 ·

A device includes a memory configure to store instructions and one or more processors configured to execute the instructions to obtain spatial audio data at a first audio output device. The one or more processors are further configured to perform data exchange, between the first audio output device and a second audio output device, of exchange data based on the spatial audio data. The one or more processors are also configured to generate first monaural audio output at the first audio output device based on the spatial audio data.

STEREOPHONIC AUDIO REARRANGEMENT BASED ON DECOMPOSED TRACKS
20220386062 · 2022-12-01 ·

The present invention provides a method for processing audio data, comprising providing input audio data containing a mixture of different timbres, decomposing the input audio data to generate decomposed data representing a predetermined timbre selected from the timbres contained in the input audio data, determining a set point position of a virtual sound source outputting the predetermined timbre relative to a position of a virtual listener, and generating stereophonic output data based on the decomposed data and the determined set point position.

CONVEYING MOTION DATA VIA MEDIA PACKETS

A device includes a memory configured to store instructions and one or more processors configured to execute the instructions to receive a media packet and to determine, based on a field of the media packet, whether the media packet includes motion data. The one or more processors are also configured to execute the instructions to, based on the media packet including motion data, extract the motion data from the media packet.

APPARATUS AND METHOD FOR PROCESSING MULTI-CHANNEL AUDIO SIGNAL

An apparatus for processing audio includes at least one processor configured to obtain a down-mixed audio signal from a bitstream, to obtain down-mixing-related information from the bitstream, to de-mix the down-mixing-related information by using down-mixing-related information, and to reconstruct an audio signal including at least one frame based on the de-mixed audio signal. The down-mixing-related information is information generated in units of frames by using an audio scene type.