Patent classifications
H04S2420/11
SPATIAL AUDIO PARAMETER ENCODING AND ASSOCIATED DECODING
An apparatus comprising means configured to obtain direction parameter values (108) associated with at least two time-frequency parts (202) of at least one audio signal (102); and encode the obtained direction parameter values based on a codebook (206), wherein the codebook comprises two or more quantization levels arranged such that a first quantization level comprises a first set of quantization values, and a second or succeeding quantization level comprises a second or further set of quantization values and preceding quantization level quantization values.
XR RENDERING FOR 3D AUDIO CONTENT AND AUDIO CODEC
A device includes a memory configured to store instructions and also includes one or more processors configured to execute the instructions to obtain audio data corresponding to a sound source and metadata indicative of a direction of the sound source. The one or more processors are configured to execute the instructions to obtain direction data indicating a viewing direction associated with a user of a playback device. The one or more processors are configured to execute the instructions to determine a resolution setting based on a similarity between the viewing direction and the direction of the sound source. The one or more processors are also configured to execute the instructions to process the audio data based on the resolution setting to generate processed audio data.
METHOD FOR DECODING A HIGHER ORDER AMBISONICS (HOA) REPRESENTATION OF A SOUND OR SOUNDFIELD
When compressing an HOA data frame representation, a gain control (15, 151) is applied for each channel signal before it is perceptually encoded (16). The gain values are transferred in a differential manner as side information. However, for starting decoding of such streamed compressed HOA data frame representation absolute gain values are required, which should be coded with a minimum number of bits. For determining such lowest integer number (β.sub.e) of bits the HOA data frame representation (C(k)) is rendered in spatial domain to virtual loudspeaker signals lying on a unit sphere, followed by normalisation of the HOA data frame representation (C(k)). Then the lowest integer number of bits is set to
β.sub.e=┌log.sub.2(┌log.sub.2(√{square root over (K.sub.MAX)}.Math.O)┐+1)┐.
METHOD FOR DECODING A HIGHER ORDER AMBISONICS (HOA) REPRESENTATION OF A SOUND OR SOUNDFIELD
When compressing an HOA data frame representation, a gain control (15, 151) is applied for each channel signal before it is perceptually encoded (16). The gain values are transferred in a differential manner as side information. However, for starting decoding of such streamed compressed HOA data frame representation absolute gain values are required, which should be coded with a minimum number of bits. For determining such lowest integer number (β.sub.e) of bits the HOA data frame representation (C(k)) is rendered in spatial domain to virtual loudspeaker signals lying on a unit sphere, followed by normalisation of the HOA data frame representation (C(k)). Then the lowest integer number of bits is set to
β.sub.e=┌log.sub.2(┌log.sub.2(√{square root over (K.sub.MAX)}.Math.O)┐+1)┐.
Audio Response Based on User Worn Microphones to Direct or Adapt Program Responses System and Method
A system, method, and wireless earpieces for communicating with a virtual reality headset. A position and an orientation of a head of a user are detected utilizing at least wireless earpieces. Audio content is received. The audio content is enhanced utilizing the position and the orientation of the head of the user. The audio content is immediately delivered to the user. The method may further include communicating the position and the orientation of the head of the user to the virtual reality headset. The audio content may be based on the orientation and position of the head of the user.
AUDIO SCENE CHANGE SIGNALING
There is disclosed inter alia a method for rendering a virtual reality audio scene comprising: receiving information defining a limited area audio scene within the virtual reality audio scene (301), wherein the limited area audio scene defines a sub space of the virtual reality audio scene (304), wherein the information defines the limited area audio scene by defining an extent a user can move within the virtual audio scene; determining if the movement of the user within the limited area audio scene meets a condition of an audio scene change (302); and processing the audio scene change when the movement of the user within the limited area audio scene meets the condition of an audio scene change (306).
VIDEO PROCESSING DEVICE AND METHOD
A video processing apparatus includes a memory storing instructions, and at least one processor configured to execute the instructions to generate a plurality of feature information by analyzing a video signal comprising a plurality of images based on a first DNN, extract a first altitude component and a first planar component corresponding to a movement of an object in a video from the video signal based on a second DNN, extract a second planar component corresponding to a movement of a sound source in audio from a first audio signal based on a third DNN, generate a second altitude component based on the first altitude component, the first planar component, and the second planar component, output a second audio signal comprising the second altitude component based on the feature information, and synchronize the second audio signal with the video signal and output the synchronized second audio signal and video signal.
Augmented hearing system
Some implementations may involve receiving, via an interface system, personnel location data indicating a location of at least one person and receiving, from an orientation system, headset orientation data corresponding with the orientation of a headset. First environmental element location data, indicating a location of at least a first environmental element, may be determined. Based at least in part on the headset orientation data, the personnel location data and the first environmental element location data, headset coordinate locations of at least one person and at least the first environmental element in a headset coordinate system corresponding with the orientation of the headset may be determined. An apparatus may be caused to provide spatialization indications of the headset coordinate locations. Providing the spatialization indications may involve controlling a speaker system to provide environmental element sonification corresponding with at least the first environmental element location data.
Apparatus, method, computer program for enabling access to mediated reality content by a remote user
An apparatus comprising means for: simultaneously controlling content rendered by a hand portable device and content rendered by a spatial audio device; and providing for rendering to a user, in response to an action by the user, of a first part, not a second part, of a spatial audio content via the hand portable device not the spatial audio device.
Spatial audio for interactive audio environments
Systems and methods of presenting an output audio signal to a listener located at a first location in a virtual environment are disclosed. According to embodiments of a method, an input audio signal is received. For each sound source of a plurality of sound sources in the virtual environment, a respective first intermediate audio signal corresponding to the input audio signal is determined, based on a location of the respective sound source in the virtual environment, and the respective first intermediate audio signal is associated with a first bus. For each of the sound sources of the plurality of sound sources in the virtual environment, a respective second intermediate audio signal is determined. The respective second intermediate audio signal corresponds to a reverberation of the input audio signal in the virtual environment. The respective second intermediate audio signal is determined based on a location of the respective sound source, and further based on an acoustic property of the virtual environment. The respective second intermediate audio signal is associated with a second bus. The output audio signal is presented to the listener via the first bus and the second bus.