H04S2420/01

DELAYED AUDIO FOLLOWING
20230020792 · 2023-01-19 ·

Disclosed herein are systems and methods for presenting mixed reality audio. In an example method, audio is presented to a user of a wearable head device. A first position of the user's head at a first time is determined based on one or more sensors of the wearable head device. A second position of the user's head at a second time later than the first time is determined based on the one or more sensors. An audio signal is determined based on a difference between the first position and the second position. The audio signal is presented to the user via a speaker of the wearable head device. Determining the audio signal comprises determining an origin of the audio signal in a virtual environment. Presenting the audio signal to the user comprises presenting the audio signal as if originating from the determined origin. Determining the origin of the audio signal comprises applying an offset to a position of the user's head.

Sound Localization for an Electronic Call
20230224658 · 2023-07-13 ·

During an electronic call between two individuals, a sound localization point simulates a location in empty space from where an origin of a voice of one individual occurs for the other individual.

ARRANGEMENT FOR PRODUCING HEAD RELATED TRANSFER FUNCTION FILTERS
20230222819 · 2023-07-13 ·

When three-dimensional audio is produced by using headphones, particular HRTF-filters are used to modify sound for the left and right channels of the headphone. As the morphology of every ear is different, it is beneficial to have HRTF-filters particularly designed for the user of headphones. Such filters may be produced by deriving ear geometry from a plurality of images taken with an ordinary camera, detecting necessary features from images and fitting said features to a model that has been produced from accurately scanned ears comprising representative values for different sizes and shapes. Taken images are sent to a server (52) that performs the necessary computations and submits the data further or produces the requested filter.

Supplementing Content

An apparatus, method and computer program product for: providing spatial audio content for output via at least one loudspeaker, determining a position of at least one audio device operatively connected to the at least one loudspeaker, providing, in response to determining that the position of the at least one audio device corresponds to an audio zone associated with additional spatial audio content, the additional spatial audio content for output via the at least one audio device, receiving an instruction to include the additional spatial audio content in the spatial audio content, and supplementing, in response to receiving the instruction to include the additional spatial content in the spatial audio content, the spatial audio content with the additional spatial audio such that the additional spatial audio content is provided for output independent of the audio zone.

Multi-viewpoint multi-user audio user experience

An apparatus including circuitry configured for receiving a spatial media content file including a plurality of viewpoints; circuitry configured for determining a first viewpoint from the plurality of viewpoints for a first user consuming the spatial media content file; circuitry configured for receiving an indication that affects an audio rendering of the first viewpoint for the first user, wherein the indication is associated with one or more actions of at least one second user consuming the spatial media content file; and circuitry configured for controlling the audio rendering of the first viewpoint for the first user in response to the receiving of the indication based on at least one of: a position and/or orientation of the first user, and the one or more actions of the second user.

SYSTEMS AND METHODS FOR HEAD RELATED TRANSFER FUNCTION PERSONALIZATION

A head-related transfer function (HRTF) generation system includes one or more processors configured to retrieve first image data of a first ear of a subject, compare the first image data with second image data of a plurality of second ears to identify a particular second ear of the plurality of second ears matching the first ear, identify a template HRTF associated with the particular second ear, and assign an HRTF to the subject based on the template HRTF.

3D Spatialisation of Voice Chat

The invention provides techniques for intelligently positioning speech in a virtual environment. Factors such as user preferences, the location of virtual environment audio and/or visual events, avatar location, and others can be taken into account when selecting a suitable location for the speech. The virtual environment can be a game environment, a meeting environment, an augmented reality environment, a virtual reality environment, and the like. The invention can be implemented by an audio processing unit which may be part of a game console.

Sound field adjustment

A device includes one or more processors configured to receive, via wireless transmission from a streaming device, encoded ambisonics audio data representing a sound field. The one or more processors are also configured to perform decoding of the ambisonics audio data to generate decoded ambisonics audio data. The decoding of the ambisonics audio data includes base layer decoding of a base layer of the encoded ambisonics audio data and selectively includes enhancement layer decoding in response to an amount of movement of the device. The one or more processors are further configured to adjust the decoded ambisonics audio data to alter the sound field based on data associated with at least one of a translation or an orientation associated with the movement of the device. The one or more processors are also configured to output the adjusted decoded ambisonics audio data to two or more loudspeakers for playback.

GENERATING SPATIAL AUDIO AND CROSS-TALK CANCELLATION FOR HIGH-FREQUENCY GLASSES PLAYBACK AND LOW-FREQUENCY EXTERNAL PLAYBACK
20230224663 · 2023-07-13 ·

The present disclosure relates to systems, methods, and non-transitory computer-readable media that present spatial audio using the speakers of a wearable audio device and speakers external to the wearable audio device. In particular, in one or more embodiments, the disclosed systems generate spatial audio having a high-frequency component and a low-frequency component. The disclosed systems further generate cross-talk cancellation filters for the low-frequency component of the spatial audio. The disclosed systems can provide the high-frequency component for presentation via speakers of the wearable audio device and the low-frequency component for presentation via the external speakers using the cross-talk cancellation filters. In some cases, the disclosed systems generate the spatial audio or the cross-talk cancellation filters using a personalized interaural display model and/or head-related transfer functions.

APPARATUS FOR IMMERSIVE SPATIAL AUDIO MODELING AND RENDERING

Disclosed is an apparatus for immersive spatial audio modeling and rendering for effectively transmitting and playing immersive spatial audio content. The apparatus for immersive spatial audio modeling and rendering disclosed herein may model a spatial audio scene, generate and transmit parameters necessary for spatial audio rendering, and generate various spatial audio effects using the spatial audio parameters, to provide an immersive three-dimensional (3D) audio source coinciding with visual experience in a virtual reality space in response to free changes in the position and direction of a remote user in the space.