Patent classifications
H04S3/00
Efficient coding of audio scenes comprising audio objects
There is provided encoding and decoding methods for encoding and decoding of object based audio. An exemplary encoding method includes inter alia calculating M downmix signals by forming combinations of N audio objects, wherein M≤N, and calculating parameters which allow reconstruction of a set of audio objects formed on basis of the N audio objects from the M downmix signals. The calculation of the M downmix signals is made according to a criterion which is independent of any loudspeaker configuration.
Calibrating listening devices
Techniques for calibrating listening devices are disclosed herein. The techniques include emitting a predetermined audio signal using an outward-facing transducer located on a first portion of a head-mounted device worn by the user, receiving the predetermined audio signal at a microphone located on a second portion of the head-mounted device, the second portion being different from the first portion, determining a transfer function for the user based on the received predetermined audio signal, and applying the transfer function to audio signals transmitted to the user.
Display apparatus
A display apparatus is capable of outputting a stereo sound. The display apparatus includes a display panel configured to display an image; a sound generating device on a rear surface of the display panel; a rear cover on the rear surface of the display panel and configured to support the sound generating device; a partition member between the rear surface of the display panel and the rear cover and configured to divide the display panel into first, second, third, fourth and fifth areas; and first, second, third, fourth, and fifth sound generating devices attached to the rear surface of the display panel and configured to vibrate the display panel. The first, second, third, fourth and fifth sound generating devices are in the first, second, third, fourth and fifth areas, respectively.
Audio Generation Methods and System
A method of generating audio assets, comprising the steps of: receiving an input multi-layered audio asset comprising a plurality of audio layers, generating an input multi-channel image, wherein each channel of the input multi-channel image comprises an input image representative of one of the audio layers, training a generative model on the input multi-channel image and implementing the trained generative model to generate an output multi-channel image, wherein each channel of the output multi-channel image comprises an output image representative of an output audio layer, and generating an output multi-layered audio asset based on a combination of output audio layers derived from the output images.
SPATIALIZED AUDIO CHAT IN A VIRTUAL METAVERSE
Implementations described herein relate to methods, systems, and computer-readable media to provide spatialized audio in virtual experiences. The spatialized audio may be used in voice communications such as, for example, voice and/or video chats. The chats may include spatialized audio that is combined at a client device, or at an online experience platform, and is targeted to a particular user. Individual audio streams may be collected from a plurality of avatars and other objects, and combined based on the target user. The audio may also include background and/or ambient sounds to provide a rich, immersive audio stream in virtual experiences.
METHOD AND SYSTEM FOR ARTIFICIAL REVERBERATION EMPLOYING REVERBERATION IMPULSE RESPONSE SYNTHESIS
The present embodiments relate to audio effect processing, and more particularly to a method for creating reverberation impulse responses from prerecorded or live source materials forms the basis of a family of reverberation effects. In one embodiment, segments of audio are selected and processed to form an evolving sequence of reverberation impulse responses that are applied to the original source material—that is, an audio stream reverberating itself. In another embodiment, impulse responses derived from one audio track are applied to another audio track. In a further embodiment, reverberation impulse responses are formed by summing randomly selected segments of the source audio, and imposing reverberation characteristics, including reverberation time, wet equalization, wet-dry mix, and predelay. By controlling the number and timing of the selected source audio segments, the method produces a collection of impulse responses that represent a trajectory through the source material. In so doing, the evolving impulse responses will have the character of room reverberation while also expressing the changing timbre and dynamics of the source audio.
Sound Localization for an Electronic Call
During an electronic call between two individuals, a sound localization point simulates a location in empty space from where an origin of a voice of one individual occurs for the other individual.
SURFACE APPROPRIATE COLLISIONS
Disclosed herein are systems and methods for presenting an audio signal associated with presentation of a virtual object colliding with a surface. The virtual object and the surface may be associated with a mixed reality environment. Generation of the audio signal may be based on at least one of an audio stream from a microphone and a video stream form a sensor. In some embodiments, the collision between the virtual object and the surface is associated with a footstep on the surface.
METHOD AND APPARATUS FOR AMBISONIC SIGNAL REPRODUCTION IN VIRTUAL REALITY SPACE
Provided is a method of reproducing an ambisonic signal in a virtual reality (VR) space. The ambisonic signal reproduction method may include receiving an ambisonic signal, mapping the ambisonic signal to channels localized on a sphere according to an equivalent spatial domain (ESD) standard corresponding to an order of the ambisonic signal, and performing a sound field reproduction in the VR space based on the channels localized on the sphere.
VIRTUAL AND MIXED REALITY AUDIO SYSTEM ENVIRONMENT CORRECTION
A virtual reality (VR), augmented reality (AR) and/or mixed reality (MR) system in a physical environment with a plurality of loudspeakers includes a user-worn head mounted display (HMD), a VR/AR/MR processor, and a VR/AR/MR user tracking processor. The HMD includes a microphone and a user tracking device configured to track a user orientation and position. The VR/AR/MR processor delivers a digital video signal to the head-mounted display, and a digital control signal and a digital audio signal to a receiver/preamplifier. The VR/AR/MR user tracking processor receives user tracking data from the HMD user tracking device and provides a digital user tracking data signal to the receiver preamplifier. the receiver/preamplifier receives the digital user tracking data signal, the digital control signal, the digitized microphone signal, and the digital audio signal, and provides a processed audio signal to the amplifier. An amplifier receives the processed audio signal and provides amplified audio signals.