H04S3/008

Capturing and synchronizing data from multiple sensors

Processes, methods, systems, and devices are disclosed for synchronizing multiple wireless data streams captured in action by various sensors, with lost data recovery. For example, a source device may have multiple sensors acquiring data and sending the data streams (e.g., via Bluetooth connections) to a target device. Timing information may be appended for each of the data streams. Data packets of the multiple data streams may be formed with the timing information. The data packets may be transmitted to a target device that is configured to synchronize the multiple data streams using the timing information. The target device, applying the example processes or techniques of this disclosure, may accurately synchronize the multiple data streams. In some cases, the target device may capture additional data streams and the processor synchronizes all data streams of both the source and the target devices.

Determining corrections to be applied to a multichannel audio signal, associated coding and decoding
20220358937 · 2022-11-10 ·

A method and device for determining a set of corrections to be made to a multichannel sound signal, in which the set of corrections is determined on the basis of an item of information representative of a spatial image of an original multichannel signal and an item of information representative of a spatial image of the original multichannel signal that has been coded and then decoded.

DYNAMICS PROCESSING ACROSS DEVICES WITH DIFFERING PLAYBACK CAPABILITIES

Individual loudspeaker dynamics processing configuration data, for each of a plurality of loudspeakers of a listening environment, may be obtained. Listening environment dynamics processing configuration data may be determined, based on the individual loudspeaker dynamics processing configuration data. Dynamics processing may be performed on received audio data based on the listening environment dynamics processing configuration data, to generate processed audio data. The processed audio data may be rendered for reproduction via a set of loudspeakers that includes at least some of the plurality of loudspeakers, to produce rendered audio signals. The rendered audio signals may be provided to, and reproduced by, the set of loudspeakers.

METHOD AND APPARATUS FOR RENDERING VOLUME SOUND SOURCE

A method and apparatus for rendering a volume sound source are disclosed. The method of rendering a volume sound source may include identifying information about a listener and information about the volume sound source, determining a corresponding area in which a source element is disposed in the volume sound source in consideration of the information about the listener, determining an angle between the listener and the corresponding area based on the information about the listener and the information about the volume sound source, determining a number of source elements disposed in the corresponding area according to the angle, determining a position and a gain of the source element using i) the number of source elements and ii) a distance between the listener and the volume sound source, and rendering the volume sound source according to the position and the gain of the source element.

SYSTEMS AND METHODS FOR GENERATING VIDEO-ADAPTED SURROUND-SOUND
20220360933 · 2022-11-10 ·

Audiovisual presentations, such as film recordings, may have been originally created having an audio soundtrack with multiple audio tracks mixed for a surround sound system that includes a set of speakers physically surrounding a user. The present disclosure presents systems and methods to remix these soundtracks into 3D audio that when presented to the ears of a user can be perceived as a virtual surround sound system that mimics the physical system. What is more, the disclosed systems and methods can enhance the virtual surround sound system by adjusting virtual speakers of the virtual surround sound system according to video content of the audiovisual presentation. Further enhancement may be possible by adjusting the virtual speakers of the virtual surround sound system according to a sensed position of a user.

SYSTEM AND METHOD FOR ADAPTIVE AUDIO SIGNAL GENERATION, CODING AND RENDERING

Embodiments are described for an adaptive audio system that processes audio data comprising a number of independent monophonic audio streams. One or more of the streams has associated with it metadata that specifies whether the stream is a channel-based or object-based stream. Channel-based streams have rendering information encoded by means of channel name; and the object-based streams have location information encoded through location expressions encoded in the associated metadata. A codec packages the independent audio streams into a single serial bitstream that contains all of the audio data. This configuration allows for the sound to be rendered according to an allocentric frame of reference, in which the rendering location of a sound is based on the characteristics of the playback environment (e.g., room size, shape, etc.) to correspond to the mixer's intent. The object position metadata contains the appropriate allocentric frame of reference information required to play the sound correctly using the available speaker positions in a room that is set up to play the adaptive audio content.

WEARABLE DEVICE AND METHOD FOR CONTROLLING AUDIO OUTPUT USING MULTI DIGITAL TO ANALOG CONVERTER PATH

A wearable device is provided and includes a plurality of speakers including a first speaker, a second speaker, and an N.sup.th speaker, a plurality of digital to analog converter (DAC)s including a first DAC connected to the first speaker, a second DAC connected to the second speaker, and an N.sup.th DAC connected to the N.sup.th speaker, an audio signal processing module including N DAC output paths configured to filter an audio signal according to each frequency band and output the audio signal, a memory; and a processor electrically connected to the plurality of DACs, the audio signal processing module, and the memory, wherein the memory includes instructions causing the processor to, when the audio signal is reproduced, analyze a frequency component included in the audio signal, activate the N DAC output paths when the frequency component included in the audio signal has a full band range, activate only a DAC output path for processing a specific frequency band among the N DAC output paths when the frequency component included in the audio signal has only the specific frequency band, and output the audio signal through a speaker connected to the activated DAC output path.

Parametric joint-coding of audio sources

The following coding scenario is addressed: A number of audio source signals need to be transmitted or stored for the purpose of mixing wave field synthesis, multi-channel surround, or stereo signals after decoding the source signals. The proposed technique offers significant coding gain when jointly coding the source signals, compared to separately coding them, even when no redundancy is present between the source signals. This is possible by considering statistical properties of the source signals, the properties of mixing techniques, and spatial hearing. The sum of the source signals is transmitted plus the statistical properties of the source signals, which mostly determine the perceptually important spatial cues of the final mixed audio channels. Source signals are recovered at the receiver such that their statistical properties approximate the corresponding properties of the original source signals. Subjective evaluations indicate that high audio quality is achieved by the proposed scheme.

Methods and systems for recording mixed audio signal and reproducing directional audio

Methods and systems are provided for recording mixed audio signal and reproducing directional audio. A method includes receiving a mixed audio signal via plurality of microphones; determining an audio parameter associated with the mixed audio signal received at each of the plurality of microphones; determining active audio sources and a number of the active audio sources from the mixed audio signal; determining direction and positional information of each of the active audio source; dynamically selecting a set of microphones from the plurality of microphones based on at least one of the number of the active audio sources, the direction of each of the active audio sources, the positional information of each of the active audio sources, the audio parameter, or a predefined condition; and recording, based on the selected set of microphones, the mixed audio signal for reproducing directional audio.

NEAR-FIELD AUDIO RENDERING

Examples of the disclosure describe systems and methods for presenting an audio signal to a user of a wearable head device. According to an example method, a source location corresponding to the audio signal is identified. For each of the respective left and right ear of the user, a virtual speaker position, of a virtual speaker array, is determined, the virtual speaker position collinear with the source location and with a position of the respective ear. For each of the respective left and right ear of the user, a head-related transfer function (HRTF) corresponding to the virtual speaker position and to the respective ear is determined; and the output audio signal is presented to the respective ear of the user via one or more speakers associated with the wearable head device. Processing the audio signal includes applying the HRTF to the audio signal.