H04S7/30

DEVICES AND METHODS FOR AUDITORY REHABILITATION FOR INTERAURAL ASYMMETRY
20230049597 · 2023-02-16 ·

A device, system and related methods to provide assessment and treatment of amblyaudia through standardized methods that do not require advanced training or a booth with loudspeakers for the operator to administer. The ARIA stimuli protocols for both assessment and treatment, encoded in or to be used by a software program or application, are transferred to a stand-alone set of specialized noise-cancelling headphones attached or connected to, wired or wirelessly, a software platform on an electronic computing device. or integrated with the headphones. The program administers assessment tests to individuals through the noise-cancelling earphones. The device enables someone with minimal instructions to administer automatically or semi-automatically both assessment and treatment protocols, generate results, make interpretations, store data, and produce reports. The device or system may be loaded with standard protocols for English-speaking individuals, as well as dichotic speech material in any language.

Stereo Sound Pickup Method and Apparatus, Terminal Device, and Computer-Readable Storage Medium
20230048860 · 2023-02-16 ·

A stereo sound pickup method and apparatus, a terminal device, and a computer-readable storage medium. The method includes configuring a terminal device to record a video, wherein the terminal device comprises a plurality of microphones, configuring the plurality of microphones to capture a sound, and forming a stereo beam based on the captured sound. The stereo beam is related to a video recording scenario of the terminal device, and the video recording scenario includes a posture of the terminal device and usage of a camera, the posture includes that the terminal device is in a landscape mode or a portrait mode, and the usage of the camera includes that a rear-facing camera is used or a front-facing camera is used.

MULTI-TRACK AUDIO IN A SECURITY SYSTEM
20230049433 · 2023-02-16 ·

A method, system, server and device are disclosed. According to one or more embodiments, a server is provided. A first audio track is received which includes first audio originating from a premises client at a premises location. A second audio track is received which includes second audio originating from a remote client. A first pan angle is determined for the first audio track and a second pan angle is determined for the second audio track. The second pan angle is different from the first pan angle. A stereo composite track is generated based on the first pan angle and the second pan angle, where the stereo composite track includes the first audio track and the second audio track.

Hearing system comprising a personalized beamformer

A hearing system configured to be located at or in the head of a user, comprises a) at least two microphones providing at least two electric input signals, b) an own voice detector, c) access to a database (O.sub.l, H.sub.l) comprising c1) relative or absolute own voice transfer function(s), and corresponding c2) absolute or relative acoustic transfer functions for a multitude of test-persons, d) a processor connectable to the at least two microphones, to the own voice detector, and to the database. The processor is configured A) to estimate an own voice relative transfer function for sound from the user's mouth to at least one of the at least two microphones, and B) to estimate personalized relative or absolute head related acoustic transfer functions from at least one spatial location other than the user's mouth to at least one of the microphones of the hearing system in dependence of the estimated own voice relative transfer function(s) and the database (O.sub.l, H.sub.l). The hearing system further comprises e) a beamformer configured to receive the at least two electric input signals, or processed versions thereof, and to determine personalized beamformer weights based on the personalized relative or absolute head related acoustic transfer functions or impulse responses. A method of determining personalized beamformer coefficients (w.sub.k) is further disclosed.

Reconstruction of audio scenes from a downmix

Audio objects are associated with positional metadata. A received downmix signal comprises downmix channels that are linear combinations of one or more audio objects and are associated with respective positional locators. In a first aspect, the downmix signal, the positional metadata and frequency-dependent object gains are received. An audio object is reconstructed by applying the object gain to an upmix of the downmix signal in accordance with coefficients based on the positional metadata and the positional locators. In a second aspect, audio objects have been encoded together with at least one bed channel positioned at a positional locator of a corresponding downmix channel. The decoding system receives the downmix signal and the positional metadata of the audio objects. A bed channel is reconstructed by suppressing the content representing audio objects from the corresponding downmix channel on the basis of the positional locator of the corresponding downmix channel.

Generating binaural audio in response to multi-channel audio using at least one feedback delay network

In some embodiments, virtualization methods for generating a binaural signal in response to channels of a multi-channel audio signal, which apply a binaural room impulse response (BRIR) to each channel including by using at least one feedback delay network (FDN) to apply a common late reverberation to a downmix of the channels. In some embodiments, input signal channels are processed in a first processing path to apply to each channel a direct response and early reflection portion of a single-channel BRIR for the channel, and the downmix of the channels is processed in a second processing path including at least one FDN which applies the common late reverberation. Typically, the common late reverberation emulates collective macro attributes of late reverberation portions of at least some of the single-channel BRIRs. Other aspects are headphone virtualizers configured to perform any embodiment of the method.

Sound effect simulation by creating virtual reality obstacle

According to one embodiment, a method, computer system, and computer program product for modulating external sounds to reflect the acoustic effects of virtual objects in a mixed-reality environment is provided. The present invention may include creating a knowledge corpus, recording a sound effect occurring externally to a mixed-reality environment experienced by a user operating the mixed-reality device; identifying one or more objects within the mixed-reality environment, including at least one virtual object; modulating the sound effect based on the knowledge corpus to simulate one or more acoustic effects of the one or more objects within the MR environment; and playing the modulated sound effect to the user.

METHOD AND SYSTEM FOR INSTRUMENT SEPARATING AND REPRODUCING FOR MIXTURE AUDIO SOURCE

A method and a system for instrument separating and reproducing for a mixture audio source is provided. The method and/or the system includes inputting selected music into an instrument separation model for extracting features therefrom, determining audio source signals of multiple channels for the separation of all instruments, each channel containing sound of one instrument, and transmitting the signals of the different channels to multiple speakers placed at designated positions for playing, which can reproduce or recreate an immersive sound field listening experience for users.

SYSTEM AND METHOD FOR VIRTUAL SOUND EFFECT WITH INVISIBLE LOUDSPEAKER(S)

In at least one embodiment, an apparatus for providing a virtual sound effect in a listening environment is provided. The apparatus includes at least one controller and an audio playback device. The audio playback device includes the at least one controller that is programmed to receive an audio input signal from an audio input source and to apply a head related transfer function (HRTF) to the audio input signal. The at least one controller is further programmed to apply crosstalk cancellation to the audio input signal and to generate an audio output signal after applying the HRTF and the crosstalk cancellation to the audio input signal for playback by at least one loudspeaker that is invisible to a listener in the listening environment.

ADJUSTING AUDIO AND NON-AUDIO FEATURES BASED ON NOISE METRICS AND SPEECH INTELLIGIBILITY METRICS

Some implementations involve determining a noise metric and/or a speech intelligibility metric and determining a compensation process corresponding to the noise metric and/or the speech intelligibility metric. The compensation process may involve altering a processing of audio data and/or applying a non-audio-based compensation method. In some examples, altering the processing of the audio data does not involve applying a broadband gain increase to the audio signals. Some examples involve applying the compensation process in an audio environment. Other examples involve determining compensation metadata corresponding to the compensation process and transmitting an encoded content stream that includes encoded compensation metadata, encoded video data and encoded audio data from a first device to one or more other devices.