HEARING DEVICE OR SYSTEM COMPRISING A NOISE CONTROL SYSTEM

20230143325 · 2023-05-11

Assignee

Inventors

Cpc classification

International classification

Abstract

A hearing system comprises a hearing device, e.g. a hearing aid or a headset, configured to be worn by a user. The hearing device comprises at least one input transducer for providing at least one electric input signal representative of sound in the environment of the hearing device, wherein said at least one electric input signal comprises a target signal component assumed to be of current interest to the user, and a noise component. The hearing device further comprises a noise control system configured to provide an estimate of said target signal component and an estimate of said noise component and to apply a statistical structure to said noise component to thereby provide a modified noise component comprising said statistical structure; and to determine a modified estimate of said target signal component in dependence of said modified noise component. Thereby an improved segregation of sound sources may be provided.

Claims

1. A hearing system comprising A hearing device configured to be worn by a user, the hearing device comprising At least one input transducer for providing at least one electric input signal representative of sound in the environment of the hearing device, wherein said at least one electric input signal comprises a) a target signal component assumed to be of current interest to the user, and b) a noise component; An output unit configured to provide an output signal based on said at least one electric input signal, either comprising stimuli for being presented to the user, and/or for being transmitted to another device; A noise control system configured to provide an estimate of said target signal component and an estimate of said noise component in said at least one electric input signal, or in a signal originating therefrom; wherein the noise control system is further configured to apply a statistical structure to said estimate of said noise component to thereby provide a modified noise component comprising said statistical structure; and to determine a modified estimate of said target signal component in dependence of said modified noise component, and wherein said output signal comprises said modified estimate of said target signal component, or a further processed version thereof.

2. A hearing system according to claim 1 wherein said noise control system is configured to apply said statistic structure to said noise component by modulation.

3. A hearing system according to claim 1 wherein said statistical structure is constituted by or comprises auditory texture in the form of sounds produced by the addition of a multitude of similar sound sources.

4. A hearing system according to claim 1 wherein said statistical structure is constituted by or comprises amplitude modulation in a rhythmic pattern.

5. A hearing system according to claim 1 wherein said at least one input transducer comprises a multitude of input transducers, each providing an electric input signal representative of sound in the environment of the hearing device, and wherein said noise control system comprises a directional system comprising at least one beamformer configured to receive as inputs said multitude of electric input signals, or signals originating therefrom, and to provide an estimate of said target signal component in dependence of said inputs and predefined or adaptively updated beamformer weights.

6. A hearing system according to claim 5 wherein said directional system comprises a linear constraint minimum variance (LCMV) beamformer.

7. A hearing system according to claim 5 wherein said at least one beamformer comprises first and second beamformers, wherein said first beamformer comprises said target signal component, and wherein said second beamformer is a target-cancelling beamformer comprising said noise component.

8. A hearing system according to claim 7 wherein said statistical structure is applied to said noise component in that the statistical structure is added directly to the noise component (i.e. the noise component itself is modified); and/or in that the statistical structure is added to the noise component in combination with other processing done on the noise component; and/or in that the statistical structure is added to the noise component and added to the output signal, after the original noise component provided by the second beamformer has been removed from the target signal component signal provided by the first beamformer.

9. A hearing system according to claim 1 comprising at least one analysis filter bank for providing said at least one electric input signal in a time frequency representation (k,l), where (k,l) represents a time-frequency tile, and k is a frequency index and l is a time index.

10. A hearing system according to claim 9 wherein auditory texture is added to time-frequency regions that are attenuated in the noise control system of the hearing device.

11. A hearing system according to claim 1 comprising an auxiliary device wherein a part of the processing of the hearing system is performed.

12. A hearing system according to claim 1 being constituted by the hearing device.

13. A hearing system according to claim 1 wherein the hearing device is constituted by or comprises a hearing aid or first and second hearing aids of a binaural hearing aid system, or a headset, or a combination thereof.

14. A hearing system according to claim 1 comprising a further hearing device, wherein the hearing device and the further hearing device each comprise appropriate antenna and transceiver circuitry allowing them to exchange data, either directly or via an auxiliary device.

15. A hearing system according to claim 14 when dependent on claim 9 configured to provide that the phase of the complex time frequency tile of the at least one analysis filter bank of a given hearing device may be altered by multiplying the at least one electric input signal with a random or a pseudorandom phase.

16. A method of operating a hearing system comprising a hearing device configured to be worn by a user, the method comprising providing at least one electric input signal representative of sound in the environment of the hearing device, wherein said at least one electric input signal comprises a) a target signal component assumed to be of current interest to the user, and b) a noise component; providing an output signal based on said at least one electric input signal, either comprising stimuli for being presented to the user, and/or for being transmitted to another device; providing an estimate of said target signal component and an estimate of said noise component in said at least one electric input signal, or in a signal originating therefrom, applying a statistical structure to said estimate of said noise component to thereby provide a modified noise component comprising said statistical structure; and determining a modified estimate of said target signal component in dependence of said modified noise component, and providing that said output signal comprises said modified estimate of said target signal component, or a further processed version thereof.

17. A computer program comprising instructions which, when the program is executed by a computer, cause the computer to carry out the method of claim 16.

18. A hearing system comprising a hearing device configured to be worn by a user, the hearing device comprising at least two input transducers configured to provide respective at least two electric input signals representative of sound, wherein at least one of said at least two electric input signals comprises a) a target signal component assumed to be of current interest to the user, and wherein at least one of said at least two electric input signals comprises b) a noise component; an output unit configured to provide an output signal independence of said at least two electric input signals, either comprising stimuli for being presented to the user, and/or for being transmitted to another device; a noise control system configured to provide an estimate of said target signal component and an estimate of said noise component in said at least two electric input signals, or in a signal or signals originating therefrom; wherein the noise control system is configured to apply a statistical structure to said estimate of said noise component to thereby provide a modified noise component comprising said statistical structure; and to determine a modified estimate of said target signal component in dependence of said modified noise component, and wherein said output signal comprises said modified estimate of said target signal component, or a further processed version thereof.

19. A hearing system according to claim 18 wherein the at least two input transducers comprise at least one acoustic to electric transducer each being configured to provide an electric input signal comprising sound from the environment of the user wearing the hearing system and an audio receiver configured to provide an electric input signal representing sound received from another device or system.

20. A hearing system according to claim 19 wherein the target signal component originates from the electric input signal provided by the audio receiver, and wherein at least a part of the noise component originates from the electric input signal provided by the at least one acoustic to electric transducer.

Description

BRIEF DESCRIPTION OF DRAWINGS

[0119] The aspects of the disclosure may be best understood from the following detailed description taken in conjunction with the accompanying figures. The figures are schematic and simplified for clarity, and they just show details to improve the understanding of the claims, while other details are left out. Throughout, the same reference numerals are used for identical or corresponding parts. The individual features of each aspect may each be combined with any or all features of the other aspects. These and other aspects, features and/or technical effect will be apparent from and elucidated with reference to the illustrations described hereinafter in which:

[0120] FIG. 1A, 1B, 1C shows simplified block diagrams of first, second and third embodiments of a hearing device according to the present disclosure,

[0121] FIG. 2A, 2B shows two options for where a statistical structure, e.g. a rhythmic pattern of modulation over time, may be added to the processing pipeline of a single hearing aid with two microphones,

[0122] FIG. 3 shows an embodiment of a binaural hearing aid system according to the present disclosure, and

[0123] FIG. 4A shows the estimated coherence function as well as the true coherence between two microphones as a function of frequency for a cylindrical isotropic noise field; and

[0124] FIG. 4B shows the estimated coherence function as well as the true coherence between two microphones as a function of frequency for a spherically isotropic noise field.

[0125] The figures are schematic and simplified for clarity, and they just show details which are essential to the understanding of the disclosure, while other details are left out. Throughout, the same reference signs are used for identical or corresponding parts.

[0126] Further scope of applicability of the present disclosure will become apparent from the detailed description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the disclosure, are given by way of illustration only. Other embodiments may become apparent to those skilled in the art from the following detailed description.

DETAILED DESCRIPTION OF EMBODIMENTS

[0127] The detailed description set forth below in connection with the appended drawings is intended as a description of various configurations. The detailed description includes specific details for the purpose of providing a thorough understanding of various concepts. However, it will be apparent to those skilled in the art that these concepts may be practiced without these specific details. Several aspects of the apparatus and methods are described by various blocks, functional units, modules, components, circuits, steps, processes, algorithms, etc. (collectively referred to as “elements”). Depending upon particular application, design constraints or other reasons, these elements may be implemented using electronic hardware, computer program, or any combination thereof.

[0128] The electronic hardware may include micro-electronic-mechanical systems (MEMS), integrated circuits (e.g. application specific), microprocessors, microcontrollers, digital signal processors (DSPs), field programmable gate arrays (FPGAs), programmable logic devices (PLDs), gated logic, discrete hardware circuits, printed circuit boards (PCB) (e.g. flexible PCBs), and other suitable hardware configured to perform the various functionality described throughout this disclosure, e.g. sensors, e.g. for sensing and/or registering physical properties of the environment, the device, the user, etc. Computer program shall be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.

[0129] The present application relates to the field of hearing devices, e.g. hearing aids or headsets or combinations thereof.

[0130] In the present disclosure, it is proposed to impose a perceptible statistical structure on the background sound signal, e.g. to modify the background sound signal such that its statistical regularity over time is increased, in order to improve auditory object formation in hearing aid users. In some situations, this may have a negative effect on the overall SNR (by increasing the energy in the background sound). It may, however, have the overall positive effect of making multiple incohesive background sounds ‘group together’ more into an auditory texture, hence making the auditory scene simpler, and the task of attending to a foreground sound easier. This enhanced statistical structure may be provided in a number of ways, cf. e.g. two examples outlined below.

[0131] The present disclosure provides an improvement on previous solutions to dealing with background noise in hearing devices because, unlike further attenuation of the background noise, the proposed solution can improve the listener's perception of the target sounds with only minimal further limitations on the audibility of the surrounding auditory scene.

[0132] The technical means of the invention may include: [0133] A monoaural or binaural hearing device (e.g. a hearing aid or headset) or system comprising a noise reduction system, e.g. implemented as a multi microphone input beamformer, e.g. an MVDR beamformer and a (single channel) post filter, with two or more microphones on each device in a monaural solution and at least one microphone in each device in a binaural solution. [0134] An interface for activating the technology, either by the user's real-time choice, a prescribed program of the hearing device (or system), or by an automatic estimation of the complexity of the listening environment (e.g. provided by the hearing device (or system) itself, or optionally using external sensors or devices). [0135] One or more rhythmic patterns for object formation (e.g. optimized for object formation). [0136] A processor and algorithm to apply amplitude-modulation of the background noise.

[0137] FIGS. 1A, 1B and 1C shows simplified block diagrams of first, second and third embodiments of a hearing device according to the present disclosure.

[0138] FIG. 1A shows a hearing system (according to the first aspect of the present disclosure) comprising a hearing device (e.g. a hearing aid) configured to be worn by a user, e.g. at or in an ear (or fully or partially implanted in the head of the user). The hearing device comprises an input transducer (IT1), e.g. a microphone, for providing at least one electric input signal (X1) representative of sound in the environment of the hearing device. The electric input signal (X1) comprises a) a target signal component assumed to be of current interest to the user, and b) a noise component. The hearing device further comprises an output unit (OU) configured to provide an output signal based on the at least one electric input signal (X1), either comprising stimuli for being presented to the user, and/or (a signal) for being transmitted to another device. The output unit (OU) may comprise an output transducer, e.g. a loudspeaker or a vibrator. The output unit (OU) may comprise an electrode array or a wireless transceiver. The hearing device further comprises a noise control system (NCS) configured to provide a noise reduced signal (Y.sub.NR). The noise control system comprises a target-and-noise estimator (TNE) for providing an estimate of the target signal component (TE) and an estimate of the noise component (NE) in the at least one electric input signal (X1), or in a signal originating therefrom. The noise control system comprises a noise modifier (MOD) configured provide and apply a statistical structure to the noise component (NE) to thereby provide a modified noise component (MNE) comprising the statistical structure. The noise control system (NCS) id further configured to determine a modified estimate (Y.sub.NR) of the target signal component (TE) in dependence of the modified noise component (MNE). The hearing system is configured to provide that the output signal (e.g. presented to the user) comprises the modified estimate (Y.sub.NR) of the target signal component, or a further processed version thereof.

[0139] The noise modifier may modify the noise using either linear or non-linear processing. Also, the noise may be modified by adding another signal (with specific statistical properties) to the noise.

[0140] The embodiment of FIG. 1B exemplifies a hearing system according to the second aspect of the present disclosure. The hearing system consists of or comprises a hearing device (e.g. a hearing aid) configured to be worn by a user. The hearing device of FIG. 1B is similar to the embodiment of FIG. 1A, but comprises (at least) two input transducers instead of (at least) one (and the target and (at least a part of the) noise components may be provided by separate input transducers). The hearing device (HD, e.g. a hearing aid) comprises at least two input transducers (IT1, IT2) configured to provide respective at least two electric input signals (X1, X2) representative of sound. A first one (e.g. X1) of the at least two electric input signals comprises a target signal component assumed to be of current interest to the user. A second one (e.g. X2) of said at least two electric input signals comprises a noise component. The first electric input signal (X1) (comprising the target signal component) may originate from an audio receiver (the first input transducer IT1). The second electric input signal (X2) (comprising at least a part of the noise component) may originate from an acoustic to electric transducer (the second input transducer IT2). The target-and-noise estimator (TNE) provides an estimate of the target signal component (TE) in the first one of at least one electric input signal (xl) (or in a signal originating therefrom). The target-and-noise estimator (TNE) further provides an estimate of a noise component (NE) in the second one of at least one electric input signal (X1) (or in a signal originating therefrom). Hence the target component and the noise component (or at least a part thereof) are determined from two different electric input signals (end thus from two different input transducers).

[0141] Instead of an acoustic to electric transducer and an audio receiver, the at least two input transducers may comprise at least two acoustic to electric transducers, e.g. microphones and/or vibration sensors. The target signal component and (at least a part of) the noise component may be determined in dependence of the at least two electric input signals provided by the at least two acoustic to electric transducers. The target signal component and (at least a part of) the noise component may e.g. be determined from different acoustic to electric transducers, e.g. from a microphone relatively close to the target sound source and a microphone relatively far away from the target sound source, respectively.

[0142] The embodiment of FIG. 1C is similar to the embodiments of FIG. 1A or 1B. The following differences are: A. The hearing device (HD) comprises two input transducers (IT1, IT2), e.g. two microphones, providing respective first and second electric input signals (X1, X2) comprising sound from the environment of the user wearing the hearing device. B. The noise control system (NCS), e.g. the target-and-noise estimator (TNE), comprises a directional system (DIRS) comprising at least one beamformer (e.g. two beamformers) and configured to provide the estimate of the target signal component (TE) and the estimate of the noise component (NE), respectively, based on the first and second electric input signals (X1, X2).

[0143] C. The hearing device further comprises a signal processing unit (SPU) in the forward (audio) path, e.g. configured to apply one or more processing algorithms (e.g. to compensate for the user's hearing impairment) to the (noise reduced) signal (Y.sub.NR) from the noise control system (NCS). The signal processing unit (SPU) provides a processed signal (Your) in dependence of the (noise reduced) signal (Y.sub.NR). The output unit (OU) is configured to provide the stimuli of the output signal based on the processed signal (Y.sub.OUT).

[0144] More specific embodiments are provided in FIGS. 2A, and 2B, and described further below.

[0145] Given current theories on auditory object formation, different patterns of background noise modulation are possible. In the present disclosure it is proposed to modify the background noise in order to render it more perceptually coherent, e.g. to specifically add modulation or specific signal characteristics into a noise pathway of a noise reduction system of the hearing device. Four suggestions include:

[0146] Method 1: Add Sound Textures to the Background Sound Signal.

[0147] The statistical structure which could be added to the background could take the form of an ‘auditory texture’ for example. Natural ‘auditory textures’ are sounds produced by the addition of many similar sound sources, such as the sound of a room full of people talking, or a swarm of bees, or of raindrops (cf. e.g. [McWalter & McDermott; 2019]). These sound textures tend to be stable over fairly long periods of time and, although they are complex in the sense that they are formed from many individual sources, they tend to be perceived as a single ‘background’ sound. Auditory textures can be characterized by low-order summary statistics (such as the between-channel correlations when the signal is passed through a filter bank of an auditory processing model), and can be synthesized by imposing these simple statistics on noise (cf. e.g. [McDermott & Simonelli; 2011]). Rather than modulating a pure noise source, the perception of auditory texture may be imposed on the noise signal in a hearing device, with the aim of rendering the noise signal more perceptually coherent. Modulations of some of these features (for example between-channel correlation structure) may be applied to the frequency-domain noise signal (cf. e.g. on the output of a target cancelling beamformer) in a hearing aid that employs MVDR noise reduction, resulting in a ‘textured’ background sound signal that may have the property of being perceived as a single background object rather than a complex background scene.

[0148] Furthermore, in a binaurally-connected hearing system, the IAC of the subsequent ‘texturized background signal’ could be manipulated so that it has an artificially lower IAC (by inter-aurally decorrelating some of the texture modulations for example). Natural ‘background sounds’ tend to have low inter-aural correlation because very different signals reach the two ears (they are reflected many times from different directions by the physical properties of the listening space). However, after a binaural beamformer is applied, the resulting ‘noise’ tends to be highly localized (which implies a high or constant level of IAC). If we apply the textures above separately to the two ears, we may ‘break’ this highly-coherent noise source and render it more diffuse, and potentially then easier for the brain to separate from target signals. The noise is not only highly localized—it tends to be localized from the same direction as the target, because the binaural beamformer yields more or less the same signal to be presented to both ears. This may e.g. be implemented by “randomizing” the phase of the time-frequency components which are dominated by the noise (cf. ‘Method 4’ below).

[0149] In addition, in order to make these imposed sound textures merge and be identifiable as a background sound rather than an additional foreground sound, the particular selection of modulations applied could be selected to correspond closely with the real background noise signal We can analyse the acoustic properties of the current background sound (spectral centroid, between-channel correlation etc.) and then apply a texture from a library of possible textures which has acoustic features that (fairly closely; e.g. best among a number of examples, e.g. from a library) match the natural environment. The analysis may e.g. be performed using a sound scene classifier. For example, in a noise-speech environment, apply the sound texture for ‘babble background,’ but if the noise is from outdoors, apply the ‘rain’ texture, etc. This option may also, at least in part, apply to ‘Method 2’ outlined below. The properties of the background sound can be compared to a library of sound textures to find one that has the best match with the acoustic properties of the natural environment.

[0150] Method 2: Add Regular Amplitude Modulation to the Background Sound.

[0151] The statistical structure applied to the background sound may also be a simple amplitude modulation in a rhythmic pattern. In other words, the statistical regularity may come from repeating a pattern of amplitude-modulation over time.

[0152] Because the brain is substantially attuned to detect patterns in sound, only minimal modification of the sound output may be necessary to achieve a listening improvement. Effective patterns for eliciting object formation can be inspired by previous research (e.g. [Aman et al.; 2021]), and further optimised for hearing aid users and their listening environments. The pattern can be accomplished across many different manipulations of the sound (e.g. how long the pattern takes before repeating, the min/max duration between the up/down and down/up segments of the amplitude modulation, the min/max levels in amplitude of the amplitude modulation). Due to their hearing loss, and how that loss interacts with different listening challenges (e.g. a train station, a canteen, inside a bus), the hearing aid users may be better/worse at perceiving patterns with different characteristics within that range of manipulations. A first example: Since hearing-aid users tend to be older and may have challenges to their working-memory, they need a pattern that takes only a small amount of time before repeating. A second example A hearing-impaired listener with greater hearing loss needs a larger change in the levels of the amplitude-modulation in order to perceive that it is changing in a pattern, and this means that the down/up to up/down period of the amplitude modulation has to be shorter to minimize the energetic masking that it adds. These patterns can be applied to the background sound estimate in a hearing aid that uses directional beamforming, see e.g. EP2701145A1.

[0153] Method 3: Addition of Textured Noise to Monaural Noise Reduction Stage.

[0154] In addition to the modifications to the ‘noise signal’ in the beamformer as described above, it may also be possible to add textured noise to the time-frequency regions that are attenuated in the monaural noise reduction stage (e.g. in the noise control system) of the hearing device.

[0155] Normally, the noise reduction system finds time-frequency regions in the signal where the noise is more energetic than the target signal, and then attenuates those regions by a small amount (for example 7 dB) in order to avoid introducing ‘musical tones’ that occur with more aggressive noise attenuation. The amount of attenuation may depend on e.g. SNR, type of background noise, or frequency resolution. It is proposed to attenuate noisy regions more aggressively, e.g. attenuating by 20 dB, but e.g. also to add ‘textured’ background noise to the specific time-frequency units, where we attenuated. In that way a more pleasant background noise, without audible artefacts, may be obtained for presentation to the listener.

[0156] In the case where noise is only added to low-SNR or level regions in time and in frequency, it may only be beneficial to add noise, if the amount of noisy regions is high (a minimum number of noisy regions are needed in order to allow the added noise to be grouped together).

[0157] FIG. 2A, 2B shows two options for where a statistical structure, e.g. a rhythmic pattern of modulation over time, may be added to the processing pipeline of a single hearing aid with two microphones.

[0158] FIG. 2A, 2B each shows a part of a hearing aid comprising first and second microphones (M.sub.1, M.sub.2) providing respective first and second electric input signals IN.sub.1 and IN.sub.2, and a noise reduction system. The noise reduction system comprises a directional system (DIRS) providing a noise reduced (e.g. at least beamformed) signal Y.sub.BF based on the first and second electric input signals. A direction from the target signal to the hearing aid is e.g. defined by the microphone axis and indicated in FIG. 2A, 2B by arrow denoted Target sound. The target direction can be any direction in the environment. The target direction may e.g. be a direction to a speaker of interest in the user's environment. An adaptive beam pattern (Y (Y(k))), for a given frequency band k, k being a frequency band index, is obtained by linearly combining a delay-and-sum-beamformer (O(O(k))) and a delay-and-subtract-beamformer (C(C(k))) in that frequency band. The delay-and-sum-beamformer may e.g. have a substantially omni-directional characteristic, as indicated by the circular symbol denoted O in FIG. 2A, 2B. The delay-and-subtract-beamformer may e.g. have a characteristic of cancelling the target signal component (and being termed a ‘target cancelling beamformer’) as indicated by the cardioid symbol denoted C in combination with the target direction (cf. arrow denoted ‘Target sound’) in FIG. 2A, 2B. The first (omni-directional) and second (target-cancelling) beamformers (denoted O and C in FIG. 2A, 2B) provide beamformed signals O and C, respectively, as linear combinations of the first and second electric input signals IN.sub.1 and IN.sub.2, where first and second sets of (frequency dependent) complex weighting constants (W.sub.o1(k), W.sub.o2 (k)) and (W.sub.c1(k), W.sub.c2(k)) representative of the respective beam patterns are stored in memory unit (MEM). The complex weighting constants are applied to the first and second electric input signals via respective multiplication units (‘×’) and the weighted input signals are added (or subtracted) by respective sum-units (‘+’) as shown in FIG. 2A, 2B. The adaptive beam pattern arises by scaling the delay-and-subtract-beamformer (C(k)) by a complex-valued, frequency-dependent, adaptive scaling factor β(k) (generated by beamformer BF) before subtracting it from the delay-and-sum-beamformer (O(k)), i.e. providing the beam pattern Y.


Y(k)=O(k)−β(k)C(k).

[0159] It should be noted that the sign in front of β(k) might as well be +, if the sign(s) of the weights constituting the delay-and-subtract beamformer C are appropriately adapted. Further, β(k) may be substituted by β*(k), where * denotes complex conjugate, such that the beamformed signal Y.sub.BF is expressed as Y.sub.BF=(w.sub.o(k)−β*(k).Math.w.sub.c(k)).sup.H.Math.IN(k).

[0160] An adaptive beamformer may also be obtained by linear combination of other beamformers.

[0161] Preferably, one of the beamformers represents a noise estimate (target cancelling beamformer).

[0162] The directional system (DIRS) may e.g. be adapted to work optimally in situations where the microphone signals comprise a localized target sound source (e.g. a target speaker) in the presence of additive noise sources. Given this situation, the scaling factor β(k) (β in FIG. 2A, 2B) is adapted to minimize the noise under the constraint that the sound impinging from the target direction (at least at one frequency) is essentially unchanged. For each frequency band k, the adaptation factor β(k) can be found in different ways. The solution may be found in closed form as

[00001] β ( k ) = .Math. C * O .Math. .Math. .Math. "\[LeftBracketingBar]" C .Math. "\[RightBracketingBar]" 2 .Math. ,

[0163] where * denote the complex conjugation and custom-charactercustom-character denotes the statistical expectation operator, which may be approximated in an implementation as a time average, e.g. comprising a low-pass filter. The expectation operator custom-charactercustom-character may e.g. be implemented using a first order IIR filter, possibly with different attack and release time constants. Alternatively, the expectation operator may be implemented using an FIR filter.

[0164] The adaptive beamformer (BF) may be configured to determine the adaptation parameter β.sub.opt(k) from the following expression

[00002] β opt = w O H C v w C w C H C v w C ,

[0165] where w.sub.o and w.sub.c are the beamformer weights for the delay and sum O and the delay and subtract C beamformers, respectively, C.sub.v is the noise covariance matrix, and H denotes Hermitian transposition.

[0166] Each of the embodiments of FIG. 2A, 2B comprises a different solution of applying a statistical structure to the noise component of the electric input signals (IN.sub.1, IN.sub.2) to thereby provide a modified noise signal component comprising the statistical structure. The application of a statistical structure may e.g. comprise one or more of a) applying modulation to the noise estimate, b) randomizing the phase of the noise estimate, c) applying auditory texture to the noise estimate.

[0167] In the embodiment of FIG. 2A the noise modifier (MOD) is located after the adaptive beamformer (ABF) providing the adaptive (noise attenuating) parameter β (or matrix β in case of more than two electric input signals), thereby providing a modified parameter β.sub.mod (or matrix β.sub.mod) comprising the statistical structure. The applied statistical structure is provided by modification control signal (STST) or it may be a fixed feature of the noise modifier (MOD). In the embodiment of FIG. 2A, the modified adaptive parameter β.sub.mod is multiplied to the noise component (C) provided by the target cancelling beamformer. The resulting beamformed signal Y.sub.BF is based on the signal from the omnidirectional beamformer (comprising the target signal component and noise) and the noise component from the target cancelling beamformer multiplied by the modified adaptive (noise cancelling) parameter β.sub.mod to provide the noise reduced signal Y.sub.BF=O-β.sub.modC.

[0168] The embodiment of FIG. 2B resembles the embodiment of FIG. 2A, where the noise modifier (MOD) is located after the target-cancelling beamformer, but in FIG. 2B the modified noise estimate (MNE, denoted β.sub.mod in FIG. 2A) is combined with the beamformed signal, e.g. in a combination unit, e.g. a sum unit or a multiplication unit, or more generally a filter. A (single channel) post filter (PF) may be inserted before or after the combination unit. The combination unit may form part of the post filter (PF) as shown in FIG. 2B.

[0169] Method 4: Binaural Beamforming and Addition of Textured Noise to Monaural Noise Reduction Stage (e.g. in the Noise Control System).

[0170] In general, the amount of noise added may depend on an overall sound level, or on the estimated signal to noise ratio (SNR) in the mixture, e.g. in situations with only little noise, it may not be necessary to add background noise compared to more difficult scenarios.

[0171] For a system with more than two microphones, noise estimates from more than one direction may be obtained with a generalized sidelobe canceller. If implemented in a binaural hearing aid system, the background noise estimates may be further modulated differently at each ear, to introduce a frequency-dependent interaural timing difference (or the phase of the background noises may be randomized in order to make the noise more diffuse). This may e.g. be accomplished in the sound processing pipeline after the sounds from the microphones have been filtered into separate frequency channels. Then, the signal carried by the frequency channel from the inputs of one hearing aid would be slightly delayed compared to the signal in the corresponding frequency channel of the other hearing aid, to simulate that this signal arrives at each ear with a delay.

[0172] The timing difference may e.g. be adjusted to simulate: [0173] The laterality of the background noise. The “laterality” in this case is referencing from the hearing aid user's head. So sound signals can be simulated to have originated from either more to the left or more to the right of the listener's head's midline. This new directionality may help with the object formation of the background. [0174] The directional diffusivity (the “size” of the sound source in space). Diffuseness may be quantified based on an estimated noise covariance matrix. If the noise between the microphones is highly correlated, it indicates that the noise is impinging from a single direction (i.e. the opposite of a diffuse noise source). In theory, more diffuse sounds are better classified as background by the brain, whereas less diffuse sounds are better perceived as objects.

[0175] As the binaural beamformer for both ears rely (at least partly) on the same input signals, the target and the remaining noise in the noise reduced signal will appear as if the target and the noise is co-located. Hereby the listener's ability to segregate the target from the remaining noise is degraded.

[0176] FIG. 3 shows an embodiment of a binaural hearing aid system according to the present disclosure. The binaural hearing aid system comprises left and right hearing aids (HD.sub.L, HD.sub.R) as indicated by the brackets denoted ‘HD.sub.L, HD.sub.R’ in the left part of FIG. 3. Each hearing aid comprises at least one microphone (her one in each hearing aid is indicated (M.sub.L, M.sub.R)). Each of the left and right hearing aids (HD.sub.L, HD.sub.R) comprises appropriate antennas and transceiver circuitry to establish a communication link between the two hearing aids (possibly via a third intermediate device, e.g. a processing device, e.g. a smartphone). The communication link may be configured to transmit and receive audio data, as indicated by dashed arrows from the left to the right and from the right to the left hearing device. Each hearing device may comprise more than one microphone. More than one microphone signal (or a part thereof, e.g. filtered or down-sampled versions thereof) may be exchanged between the left and right hearing aids (HD.sub.L, HD.sub.R). Each of the left and right hearing aids (HD.sub.L, HD.sub.R) comprises and noise control system comprising a binaural beamformer (Binaural Beamformer (L), Binaural Beamformer (R)). Each of the binaural beamformers gets as inputs a locally originating microphone signal and a microphone signal (or a filtered or down-sampled version thereof) received from the opposite hearing aid (via the communication link). Each of the binaural beamformers provides a binaurally beamformed signal which is fed to fed to a post processing unit (denoted Post-Processing (L) and Post-Processing (R) in the left and right hearing aids, respectively). The binaural beamformer or the post-processing unit of each of the left and right hearing aids may comprise a post filter for further reducing noise in the beamformed (spatially filtered) signal. The post-processing units of each of the left and right hearing aids are configured to apply one or more processing algorithms to the signal from the noise reduction system (e.g. from the binaural beamformer) and to provide a processed signal to an output transducer. The post-processing units may be configured to apply a frequency and/or level dependent gain to a signal for the forward (audio) path of the respective hearing aid, e.g. to compensate for a hearing impairment of the user. In the embodiment of FIG. 3, the output transducer is a loudspeaker (denoted SPK.sub.L, SPK.sub.R in the left and right hearing aids, respectively) configured to play processed sound to the respective left and right ears of the user (U). The left and right hearing aids are thus air conduction hearing aids. They may, however, be or comprise bone conduction hearing aids or cochlear implant type hearing aids (or a combination thereof).

[0177] In order to ease the listener's ability to segregate speech from the noise (according to the present disclosure), further processing may be applied (here shown as a post-processing block in FIG. 3). The post-processing may contain a single channel noise reduction system, which is able to identify regions (in time and frequency) where either the target signal or the background noise is dominant. The post-processing block may have more inputs compared to what is shown in the FIG. 3. Such additional input may be: a noise estimate e.g. from a target cancelling beamformer; also a voice activity detector may be used to identify whether target or background noise dominate the time-frequency unit. In the time-frequency units, where noise is dominant, further noise reduction may be applied, e.g. in terms of a gain reduction.

[0178] Furthermore, the phase of the (complex) time frequency tile may be altered (by multiplying the signal with a “random” or a “pseudorandom” phase change (by a multiplication to the signal by exp(jφ))). The phase φ may e.g. be a altered such that the noise will appear from a different direction than the target (this may also be obtained by applying a different HRTF for left and right ear (here also the amplitude may be altered)). The phase may also be randomized such that the noise field becomes diffuse (e.g. a spherically diffuse noise field or a cylindrically diffuse noise field, see examples below). This can be obtained by for each instrument (HD.sub.L, HD.sub.R) drawing the angle φ from a different random distribution (where the maximum and the minimum φ corresponds to the maximum possible delay depending on the microphone distance As the maximum delay between the two microphones is given by exp(−j2π.Math.d/c), where d is the microphone distance and c is the sound velocity, and f is the frequency; φ should be in the interval given by [−π.Math.f.Math.d/c; +π.Math.f.Math.d/c]).

Examples of Conversion of Co-Located Background Noise to Binaural Diffuse Noise

[0179] Regarding the conversion of co-located background noise into binaural diffuse noise, an example of converting the noise into a cylindrical or spherical diffuse noise field by (for each frequency unit) multiplying the noisy time frequency unit by a random phase is provided in the following.

[0180] As the horizontal angles in a cylindrical noise field are equally likely, we can for each frequency band draw the angles from a uniform distribution. However, having a uniform distribution of angles does not mean that the phase is uniformly distributed. In a free-field cylindrically diffuse noise field, the phase difference between two microphones is given by

[00003] φ = 2 π f d cos ( Θ ) c

[0181] where f is the frequency, d is the microphone distance, c is the sound velocity, and Θ is a uniformly random distribution in the interval [0, 2π] (or [0, π], as the noise field is symmetric).

[0182] By for each time and frequency unit (t,f) dominated by noise multiplying exp(−iφ(t,f)), where i=√{square root over (−1)}, we can convert the background noise into a diffuse noise signal with a desired coherence function.

[0183] FIG. 4A shows the estimated coherence function as well as the true coherence between two microphones as function of frequency for a cylindrical isotropic noise field. In a cylindrically isotropic noise field the coherence between two microphones is given by B.sub.0(2πfd/c), where B.sub.0 is a Bessel function of 0.sup.th order. In the plot of FIG. 4A, d=0.17 m and c=340 m/sec.

[0184] In a similar way, we may convert the background noise into other diffuse noise fields, such as e.g. a spherically isotropic noise field.

[0185] The formula for generating the random phase in a spherically isotropic noise field may be expressed as:


φ(t,f)=cos(Θ)sin(cos.sup.−1(U))

here O is a uniformly random distribution in the interval [0; 2π] (or [0; π]) and U is a uniformly random distribution in the interval [−1,1]. In this case the noise should be multiplied by exp(−iφ(t,f)), where i=√{square root over (−1)}.

[0186] FIG. 4B shows the estimated coherence function as well as the true coherence between two microphones as function of frequency for a spherically isotropic noise field (given by a sin(2πfd/c)/(2πfd/c)). In the plot of FIG. 4B, d=0.17 m and c=340 m/sec.

[0187] It may be advantageous to take into account that the phase between two consecutive frames may be correlated due to frame overlap in the filter bank.

[0188] An advantage of the proposed method is that the random phase may be applied without exchanging information about the phase between the two hearing instruments of a binaural system.

[0189] The phase randomization may be applied solely on one side, or the phase randomization may be applied to both hearing instruments, where each random distribution will be drawn such that the correlation between the microphones follows the distribution for e.g. spherically isotropic noise.

[0190] Cylindrically and spherically isotropic noise fields are two very specific idealized noise fields. Other noise fields may be considered.

[0191] The phase randomization may be applied solely above a threshold frequency, e.g. 1500 Hz. These types of phase modification may be applied as a tinnitus masker.

[0192] Embodiments of the disclosure may, e.g., be useful in applications such as hearing aids or headsets.

[0193] It is intended that the structural features of the devices described above, either in the detailed description and/or in the claims, may be combined with steps of the method, when appropriately substituted by a corresponding process.

[0194] As used, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well (i.e. to have the meaning “at least one”), unless expressly stated otherwise. It will be further understood that the terms “includes,” “comprises,” “including,” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will also be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element, but an intervening element may also be present, unless expressly stated otherwise. Furthermore, “connected” or “coupled” as used herein may include wirelessly connected or coupled. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. The steps of any disclosed method are not limited to the exact order stated herein, unless expressly stated otherwise.

[0195] It should be appreciated that reference throughout this specification to “one embodiment” or “an embodiment” or “an aspect” or features included as “may” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the disclosure. Furthermore, the particular features, structures or characteristics may be combined as suitable in one or more embodiments of the disclosure.

[0196] The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects.

[0197] The claims are not intended to be limited to the aspects shown herein but are to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more.

REFERENCES

[0198] [Aman et al.; 2021] Aman L, Picken S, Andreou L V, Chait M., Sensitivity to temporal structure facilitates perceptual analysis of complex auditory scenes. Hearing Research. 2021, Feb. 1; 400:108111. [0199] [Mishra et al.; 2021] Mishra A P, Harper N S, Schnupp J W H (2021), Exploring the distribution of statistical feature parameters for natural sound textures. PLoS ONE 16(6): e0238960. https://doi.org/10.1371/journal.pone. 0238960. [0200] [McWalter & McDermott; 2019] McWalter, R., McDermott, J. H., Illusory sound texture reveals multi-second statistical completion in auditory scene analysis. Nat. Commun 10, 5096 (2019). https://doi.org/10.1038/s41467-019-12893-0. [0201] [McWalter and McDermott; 2018] McWalter, R. I. & McDermott, J. H., Adaptive and selective time-averaging of auditory scenes. Curr. Biol. 28, 1405-1418 (2018). [0202] [McDermott & Simoncelli; 2011] McDermott J H, Simoncelli E P, Sound texture perception via statistics of the auditory periphery: evidence from sound synthesis. Neuron. 2011; 71(5): 926-940. https://doi.org: 10.1016/j.neuron. 2011.06.032. [0203] EP2701145A1 (Oticon) 26.02.2014. [0204] JP2010200260A (Yamaha) 09.09.2010.