HEARING AID COMPRISING A BEAM FORMER FILTERING UNIT COMPRISING A SMOOTHING UNIT
20170347206 · 2017-11-30
Assignee
Inventors
Cpc classification
H04R2430/20
ELECTRICITY
H04R25/606
ELECTRICITY
H04R25/407
ELECTRICITY
H04R2225/67
ELECTRICITY
H04R2225/0216
ELECTRICITY
H04R2430/25
ELECTRICITY
International classification
Abstract
A hearing aid comprises a resulting beam former (Y) for providing a resulting beamformed signal Y.sub.BF based on first and second electric input signals IN.sub.1 and IN.sub.2, first and second sets of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), and a resulting complex, frequency dependent adaptation parameter β(k)•β(k) may be determined as <C.sub.2*•C.sub.1>/<(|C2|.sup.2>+c), where * denotes the complex conjugation and •
denotes the statistical expectation operator, and c is a constant, and wherein said adaptive beam former filtering unit (BFU) comprises a smoothing unit for implementing said statistical expectation operator by smoothing the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2>.sup.2 over time. Alternatively, β(k) may be determined from the following expression
where w.sub.C1 and w.sub.C2 are the beamformer weights representing the first (C.sub.1) and the second (C.sub.2) beamformers, respectively, C.sub.v is a noise covariance matrix, and H denotes Hermitian transposition. Corresponding methods of operating a hearing aid, and a hearing aid utilizing smoothing β(k) based on adaptive covariance smoothing are disclosed.
Claims
1. A hearing aid adapted for being located in an operational position at or in or behind an ear or fully or partially implanted in the head of a user, the hearing aid comprising first and second microphones (M.sub.BTE1, M.sub.BTE2) for converting an input sound to first IN.sub.1 and second IN.sub.2 electric input signals, respectively, an adaptive beam former filtering unit (BFU) for providing a resulting beamformed signal Y.sub.BF, based on said first and second electric input signals, the adaptive beam former filtering unit comprising, a first memory comprising a first set of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) representing a first beam pattern (C1), where k is a frequency index, k=1, 2, . . . , K, a second memory comprising a second set of complex frequency dependent weighting parameters W.sub.21(k), W.sub.22(k) representing a second beam pattern (C2), where said first and second sets of weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), respectively, are predetermined and possibly updated during operation of the hearing aid, an adaptive beam former processing unit for providing an adaptively determined adaptation parameter β(k) representing an adaptive beam pattern (ABP) configured to attenuate unwanted noise as much as possible under the constraint that sound from a target direction is essentially unaltered, and a resulting beam former (Y) for providing said resulting beamformed signal Y.sub.BF based on said first and second electric input signals IN.sub.1 and IN.sub.2, said first and second sets of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), and said resulting complex, frequency dependent adaptation parameter β(k), where β(k) may be determined as •
denotes the statistical expectation operator, and c is a constant, wherein said adaptive beam former filtering unit (BFU) comprises a smoothing unit for implementing said statistical expectation operator by smoothing the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2 over time.
2. A hearing aid according to claim 1, wherein the smoothing unit is configured to apply substantially the same smoothing time constants for the smoothing of the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2.
3. A hearing aid according to claim 1, wherein the smoothing unit is configured to smoothe a resulting adaptation parameter β(k).
4. A hearing aid according to claim 4, wherein the smoothing unit is configured to provide that the attack and release time constants involved in the smoothing of the resulting adaptation parameter β(k) is larger than the corresponding attack and release time constants involved in the smoothing of the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2.
5. A hearing aid according to claim 1, wherein the smoothing unit is configured to provide that the attack and release time constants involved in the smoothing of the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2 are adaptively determined.
6. A hearing aid according to claim 1, wherein the smoothing unit is configured to provide that the attack and release time constants involved in the smoothing of the resulting adaptation parameter β(k) are adaptively determined.
7. A hearing aid according to claim 1, wherein the smoothing unit comprises a low pass filter implemented as an IIR filter with a fixed time constant, and an IIR filter with a configurable time constant.
8. A hearing aid according to claim 7 wherein the smoothing unit is configured to determine the configurable time constant by a function unit providing a predefined function of the difference between a first filtered value of the real expression |C.sub.2|.sup.2 when filtered by an IIR filter with a first time constant, and a second filtered value of the real expression |C.sub.2|.sup.2 when filtered by an IIR filter with a second time constant, wherein the first time constant is smaller than the second time constant.
9. A hearing aid according to claim 8 wherein the function unit comprises an ABS unit providing an absolute value of the difference between the first and second filtered values.
10. A hearing aid according to claim 8 wherein the first and second time constants are fixed time constants.
11. A hearing aid according to claim 9 wherein the first time constant the fixed time constant and the second time constant is the configurable time constant.
12. A hearing aid according to claim 8 wherein the predefined function is a decreasing function of the difference between the first and second filtered values.
13. A hearing aid according to claim 12 wherein the predefined function is one of a binary function, a piecewise linear function, and a continuous monotonous function.
14. A hearing aid according to claim 8 wherein the smoothing unit comprises respective low pass filters implemented as IIR filters using said configurable time constant for filtering real and imaginary parts of the expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2, and wherein said configurable time constant is determined from |C.sub.2|.sup.2.
15. A hearing aid adapted for being located in an operational position at or in or behind an ear or fully or partially implanted in the head of a user, the hearing aid comprising first and second microphones (M.sub.BTE1, M.sub.BTE2) for converting an input sound to first IN.sub.1 and second IN.sub.2 electric input signals, respectively, an adaptive beam former filtering unit (BFU) for providing a resulting beamformed signal Y.sub.BF, based on said first and second electric input signals, the adaptive beam former filtering unit comprising, a first memory comprising a first set of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) representing a first beam pattern (C1), where k is a frequency index, k=1, 2, . . . , K, a second memory comprising a second set of complex frequency dependent weighting parameters W.sub.21(k), W.sub.22(k) representing a second beam pattern (C2), where said first and second sets of weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), respectively, are predetermined and possibly updated during operation of the hearing aid, an adaptive beam former processing unit for providing an adaptively determined adaptation parameter β(k) representing an adaptive beam pattern (ABP) configured to attenuate unwanted noise as much as possible under the constraint that sound from a target direction is essentially unaltered, and a resulting beam former (Y) for providing said resulting beamformed signal Y.sub.BF based on said first and second electric input signals IN.sub.1 and IN.sub.2, said first and second sets of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), and said resulting complex, frequency dependent adaptation parameter β(k), wherein the adaptive beamformer processing unit is configured to determine the adaptation parameter β(k) from the following expression
16. A hearing aid according to claim 1 comprising a hearing instrument adapted for being located at or in an ear of a user or for being fully or partially implanted in the head of a user, a headset, an earphone, an ear protection device or a combination thereof.
17. A method of operating a hearing aid adapted for being located in an operational position at or in or behind an ear or fully or partially implanted in the head of a user, the method comprising converting an input sound to, or providing, first IN.sub.1 and second IN.sub.2 electric input signals, adaptively providing a resulting beamformed signal Y.sub.BF, based on said first and second electric input signals; storing in a first memory a first set of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) representing a first beam pattern (C1), where k is a frequency index, k=1, 2, . . . , K; storing in a second memory comprising a second set of complex frequency dependent weighting parameters W.sub.21(k), W.sub.22(k) representing a second beam pattern (C2), wherein said first and second sets of weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), respectively, are predetermined and possibly updated during operation of the hearing aid, providing an adaptively determined adaptation parameter β(k) representing an adaptive beam pattern (ABP) configured to attenuate unwanted noise as much as possible under the constraint that sound from a target direction is essentially unaltered, and providing said resulting beamformed signal Y.sub.BF based on said first and second electric input signals IN.sub.1 and IN.sub.2, said first and second sets of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), and said resulting complex, frequency dependent adaptation parameter β(k), where β(k) may be determined as •
denotes the statistical expectation operator, and c is a constant, and smoothing the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2 over time.
18. A method of operating a hearing aid adapted for being located in an operational position at or in or behind an ear or fully or partially implanted in the head of a user, the method comprising converting an input sound to, or providing, first IN.sub.1 and second IN.sub.2 electric input signals, adaptively providing a resulting beamformed signal Y.sub.BF, based on said first and second electric input signals; storing in a first memory a first set of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) representing a first beam pattern (C1), where k is a frequency index, k=1, 2, . . . , K; storing in a second memory comprising a second set of complex frequency dependent weighting parameters W.sub.21(k), W.sub.22(k) representing a second beam pattern (C2), wherein said first and second sets of weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), respectively, are predetermined and possibly updated during operation of the hearing aid, providing an adaptively determined adaptation parameter β(k) representing an adaptive beam pattern (ABP) configured to attenuate unwanted noise as much as possible under the constraint that sound from a target direction is essentially unaltered, and providing said resulting beamformed signal Y.sub.BF based on said first and second electric input signals IN.sub.1 and IN.sub.2, said first and second sets of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), and said resulting complex, frequency dependent adaptation parameter β(k), wherein said resulting complex, frequency dependent adaptation parameter β(k) is determined from the following expression
19. A method according to claim 17 comprising adaptive smoothing of a covariance matrix for said electric input signals comprising adaptively changing time constants (τ.sub.att, τ.sub.rel) for said smoothing in dependence of changes (ΔC) over time in covariance of said first and second electric input signals; wherein said time constants have first values (τ.sub.att1, τ.sub.rel1) for changes in covariance below a first threshold value (ΔC.sub.th1) and second values (τ.sub.att2, τ.sub.rel2) for changes in covariance above a second threshold value (ΔC.sub.th2), wherein the first values are larger than corresponding second values of said time constants, while said first threshold value (ΔC.sub.th1) is smaller than or equal to said second threshold value (ΔC.sub.th2).
20. A method according to claim 18 comprising adaptively smoothing said noise covariance matrix C.sub.v comprising adaptively changing time constants (τ.sub.att, τ.sub.rel) for said smoothing in dependence of changes (ΔC) over time in covariance of said first and second electric input signals; wherein said time constants have first values (τ.sub.att1, τ.sub.rel1) for changes in covariance below a first threshold value (ΔC.sub.th1) and second values (τ.sub.att2, τ.sub.rel2) for changes in covariance above a second threshold value (ΔC.sub.th2), wherein the first values are larger than corresponding second values of said time constants, while said first threshold value (ΔC.sub.th1) is smaller than or equal to said second threshold value (ΔC.sub.th2).
21. A method according to claim 20 comprising that the noise covariance matrix is C.sub.v is updated when only noise is present.
22. Use of a hearing aid as claimed in claim 1.
23. A data processing system comprising a processor and program code means for causing the processor to perform the steps of the method of claim 17.
Description
BRIEF DESCRIPTION OF DRAWINGS
[0122] The aspects of the disclosure may be best understood from the following detailed description taken in conjunction with the accompanying figures. The figures are schematic and simplified for clarity, and they just show details to improve the understanding of the claims, while other details are left out. Throughout, the same reference numerals are used for identical or corresponding parts. The individual features of each aspect may each be combined with any or all features of the other aspects. These and other aspects, features and/or technical effect will be apparent from and elucidated with reference to the illustrations described hereinafter in which:
[0123]
[0124]
[0125]
[0126]
[0127]
[0128]
[0129]
[0130]
[0131]
[0132]
[0133]
[0134]
[0135]
[0136]
[0137]
[0138]
[0139]
[0140]
[0141]
[0142]
[0143]
[0144]
[0145]
[0146]
[0147]
[0148]
[0149]
[0150]
[0151]
[0152] The figures are schematic and simplified for clarity, and they just show details which are essential to the understanding of the disclosure, while other details are left out. Throughout, the same reference signs are used for identical or corresponding parts.
[0153] Further scope of applicability of the present disclosure will become apparent from the detailed description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the disclosure, are given by way of illustration only. Other embodiments may become apparent to those skilled in the art from the following detailed description.
DETAILED DESCRIPTION OF EMBODIMENTS
[0154] The detailed description set forth below in connection with the appended drawings is intended as a description of various configurations. The detailed description includes specific details for the purpose of providing a thorough understanding of various concepts. However, it will be apparent to those skilled in the art that these concepts may be practised without these specific details. Several aspects of the apparatus and methods are described by various blocks, functional units, modules, components, circuits, steps, processes, algorithms, etc. (collectively referred to as “elements”). Depending upon particular application, design constraints or other reasons, these elements may be implemented using electronic hardware, computer program, or any combination thereof.
[0155] The electronic hardware may include microprocessors, microcontrollers, digital signal processors (DSPs), field programmable gate arrays (FPGAs), programmable logic devices (PLDs), gated logic, discrete hardware circuits, and other suitable hardware configured to perform the various functionality described throughout this disclosure. Computer program shall be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.
[0156] The present application relates to the field of hearing aids, e.g. hearing aids.
C.sub.1(k)=w.sub.11(k)•X.sub.1(k)+w.sub.12(k)•X.sub.2(k)
C.sub.2(k)=w.sub.21(k)•X.sub.1(k)+w.sub.22(k)•X.sub.2(k)
[0157]
[0158]
[0159] An adaptive beampattern (Y(k)), for a given frequency band k, is obtained by linearly combining two beam formers C.sub.1(k) and C.sub.2(k). C.sub.1(k) and C.sub.2(k) are different (possibly fixed) linear combinations of the microphone signals.
[0160] The beampatterns could e.g. be the combination of an omnidirectional delay-and-sum-beam former C.sub.1(k) and a delay-and-subtract-beam former C.sub.2(k) with its null direction pointing towards the target direction (target cancelling beam former) as shown in
Y(k)=C.sub.1(k)−β(k)C.sub.2(k).
[0161] The beam former is adapted to work optimally in situations where the microphone signals consist of a point-noise target sound source in the presence of additive noise sources. Given this situation, the scaling factor β(k) is adapted to minimize the noise under the constraint that the sound impinging from the target direction is unchanged. For each frequency band k, the adaptation factor β(k) can be found in different ways. The solution may be found in closed form as
where * denote the complex conjugation and •
denotes the statistical expectation operator, which may be approximated in an implementation as a time average. As an alternative, the adaptation factor may be updated by an LMS or NLMS equation:
[0162] In the following we omit the frequency channel index k. In (1), the adaptation factor β is estimated by averaging across the input data. A simple way to average across data is by low-pass filtering the data as shown in
[0163]
[0164] Such a low-pass filter LP may e.g. be implemented by a first order IIR filter as shown in
[0165] This is illustrated in
[0166] We propose different ways to overcome this problem. A simple extension is to enable different attack and release coefficients in the low-pass filter. Such a low-pass filter is shown in
[0167]
[0168] C.sub.2*C.sub.1
and
|C.sub.2|.sup.2
estimates by applying smaller time constants. We thus obtain a faster convergence in the case, where the input level suddenly decreases. In
[0169] The advantage of smoothing the estimate of β is that the estimate is less sensitive to sudden drops in input level. Consequently, we can apply a shorter time constant to the low-pass filters used in the numerator and the denominator of (1). Hereby we can adapt faster in case of a sudden decreasing level. By post-smoothing β, we cope with the increased estimation variance.
[0170] Another option is to apply an adaptive smoothing coefficient that changes if a sudden input level change is detected. Embodiments of such low-pass filters are shown in
[0171]
[0172]
[0173] The resulting smoothing estimate from the low-pass filter shown in
[0174]
[0175]
[0176]
[0177]
[0178] The hearing aid (HD) exemplified in
[0179] The hearing aid (HD) comprises a directional microphone system (beam former filtering unit (BFU)) adapted to enhance a target acoustic source among a multitude of acoustic sources in the local environment of the user wearing the hearing aid device. In an embodiment, the directional system is adapted to detect (such as adaptively detect) from which direction a particular part of the microphone signal (e.g. a target part and/or a noise part) originates. In an embodiment, the beam former filtering unit is adapted to receive inputs from a user interface (e.g. a remote control or a smartphone) regarding the present target direction. The memory unit (MEM) may e.g. comprise predefined (or adaptively determined) complex, frequency dependent constants (W.sub.ij) defining predefined or (or adaptively determined) ‘fixed’ beam patterns (e.g. omni-directional, target cancelling, etc.), together defining the beamformed signal Y.sub.BF (cf. e.g.
[0180] The hearing aid of
[0181] The hearing aid (HD) according to the present disclosure may comprise a user interface UI, e.g. as shown in
[0182] The auxiliary device and the hearing aid are adapted to allow communication of data representative of the currently selected direction (if deviating from a predetermined direction (already stored in the hearing aid)) to the hearing aid via a, e.g. wireless, communication link (cf. dashed arrow WL2 in
[0183]
[0184]
[0185]
[0186] The method is configured to operate a hearing aid adapted for being located in an operational position at or in or behind an ear or fully or partially implanted in the head of a user.
[0187] The method comprises
[0188] S1. converting an input sound to first IN.sub.1 and second IN.sub.2 electric input signals,
[0189] S2. adaptively providing a resulting beamformed signal Y.sub.BF, based on said first and second electric input signals;
[0190] S3. storing in a first memory a first set of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) representing a first beam pattern (C1), where k is a frequency index, k=1, 2, . . . , K; storing in a second memory comprising a second set of complex frequency dependent weighting parameters W.sub.21(k), W.sub.22(k) representing a second beam pattern (C2), wherein said first and second sets of weighting parameters W.sub.11(.sub.k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), respectively, are predetermined and possibly updated during operation of the hearing aid,
[0191] S4. providing an adaptively determined adaptation parameter β(k) representing an adaptive beam pattern (ABP) configured to attenuate unwanted noise as much as possible under the constraint that sound from a target direction is essentially unaltered, and
[0192] S5. providing said resulting beamformed signal Y.sub.BF based on said first and second electric input signals IN.sub.1 and IN.sub.2, said first and second sets of complex frequency dependent weighting parameters W.sub.11(k), W.sub.12(k) and W.sub.21(k), W.sub.22(k), and said resulting complex, frequency dependent adaptation parameter β(k), where (β(k) may be determined as
where * denotes the complex conjugation and •
denotes the statistical expectation operator, and c is a constant
[0193] S6. smoothing the complex expression C.sub.2*•C.sub.1 and the real expression |C.sub.2|.sup.2 over time.
[0194] A Method of Adaptive Covariance Matrix Smoothing for Accurate Target Estimation and Tracking.
[0195] In a further aspect of the present disclosure, a method of adaptively smoothing covariance matrices is outlined in the following. A particular use of the scheme is for (adaptively) estimating a direction of arrival of sound from a target sound source to a person (e.g. a user of a hearing aid, e.g. a hearing aid according to the present disclosure).
[0196] The method is exemplified as an alternative scheme for smoothing of the adaptation parameter β(k) according to the present disclosure (cf.
[0197] Signal Model:
[0198] We consider the following signal model of the signal x impinging on the i.sup.th microphone of a microphone array consisting of M microphones:
x.sub.i(n)=s.sub.i(n)+v.sub.i(n), (1)
where s is the target signal, v is the noise signal, and n denotes the time sample index. The corresponding vector notation is
x(n)=s(n)+v(n), (2)
where x(n)=[x.sub.1(n); x.sub.2(n), . . . , x.sub.M(n)].sup.T. In the following, we consider the signal model in the time frequency domain. The corresponding model is thus given by
X(k,m)=S(k,m)+V(k,m), (3)
where k denotes the frequency channel index and m denotes the time frame index. Likewise X(k,m)=[X.sub.1(k,m), X.sub.2(k,m), . . . , X.sub.M(k,m)].sup.T. The signal at the i.sup.th microphone, x.sub.i is a linear mixture of the target signal s.sub.i and the noise v.sub.i. v.sub.i is the sum of all noise contributions from different directions as well as microphone noise. The target signal at the reference microphone s.sub.ref is given by the target signal s convolved by the acoustic transfer function h between the target location and the location of the reference microphone. The target signal at the other microphones is thus given by the target signal at the reference microphone convolved by the relative transfer function d=[1,d.sub.2, . . . , d.sub.M].sup.T between the microphones, i.e. s.sub.i=s*h*d.sub.i. The relative transfer function d depends on the location of the target signal. As this is typically the direction of interest, we term d the look vector. At each frequency channel, we thus define a target power spectral density (k, m) at the reference microphone, i.e.
σ.sub.s.sup.2(k, m)=|S(k, m)H(k, m)|.sup.2
=
|S(k, m).sub.ref|.sup.2
, (4)
where •
denotes the expected value. Likewise, the noise spectral power density at the reference microphone is given by
σ.sub.v.sup.2(k, m)=|V(k, m).sub.ref|.sup.2
, (5)
[0199] The inter-microphone cross-spectral covariance matrix at the k.sup.th frequency channel for the clean signal s is then given by
C.sub.s(k, m)=σ.sub.s(k, m)d(k, m)d.sup.H(k, m), (6)
where H denotes Hermitian transposition. We notice the M×M matrix C.sub.s(k,m) is a rank 1 matrix, as each column of C.sub.s(k,m) is proportional to d(k,m). Similarly, the inter-microphone cross-power spectral density matrix of the noise signal impinging on the microphone array is given by,
C.sub.v(k, m)=σ(k, m)Γ(k, m.sub.0), m>m.sub.0 (7)
where Γ(k, m.sub.0) is the M×M noise covariance matrix of the noise, measured some time in the past (frame index m.sub.0). Since all operations are identical for each frequency channel index, we skip the frequency index k for notational convenience wherever possible in the following. Likewise, we skip the time frame index m, when possible. The inter-microphone cross-power spectral density matrix of the noisy signal is then given by
C=C.sub.s+C.sub.v (8)
C=σ.sub.s.sup.2dd.sup.11+σ.sub.v.sup.2Γ (9)
where the target and noise signals are assumed to be uncorrelated. The fact that the first term describing the target signal, C is a rank-one matrix implies that the beneficial part (i.e., the target part) of the speech signal is assumed to be coherent/directional. Parts of the speech signal, which are not beneficial, (e.g., signal components due to late-reverberation, which are typically incoherent, i.e., arrive from many simultaneous directions) are captured by the second term.
[0200] Covariance Matrix Estimation
[0201] A look vector estimate can be found efficiently in the case of only two microphones based on estimates of the noisy input covariance matrix and the noise only covariance matrix. We select the first microphone as our reference microphone. Our noisy covariance matrix estimate is given by
where * denotes complex conjugate. Each element of our noisy covariance matrix is estimated by low-pass filtering the outer product of the input signal, XX.sup.H. We estimate each element by a first order IIR low-pass filter with the smoothing factor α∈[0; 1], i.e.
[0202] We thus need to low-pass filter four different values (two real and one complex value), i.e. Ĉ.sub.x11(m), Re{Ĉ.sub.x12(m)}, Im{Ĉ.sub.x12(m)}, and Ĉ.sub.x22(m). We don't need Ĉ.sub.x21(m) since Ĉ.sub.x 21(m)=Ĉ.sub.12*. It is assumed that the target location does not change dramatically in speech pauses, i.e. it is beneficial to keep target information from previous speech periods using a slow time constant giving accurate estimates. This means that Ĉ.sub.x is not always updated with the same time constant and does not converge to Ĉ.sub.v in speech pauses, which is normally the case. In long periods with speech absence, the estimate will (very slowly) converge towards to C.sub.no using a smoothing factor close to one. The covariance matrix C could represent a situation where the target DOA is zero degrees (front direction), such that the system prioritizes the front direction when speech is absent. C, may e.g. be selected as an initial value of C.
[0203] In a similar way, we estimate the elements in the noise covariance matrix, in that case
[0204] The noise covariance matrix is updated when only noise is present. Whether the target is present or not may be determined by a modulation-based voice activity detector. It should be noted that “Target present” (cf.
[0205] Adaptive Smoothing
[0206] The performance of look vector estimation is highly dependent on the choice of smoothing factor α, which controls the update rate of Ĉ.sub.x(m). When a is close to zero, an accurate estimate can be obtained in spatially stationary situations. When α is close to 1, estimators will be able to track fast spatial changes, for example when tracking two talkers in a dialogue situation. Ideally, we would like to obtain accurate estimates and fast tracking capabilities which is a contradiction in terms of the smoothing factor and there is a need to find a good balance. In order to simultaneously obtain accurate estimates in spatially stationary situations and fast tracking capabilities, an adaptive smoothing scheme is proposed.
[0207] In order to control a variable smoothing factor, the normalized covariance
ρ(m)=C.sub.x11.sup.−1C.sub.x12, (13)
can be observed an indicator for changes in the target DOA (where C.sub.x11.sup.−1 and C.sub.x12 are complex numbers).
[0208] In a practical implementation, e.g. a portable device, such as hearing aid, we prefer to avoid the division and reduce the number of computations, so we propose the following log normalized covariance measure
ρ(m)=Σ.sub.k{log(max{0Im{Ĉ.sub.x12}+1})−log(Ĉ.sub.x11)}, (14)
[0209] Two instances of the (log) normalized covariance measure are calculated, a fast instance {tilde over (ρ)}(m) and an instance
where {tilde over (α)} is a fast time constant smoothing factor, and the corresponding fast covariance estimate
according to
ρ(m)=Σ.sub.k{log(max{0, Im{{tilde over (C)}.sub.x12}+1})−log({tilde over (C)}.sub.x11)}, (17)
[0210] Similar expressions for the instance with variable update rate
where {tilde over (α)} is a fast time constant smoothing factor, and the corresponding fast covariance estimate
according to
ρ(m)=Σ.sub.k{log(max{0, Im{
[0211] The smoothing factor
where α.sub.0 is a slow time constant smoothing factor, i.e. α.sub.0<
[0212]
[0213]
[0214] The pre-smoothing unit (PreS) makes an initial smoothing over time (illustrated by ABS-squared units |•|.sup.2 for providing magnitude squared of the input signals X.sub.i(k,m) and subsequent low-pass filtering provided by low-pass filters LP) to provide pre-smoothed covariance estimates C.sub.x11, C.sub.x12 and C.sub.x22, as illustrated in
[0215]
[0216] The Target Present input is e.g. a control input from a voice activity detector. In an embodiment, the Target Present input (cf. signal TP in
[0217] The Fast Rel Coef, the Fast Atk Coref, the Slow Rel Coef, and the Slow Atk Coef are fixed (e.g. determined in advance of the use of the procedure) fast and slow attack and release times, respectively. Generally, fast attack and release times are shorter than slow attack and release times. In an embodiment, the time constants (cf. signals TC in
[0218] It should be noted that the goal of the computation of y=log(max(Im{x12}+1,0))−log(x11) (cf. two instances in the right part of
[0219] The adaptive low-pass filters used in
[0220]
[0221] The above scheme may e.g. be relevant for adaptively estimating a direction of arrival of alternatingly active sound sources at different locations (e.g. at different angles in a horizontal plane relative to a user wearing one or more hearing aids according to the present disclosure).
[0222]
[0223]
[0224]
[0225] It is intended that the structural features of the devices described above, either in the detailed description and/or in the claims, may be combined with steps of the method, when appropriately substituted by a corresponding process.
[0226] As used, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well (i.e. to have the meaning “at least one”), unless expressly stated otherwise. It will be further understood that the terms “includes,” “comprises,” “including,” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will also be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element but an intervening elements may also be present, unless expressly stated otherwise. Furthermore, “connected” or “coupled” as used herein may include wirelessly connected or coupled. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. The steps of any disclosed method is not limited to the exact order stated herein, unless expressly stated otherwise.
[0227] It should be appreciated that reference throughout this specification to “one embodiment” or “an embodiment” or “an aspect” or features included as “may” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the disclosure. Furthermore, the particular features, structures or characteristics may be combined as suitable in one or more embodiments of the disclosure. The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects.
[0228] The claims are not intended to be limited to the aspects shown herein, but is to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more.
[0229] Accordingly, the scope should be judged in terms of the claims that follow.