METHOD AND A HEARING DEVICE FOR IMPROVED SEPARABILITY OF TARGET SOUNDS

20170374478 · 2017-12-28

Assignee

Inventors

Cpc classification

International classification

Abstract

A hearing device, a hearing system and a method for improving a hearing impaired person's ability to perceptually separate a target sound from competing sounds, the target sound and the competing sounds forming a composite sound signal having a given frequency range, where the method comprises the steps of: (i) subdividing the frequency range of the composite sound signal into a plurality of frequency sub-bands; (ii) grouping frequency sub-bands based on comparable characteristics of the plurality of frequency sub-bands; (iii) for each of the groups calculating a group envelope; and (iv) multiplying the signal in the frequency sub-bands of each individual group by a function or functions that enhance(s) peaks of the group envelope and/or attenuates energy in troughs in the group envelope. The comparable characteristics may be the correlation between the envelope of each of the bands in the specific group of frequency sub-bands and the corresponding group envelope.

Claims

1. A hearing device for improving a hearing impaired user's ability to perceptually separate a target sound from competing sounds, the target sound and the competing sounds forming a composite sound signal having a given frequency range, the hearing aid comprising an input unit for providing a time-domain electric input signal y(n) as digital samples representing said composite sound signal in a frequency range of operation forming part of said given frequency range, n being a time-sample index, an analysis filter bank subdividing said frequency range of operation, or a part thereof, of said composite sound signal into a plurality of frequency sub-bands and providing corresponding frequency sub-band signals; a signal processing unit connected to said analysis filter bank and comprising a band grouping unit for arranging frequency sub-bands in sub-band-groups based on comparable characteristics among the plurality of frequency sub-band signals; an envelope extraction unit for calculating a group envelope for each of said sub-band groups, said group envelope comprising peaks and troughs; an enhancement unit for providing an enhancement function for each sub-band group configured to enhances said peaks in the group envelope and/or attenuate said troughs in the group envelope; and a combination unit for multiplying a signal in the frequency sub-bands of each individual sub-band-group by a respective enhancement function for the sub-band group in question, or a scaled version thereof, to provide enhanced frequency sub-band signals.

2. A hearing device according to claim 1 wherein the signal processing unit comprises a further processing unit for applying a frequency and/or level dependent gain or attenuation and/or other signal processing algorithms to said frequency sub-band signals or to said enhanced frequency sub-band signals to provide processed frequency sub-band signals

3. A hearing device according to claim 1 comprising a synthesis filter bank for converting said processed frequency sub-band signals to a time-domain electric output signal.

4. A hearing device according to claim 3 comprising an output unit for converting said time-domain electric output signal to stimuli perceivable by the user as sound.

5. A hearing device according to claim 1 comprising a hearing aid, a headset, an earphone, an ear protection device or a combination thereof.

6. A method for improving a hearing impaired person's ability to perceptually separate a target sound from competing sounds, the target sound and the competing sounds forming a composite sound signal having a given frequency range, the method comprising providing a time-domain electric input signal y(n) as digital samples representing said composite sound signal in a frequency range of operation forming part of said given frequency range, n being a time-sample index, subdividing said frequency range of operation, or a part thereof, of said composite sound signal into a plurality of frequency sub-band; arranging frequency sub-bands in sub-band-groups based on comparable characteristics among the plurality of frequency sub-bands; calculating a group envelope for each of said sub-band groups, said group envelope comprising peaks and troughs; multiplying a signal in the frequency sub-bands of each individual sub-band-group by a function that enhances said peaks of the group envelope and/or attenuates said troughs in the group envelope, thereby providing an enhancement envelope for each of said sub-band-groups.

7. Method according to claim 6, wherein said comparable characteristic comprises the correlations among the signal envelopes in said multiple frequency sub-bands.

8. A method according to claim 6, comprising the steps of: for each of said frequency sub-bands calculate the envelope of the band; for each of the sub-band-groups calculate the correlation between the envelope of each of the frequency sub-bands in the specific sub-band-group and the corresponding group envelope; for each of the sub-band groups calculate the enhancement envelope for each frequency sub-band in the sub-band-group in question; for each frequency sub-band multiply the signal in the band with the enhancement envelope of the band.

9. A method according to claim 6 comprising the steps of: calculate the correlation between the envelopes of each of said frequency sub-bands, thereby providing a correlation matrix C; based on said correlation matrix C group the frequency sub-bands into said sub-band-groups; calculate a group envelope for each of the sub-band-groups;

10. A method according to claim 6, wherein said grouping comprises the following steps: defining a threshold for correlation C_thr; selecting the row of the correlation matrix C that has the highest sum of supra-threshold values; designating the frequency sub-bands for which correlations in the selected row are greater than C_thr as the members of a first sub-band-group;

11. A method according to claim 10, wherein said grouping further comprises setting the elements in the rows and columns of the correlation matrix C corresponding to the frequency sub-bands of said first sub-band-group equal to zero, thereby providing a modified correlation matrix C.sub.M; selecting the row of the modified correlation matrix C.sub.M that has the highest sum of suprathreshold correlations; designating the frequency sub-bands for which correlations in the selected row are greater than C_thr as members of a second sub-band-group.

12. A method according to claim 6 wherein said enhancement of peaks of the group envelope and attenuation of troughs in the group envelope comprises the following steps: defining a modulation enhancement m_enh; for the defined modulation enhancement (m_enh) keeping a running tally of the group envelope's mean value, modulation depth m_group and the current amplitude offset at time n relative to said mean value, where the modulation depth is given by m_group; for each frequency sub-band in each respective sub-band-group: multiplying the signal in a current time window by (1+p(n)*c(n)*m_enh), where 0≦p(n)≦1, and where p(n) is a function of the band envelope's correlation with the group envelope.

13. A method according to claim 6 wherein said modulation enhancement m_enh is divided in two enhancement parts, one that controls the extent of peak enhancement and one that controls the extent of deepening of troughs.

14. A method according to claim 6, wherein said comparable characteristics are fundamental frequencies F.sub.0k in the amplitude variation of each separate frequency sub-band, where k is a frequency sub-band index.

15. A data processing system comprising a processor and program code means for causing the processor to perform the method of claim 6.

16. A hearing device, e.g. a hearing aid, for improving a hearing impaired user's ability to perceptually separate a target sound from competing sounds, where the hearing device comprises a data processing system according to claim 15.

17. A hearing system comprising a hearing device according to claim 1 AND an auxiliary device is moreover provided, wherein the hearing system is adapted to establish a communication link between the hearing device and the auxiliary device to provide that information, e.g. control and status signals, possibly audio signals, can be exchanged or forwarded from one to the other.

18. A hearing system according to claim 17 wherein the auxiliary device is or comprises an audio gateway device, a remote control for controlling functionality and operation of the hearing device(s), a smartphone or a combination thereof.

19. A hearing system according to claim 17 configured to run an APP allowing to control functionality of the hearing system via the auxiliary device.

20. A hearing device, e.g. a hearing aid, configured to operate at least partially on a frequency sub-band level, and configured to improve perception of a target speech signal received by the hearing device as a composite signal comprising said target speech signal and competing sound components, the hearing device comprising a perception enhancement unit based on comodulation.

21. A hearing device, according to claim 20 wherein the perception unit is configured to monitor modulation of competing sound components in at least some selected frequency sub-bands.

22. A hearing device, according to claim 20 wherein the perception unit is configured to monitor amplitude modulation of competing sound components in at least some selected frequency sub-bands.

23. A hearing device, according to claim 21 wherein the perception unit is configured to apply comodulation reflecting said modulation of the competing sound components to at least some of the frequency sub-bands.

Description

BRIEF DESCRIPTION OF DRAWINGS

[0107] The aspects of the disclosure may be best understood from the following detailed description taken in conjunction with the accompanying figures. The figures are schematic and simplified for clarity, and they just show details to improve the understanding of the claims, while other details are left out. Throughout, the same reference numerals are used for identical or corresponding parts. The individual features of each aspect may each be combined with any or all features of the other aspects. These and other aspects, features and/or technical effect will be apparent from and elucidated with reference to the illustrations described hereinafter in which:

[0108] FIGS. 1A and 1B show the basic principle that having comodulation of a masker signal over a plurality of frequency sub-bands improves auditory perception of a target signal present together with the masking signal,

[0109] FIG. 2A shows an example embodiment of a first part of a method according to the present disclosure, and

[0110] FIG. 2B shows an example embodiment of a second part of a method according to the present disclosure,

[0111] FIG. 3 shows a flow chart of a first embodiment of the method according to the present disclosure,

[0112] FIG. 4 shows a flow chart illustrating a second embodiment of the method according to the present disclosure,

[0113] FIG. 5A shows a simplified block diagram of a hearing aid according to a first embodiment of the present disclosure, and

[0114] FIG. 5B shows a simplified block diagram of a hearing aid according to a second embodiment of the present disclosure,

[0115] FIG. 6 shows a simplified block diagram of a signal processing unit according to an embodiment of the present disclosure, and

[0116] FIG. 7A shows an embodiment of a binaural hearing aid system comprising left and right hearing devices in communication with an auxiliary device, and

[0117] FIG. 7B shows the auxiliary device functioning as a user interface for the binaural hearing aid system according to the present disclosure.

[0118] The figures are schematic and simplified for clarity, and they just show details which are essential to the understanding of the disclosure, while other details are left out. Throughout, the same reference signs are used for identical or corresponding parts.

[0119] Further scope of applicability of the present disclosure will become apparent from the detailed description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the disclosure, are given by way of illustration only. Other embodiments may become apparent to those skilled in the art from the following detailed description.

DETAILED DESCRIPTION OF EMBODIMENTS

[0120] The detailed description set forth below in connection with the appended drawings is intended as a description of various configurations. The detailed description includes specific details for the purpose of providing a thorough understanding of various concepts. However, it will be apparent to those skilled in the art that these concepts may be practised without these specific details. Several aspects of the apparatus and methods are described by various blocks, functional units, modules, components, circuits, steps, processes, algorithms, etc. (collectively referred to as “elements”). Depending upon particular application, design constraints or other reasons, these elements may be implemented using electronic hardware, computer program, or any combination thereof.

[0121] The electronic hardware may include microprocessors, microcontrollers, digital signal processors (DSPs), field programmable gate arrays (FPGAs), programmable logic devices (PLDs), gated logic, discrete hardware circuits, and other suitable hardware configured to perform the various functionality described throughout this disclosure. Computer program shall be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.

[0122] The present application relates to the field of hearing devices, e.g. hearing aids.

[0123] FIGS. 1A and 1B shows the basic principle that having comodulation of a masker signal over a plurality of frequency sub-bands improves auditory perception of a target signal present together with the masking signal. A number (here 5) of frequency sub-band signals (F1, F2, F3, F4, F5) are shown with (normalized) relative amplitudes between −1 and 1 for a time segment of 1 s (cf. horizontal axis Time (s)′).

[0124] One of the key cues for improving a hearing impaired user's ability to perceptually separate a target sound from competing sounds is comodulation, where “comodulation” refers to amplitude modulations that are shared across multiple frequency sub-bands (cf. e.g. [Hall et al., 1984] or [Nelken et al., 1999]). FIGS. 1A and 1B represent a schematic illustration of comodulation and its perceptual consequence: The target sound and masker sound seem more perceptually separable when multiple comodulated masker bands are present.

[0125] In the constructed example shown in FIGS. 1A and 1B, schematic amplitude-time plots of a relatively constant-envelope target signal (reference number 1) mixed with ‘noise’ (i.e. non-target) signal(s) having a time-varying envelope (reference number 2) are shown. The target sound (1) in the middle frequency sub-band (F3) is masked by a competing sound (2) in the middle frequency sub-band (F3), and it is difficult to detect the target (FIG. 1A). The principle that is illustrated by the figure is that the presence of multiple (comodulated) ‘masker bands’ (FIG. 1B) seems to make it easier to perceptually separate out target (1) and masker (2) from one another (indicated by a clearer appearance of the constant envelope target signal (1) in frequency sub-band F3 of FIG. 1B). There is extensive evidence for this beginning with [Hall et al., 1984] and from numerous subsequent studies that found improved detection thresholds when comodulated flanking bands were added to a masker.

[0126] The plot shown in FIG. 1A illustrates the role of comodulation in enhancing the perceptual separability of a target (I) from a masker (2). In FIG. 1A, the target (1) and a masker (2) are only present in the third frequency sub-band F3 and the other bands F1, F2, F4 and F5 of FIG. 1A are silent, i.e. they contain neither a target sound nor masking sounds. In FIG. 1B, the third frequency sub-band F3 still contains both the target (1) and the masker (2), but the other frequency sub-bands F1, F2, F4 and F5 contain masking sound frequency components. Specifically, in FIG. 1B, masker energy is present in all frequency sub-bands F1 through F5 and the masker is comodulated across these bands as indicated by the arrows denoted M in the top part of FIG. 1B. The perceptual consequence of having several comodulated masker bands is that it provides a cue that helps the listener perceptually segregate the masker from the target. Although the target (I) is the same in FIGS. 1A and 1B, listeners can more easily detect the target (1) in the example shown in FIG. 1B.

[0127] It should be noted that the presence of the masker in the frequency sub-bands F1, F2, F4 and F5 makes the signal-to-noise ratio, where here we mean the sum of the energy in all of these frequency sub-bands, substantially worse than if the complete signal were present only in F3, as in FIG. 1A. Nevertheless, target detection (perception) gets better in the situation shown in FIG. 1B, as a result of comodulation of the masker between the frequency sub-bands. The principles that allow improved detection of the narrowband target in the simple example illustrated in FIGS. 1A and 1B are also believed to be important for segregating multiple broadband targets from each other.

[0128] The example shown in FIGS. 1A and 1B illustrates an essential feature that distinguishes the solution provided by the present disclosure from prior art noise reduction systems. Namely, prior art noise reduction systems treat masker energy as inherently detrimental to perception of the target, and they aim to reduce it. Contrary to this prior art approach, the solution according to the present disclosure comprises enhancing comodulation cues with the aim to (at least partially) restore the very important segregation ability of normal hearing in the hearing impaired, and that widely-held belief partly motivates the current patent application.

[0129] As it appears from FIGS. 1A and 1B, the input signal is generally a composite signal comprising both a target signal (such as a speech signal) and a competing signal (such as background noise and/or one or more competing voice signals). According to the present disclosure, a signal stream segregation is performed on this composite input signal in a process that may comprise: [0130] i. Subdividing the composite input signal in a plurality of frequency sub-bands (Band 1, Band 2, . . . Band N, cf. e.g. FIG. 2A); [0131] ii. Grouping the frequency sub-bands based on similar characteristics in the respective bands (i.e. characteristics of the individual time variant band signals that are similar for each of the bands, such as the band signal's envelopes or characteristic frequencies, e.g. fundamental frequencies, cf. e.g. FIG. 2A); [0132] iii. For each of the determined groups of frequency sub-bands calculating a group envelope of the signal in the respective band (cf. e.g. FIG. 2B); and [0133] iv. multiplying the signal in the bands of each individual group by a function that enhances (amplitude or energy in) peaks of the group envelope and/or attenuates (amplitude or energy in) troughs in the group envelope (cf. e.g. FIG. 2B).

[0134] Thus, according to a first aspect of the present disclosure (of which an embodiment is illustrated in FIG. 3) there is provided a method for improving a hearing impaired person's ability to perceptually separate a target sound from competing sounds, the target sound and the competing sounds forming a composite sound signal having a given frequency range, where the method comprises the following steps: [0135] subdividing the frequency range of said composite sound into a plurality of frequency sub-bands; [0136] grouping frequency sub-bands based on comparable characteristics of the signals of the plurality of frequency sub-bands; [0137] for each of said sub-band groups calculating a group envelope; [0138] multiplying the signal in the bands of each individual sub-band group by a function that enhances peaks of the group envelope and/or attenuates troughs in the group envelope.

[0139] In an embodiment, the magnitude of the peak enhancement is greater for some bands within the sub-band group than for other bands within the sub-band group. In an embodiment, the magnitude of the enhancement is dependent on the correlation between the individual band's envelope and the group envelope. In an embodiment of the first aspect, the magnitude of the trough attenuation is greater for some bands within the sub-band group than for other bands within the sub-band group. In an embodiment, the magnitude of the attenuation is dependent on the correlation between the individual band's envelope and the group envelope.

[0140] It should be noted that the magnitude of enhancement, or attenuation, can be made dependent on the correlation of each individual band's envelope with its sub-band group's envelope, even if non-correlation-based methods (e.g., fundamental frequency F0) are used to select the sub-band groups.

[0141] In an embodiment, the comparable characteristic is the correlations among the signal envelopes in said multiple frequency sub-bands (e.g. in those frequency bands that exhibit correlation with each other in a specific range of a correlation measure (e.g. cross-correlation) are allocated to the same sub-band group).

[0142] In an embodiment, the comparable characteristics are fundamental frequencies F.sub.0k (and/or harmonics thereof) in the amplitude variation over time of each separate frequency sub-band, where k is a frequency sub-band index.

[0143] In an embodiment, the method comprises the steps of: [0144] for each of said frequency sub-bands calculate the envelope of the band; [0145] for each of the sub-band groups calculate the correlation between the envelope of each of the bands in the specific sub-band group and the corresponding group envelope; [0146] for each of the sub-band groups calculate the enhancement envelope for each band in this sub-band group; [0147] for each band multiply the signal in the band with the enhancement envelope of the band.

[0148] In an embodiment, the method comprises the steps of: [0149] for each of said frequency sub-bands calculate the envelope of band; [0150] calculate the correlation between the envelopes of each of said frequency sub-bands, thereby providing a correlation matrix C; [0151] based on said correlation matrix C, group the frequency sub-bands into sub-band groups; [0152] calculate a group envelope for each of the sub-band groups; [0153] for each of the sub-band groups calculate the correlation between the envelope of each of the bands in the specific sub-band group and the corresponding group envelope; [0154] for each of the sub-band groups calculate the enhancement envelope for each frequency sub-band in this sub-band group; [0155] for each frequency sub-band multiply the signal in the band with the enhancement envelope of the frequency sub-band.

[0156] In an embodiment of the first aspect, the grouping comprises the steps of: [0157] a threshold for correlation C_thr is defined; [0158] the row of the correlation matrix C is selected that has the highest sum of suprathreshold values; [0159] the bands for which correlations in the selected row are greater than C_thr are designated as the members of a first group; [0160] in the rows and columns of the correlation matrix C corresponding to the bands of said first sub-band group, the elements of are set equal to zero, thereby providing a modified correlation matrix C.sub.M; [0161] the row of the modified correlation matrix C.sub.M that has the highest sum of suprathreshold correlations is selected; [0162] the bands for which correlations in the selected row are greater than C_thr are designated as members of a second sub-band group.

[0163] In an embodiment, where more than 2 groups of bands are identified, a second modified correlation matrix C.sub.M′ is preferably formed and a 3.sup.rd group of bands selected, and so on until either all OFF-diagonal elements of the modified matrix are below C_thr or until some predefined maximum number of groups is reached.

[0164] In an embodiment of the first aspect, the accentuation of peaks of the group envelope and attenuation of energy in troughs in the group envelope comprises the following steps: [0165] defining a modulation enhancement m_enh; [0166] for the defined modulation enhancement (m_enh) keeping a running tally of the group envelope's mean value, modulation depth m_group and the current amplitude offset at time n relative to said mean value, where the offset value, c(n), gives the current value of the group envelope relative to its running mean value. The time-varying function c(n) represents the group modulation envelope and is defined such that c(n) is positive when the group envelope is above its running mean and negative when the group envelope is below its running mean. [0167] for each frequency sub-band (index k) in each respective sub-band group (index j): [0168] multiplying the signal in a current time window by (1+p(n)*c(n)*m_enh) (termed the ‘enhancement envelope’ or the ‘enhancement function’ f.sub.e(j,p(k,n)) in relation to FIG. 6 below), where 0≦p(n)≦1, and where p(n) determines how much of m_enh is applied in the band (k) at a given point in time; p(n) can, for example, be set to depend on the band envelope's correlation with the group envelope; [0169] multiplying the signal by (1+p(n)*c(n)*m_enh); thereby increasing the magnitude of the peaks and deepening the troughs of the comodulation among the bands in the group.

[0170] In an embodiment the frequency sub-band specific parameter p(n) depends on inputs from detectors or classifiers.

[0171] In an embodiment of the first aspect, the modulation enhancement m_enh is divided in two enhancement parts, one that controls the extent of peak enhancement and one that controls the extent of deepening of troughs. This has the advantage that the enhancement of the two parts can be independently controlled. In an embodiment specific limitations may be put on the maximum allowed peak enhancement. In an embodiment specific limitations may be put on the maximum allowed trough attenuation, e.g., to prevent the modulation envelope from crossing zero signal amplitude, yielding a greater than 100 percent modulation (overmodulation).

[0172] According to a second aspect of the present disclosure, there is provided a hearing device for improving a hearing impaired user's ability to perceptually separate a target sound from competing sounds, where the hearing device comprises a processor configured for carrying out the method according to the first aspect of the present disclosure, thereby providing a stream segregation cue enhanced output signal for presentation to a user of the hearing device.

[0173] In an embodiment of the second aspect, the hearing device is or comprises a hearing aid.

[0174] According to a third aspect of the present disclosure, there is provided a data processing system comprising a processor provided with software adapted to perform at least some (such as a majority or all) of the steps of the method according to the first aspect of the disclosure.

[0175] According to a fourth aspect of the present disclosure, there is provided software able to perform the method according to the first aspect of the disclosure, which software may be stored on or encoded as one or more instructions or code on a tangible computer-readable medium. The computer readable medium includes computer storage media adapted to store a computer program comprising program codes, which when run on a processing system causes the data processing system to perform at least some (such as a majority or all) of the steps of the method according to the first aspect of the disclosure.

[0176] FIG. 2A shows an example embodiment of a first part of a method according to the present disclosure, and FIG. 2B shows an example embodiment of a second part of a method according to the present disclosure.

[0177] Referring to FIG. 2A, a signal 8 is provided to a filter bank (e.g. a bank of band pass filters 10, 11). In the example shown in FIGS. 1A and 1B, five such band pass filters were used, but it is understood that any suitable number of such filters may be used as deemed necessary. Each respective of the band pass filters 10, 11 provide a band passed (frequency sub-band) output signal 12. The bands may be overlapping or non-overlapping. The frequency sub-bands (1, . . . , N) may together cover a part of or the entire frequency range of operation of a hearing aid, e.g. from 0 Hz (or 20 Hz or more) to 8 kHz (or more, e.g. 10 kHz or more).

[0178] Although the example in FIG. 2A assigns the frequency sub-bands into 2 groups of frequency sub-bands, this approach can easily be extended so as to have 3 or more sub-band groups.

[0179] The band passed output signal 12 from each respective band pass filter is provided to a corresponding envelope extractor 13, 14 that determines the envelope as a function of time of the (frequency sub-band) output signal provided by the respective band pass filter. Envelope extraction may e.g. be performed by filtering, rectification and filtering, Hilbert transformation, or phase lock loop techniques.

[0180] Based on the determined signal envelopes of each respective band 1 through N, the correlations among the signal envelopes of the N frequency sub-band signals (cf. Y(k,m) in FIG. 5B, FIG. 6) are calculated, thereby obtaining a correlation matrix C. Based on the content of the correlation matrix C, a grouping of the frequency sub-bands 1 through N may be performed as follows:

[0181] Part A: Cross-correlation, thresholding and grouping of bands: [0182] a. A threshold for correlation C_thr is defined; [0183] b. The row of the correlation matrix C is selected that has the highest sum of suprathreshold values; [0184] c. The bands for which correlations in the selected row are greater than C_thr are designated as the members of Group 1; [0185] d. The correlation values of the rows and columns of the correlation matrix C corresponding to Group 1 bands are set equal to zero, thereby providing a modified correlation matrix C.sub.M; [0186] e. The row of the modified correlation matrix C.sub.M (modified in the previous step) that has the highest sum of suprathreshold correlations is selected; [0187] f. The bands for which correlations in the selected row are greater than C_thr are designated as the members of Group 2.

[0188] The above outlined procedure for obtaining the sub-band groups is illustrated by the further non-limiting example, in which the original correlation matrix is:

TABLE-US-00001 band 1 band 2 band 3 band 4 band 5 band 1 1 0.9 0.7 0.5 0.2 band 2 0.9 1 0.5 0.5 0.3 band 3 0.7 0.7 1 0.8 0.6 band 4 0.5 0.5 0.8 1 0.85 band 5 0.2 0.3 0.6 0.85 1

[0189] The threshold of correlation C_thr is set to 0.75 in this example (this value may be chosen differently, e.g. to be larger than 0.75 or smaller than 0.75, depending on the particular situation (acoustic environment, configuration of frequency sub-bands, hearing impairment of the user, etc.)). The elements of the original matrix having suprathreshold values are highlighted above.

[0190] The row of the correlation matrix C that has the highest sum of suprathreshold values is row 4 (band 4).

[0191] The bands of row 4 that have a correlation value greater than C_thr are chosen as Group 1. Thus Group 1 consists of band 3, band 4 and band 5.

[0192] The matrix elements corresponding to Group 1 are set equal to zero, thereby yielding the modified matrix:

TABLE-US-00002 band 1 band 2 band 3 band 4 band 5 band 1 1 0.9 0 0 0 band 2 0.9 1 0 0 0 band 3 0 0 0 0 0 band 4 0 0 0 0 0 band 5 0 0 0 0 0

[0193] The row of the modified matrix above that has the highest sum of supra-threshold correlations is selected. In this example both row 1 and row 2 have the sum 1.9 and the corresponding bands 1 and 2 are selected for Group 2.

[0194] According to the disclosure, the grouping of frequency sub-bands could alternatively be based on other methods than the correlation method described above.

[0195] In an embodiment of the disclosure, the grouping of frequency sub-bands is based on identification of fundamental frequencies F.sub.0k of each separate frequency sub-band k and subsequently grouping bands, which have fundamental frequencies F.sub.0k within a predefined range. Subsequent to this grouping of frequency sub-bands, the method continue as described below under Part B (cf. also FIG. 2B).

[0196] After the grouping has been performed, each or the determined sub-band groups are subjected to the steps indicated in FIG. 2B.

[0197] In step 19 the group envelope is calculated for sub-band group j (j=1 or 2 in the example shown in FIG. 2A). The group envelope can be calculated using a number of different approaches, such as averaging, e.g. frequency weighted averaging, where, for example, bands are weighted by their importance for speech comprehension. Another approach would be summing and subsequent extraction of the envelope of the resulting signal. Other weighting schemes may be used according to the application in question, e.g. depending on the expected input signal, e.g. characteristics of the input signal.

[0198] In step 20 the correlation between the envelope of each individual band belonging to Group j and the calculated group envelope is calculated.

[0199] In step 21 the enhancement envelope is calculated based on the correlations determined in step 20.

[0200] In step 22 the signal in each of the frequency sub-bands belonging to the specific sub-band group is multiplied by the enhancement envelope determined in step 21, thereby providing the desired segregation cue enhanced signal.

[0201] Calculation of the enhancement envelope comprises according to an embodiment of the disclosure the following steps:

[0202] Part B: Calculation of enhancement envelope:

[0203] a. Defining a modulation enhancement (m_enh) for a given sub-band group;

[0204] b. For said defined modulation enhancement (m_enh):

[0205] c. Keep a running tally comprising the group envelope's: [0206] i. Mean; [0207] ii. Modulation depth (m_group); and [0208] iii. Current amplitude offset at time n relative to the mean, expressed as c(n), as described elsewhere in this application.

[0209] d. For each frequency sub-band in the sub-band group: [0210] i. Determine the band envelope of the frequency sub-band; [0211] ii. Multiply the frequency sub-band signal in the current time window by (1+p(n)*c(n)*m_enh), where p(n) is between 0 and 1 and can depend on factors that include but are not limited to the band envelope's correlation with the group envelope, inputs from detectors, inputs from classifiers, etc.

[0212] Because c(n) under Item c(iii) above reflects the modulation of the group envelope, multiplying the signal by (1+p(n)*c(n)*m_enh) increases the peaks and deepens the troughs of the comodulation among the bands in the sub-band group.

[0213] According to an embodiment, the modulation enhancement m_enh is subdivided into two parts, one that controls the extent of peak enhancement and one that controls the extent of deepening of troughs.

[0214] Referring to FIG. 3 there is shown a flow chart illustrating basic steps of an embodiment of the method according to the present disclosure.

[0215] In step 23 there is provided an input signal, for instance a (processed, time variant) output signal from a microphone in a hearing aid (e.g. comprising a target signal x mixed with noise components v). In step 24 the total frequency range (or optionally a portion hereof) of the input signal is subdivided into a number of frequency sub-bands. In FIG. 1, five such bands were shown, but another number of frequency sub-bands (adjacent or separate) could also be used. In step 25 a comparable characteristic of the signals in the frequency sub-bands is determined. Examples of comparable characteristics would e.g. be the signal envelopes of each of the frequency sub-bands (k), or fundamental frequencies F.sub.0k in the amplitude variation over time of each separate frequency sub-band.

[0216] In step 26 the frequency sub-bands are grouped based on the comparable characteristics determined in step 25. In the example embodiments described, the frequency sub-bands are grouped in two sub-band groups: Group 1 and Group 2, but it is understood that other numbers of groups could also be used.

[0217] In step 27, a group envelope is calculated as described above for each of the determined sub-band groups (j=1, 2).

[0218] In steps 28 and 29, respectively, the signal in each of the frequency sub-bands is multiplied by a (enhancement) function that enhances the peaks of the group envelope for the particular sub-band group and that attenuates the troughs of the group envelope for the particular sub-band group (for Group 1 and Group 2, respectively).

[0219] FIG. 4 shows a flow chart illustrating a second embodiment of the method according to the present disclosure.

[0220] In step 30, there is provided an input signal, for instance a (processed) output signal from a microphone in a hearing, aid. In step 31, the total frequency range (or optionally a portion hereof) of the input signal is subdivided into a number of frequency sub-bands, and in step 32, the envelope of the signal in each of the frequency sub-bands are calculated.

[0221] In step 33, the correlation between the envelopes of each of the frequency sub-band signals are calculated, thereby providing a correlation matrix C (e.g. as shown in the numerical examples given above).

[0222] In step 34, grouping of the frequency sub-bands is performed based on the correlation matrix C, as e.g. described in detail above.

[0223] In step 35, a group envelope is determined for each of the sub-band groups found in step 34.

[0224] In step 36, the correlation between each band envelope and the corresponding group envelope is determined for each of the sub-band groups.

[0225] In step 37, an enhancement envelope is calculated for each frequency sub-band in each sub-band group based on the correlations found in step 36.

[0226] In step 38, the signal in each separate frequency sub-band is multiplied with the enhancement envelope of the band determined in step 37.

[0227] FIG. 5 shows a simplified block diagram of a hearing aid according to a first embodiment of the present disclosure, and FIG. 5B shows a simplified block diagram of a hearing aid according to a second embodiment of the present disclosure.

[0228] Referring to FIG. 5A there is shown a schematic block diagram of a hearing aid (HA) 39 configured to carry out the method according to the present disclosure. The hearing aid 39 comprises an input unit (IU) 41 provided with an input transducer (IT) 43, e.g. a microphone, for converting an acoustic signal (Acoustic input) 40 to an electric signal, which electric signal is provided to an A/D converter (AD) 44. The digital signal from the A/D converter is provided to a signal processing unit (SPU) 45 that comprises software code for executing the various steps of the method according to the present disclosure. The processed output digital signal is provided to a D/A converter (DA) 46 in an output unit (OU) 42, and the analogue signal from D/A converter 46 drives an output transducer (OT) 47, e.g. a loudspeaker (receiver), that converts the electrical output signal to an acoustic output signal (Acoustic output) 48. In embodiments, the output unit may (additionally or alternatively) comprise a vibrator for a bone-conduction type hearing aid or a multi-electrode array of a cochlear implant type hearing aid. The output of the signal processing unit 45 could be the stream segregation cue enhanced signal provided by the method according to the present disclosure or a processed version hereof (cf. e.g. FIG. 6). Further, the signal processing unit 45 may include an analysis filter bank (FBA) configured for sub-dividing the frequency range into a number of frequency sub-bands (for instance the five bands F1, F2, F3, F4 and F5 described in FIG. 1A, 1B above) and a corresponding synthesis filter bank (FBS) configured to recombine the frequency sub-bands into one single frequency band. An embodiment of a hearing aid as described in FIG. 5A but comprising separate analysis (AFB) and synthesis (SFB) filter banks in the forward path of the hearing aid between the input (IT) and output (OT) transducers is illustrated in FIG. SB. Additionally, the frequency sub-band signals that are input and output signals of the signal processing unit (SPU) are denoted Y(k,m) and Z(k,m), respectively, k being the frequency index (k=1, N) and m being the time frame index.

[0229] FIG. 6 shows a simplified block diagram of a signal processing unit according to an embodiment of the present disclosure. The input unit (IU) shown in FIG. 5B provides a time-domain electric input signal y(n) as digital samples representing a composite input sound signal (e.g. comprising a number of speech signal components) in a frequency range of operation of the hearing device, t being time, and n being a time-sample index. The analysis filter bank (FBA) shown in FIG. 5B subdivides the frequency range of operation of the hearing aid, or a part thereof, into a plurality of frequency sub-bands Y(k,m) of the composite sound signal, k being a frequency sub-band index (k=1, N, N being the number of sub-bands), and m being a time-frame index. Each frame comprises a number of samples, e.g. 64 or 128. The frames are non-overlapping or overlapping, typically overlapping. The signal processing unit (SPU), which is connected to the analysis filter bank FBA and receives frequency sub-band signals Y(k,m), comprises a frequency sub-band grouping unit (BGU) for arranging frequency sub-bands (k) in sub-band-groups SBG.sub.j, j=1, . . . , N.sub.SBG, based on comparable characteristics among the plurality of frequency sub-bands Y(k,m), and provides grouped sub-band signals Y.sub.SBGj(k,m). N.sub.SBG is the number of sub-band groups. N.sub.BSG depends e.g. on the type of target signal. N.sub.BSG may e.g. depend on the type and number of the currently present noise sources. N.sub.BSG is at least one, such as larger than or equal to two. In the exemplary embodiment of FIG. 6, N.sub.SBG=3. The three sub-band groups SBG1, SBG2, and SBG3 are represented by sub-band signals Y.sub.SBG1(k,m), Y.sub.SBG2(k,m), Y.sub.SBG3(k,m). In an embodiment, the three groups of frequency sub-band signals together constitute the N sub-band signals Y(k,m) of the composite sound signal (e.g. in that the respective sub-band groups together (in a mathematical sense that their union) consist of the frequency sub-bands k=1, . . . , N). The comparable characteristics among the plurality of frequency sub-bands Y(k,m) that is used to form the sub-band groups may e.g. relate to similar modulation properties among the frequency sub-bands. In an embodiment, the comparable characteristics comprises the correlations among the signal envelopes in said multiple frequency sub-bands In an embodiment, the frequency sub-band grouping unit (BGU) is configured to assign a given sub-band to a given sub-band group, if it fulfils a given criterion for the comparable characteristics assigned to that sub-band group (e.g. is within a distance measure from a given value (or is larger than or smaller than a given value, etc.) of the characteristics, e.g. a given correlation value). The signal processing unit (SPU) further comprises an envelope extraction unit (EXU) for calculating a group envelope for each of said sub-band groups SBG.sub.1, represented by sub-band signals Y.sub.SBG1(k,m), Y.sub.SBG2(k,m), Y.sub.SBG3(k,m), respectively. The envelope extraction unit (EXU) provides as an output respective group envelope signals E.sub.G(j), j=1, . . . , N.sub.SBG (here N.sub.SBG=3). Each group envelope signal E.sub.G(1), E.sub.G(2), E.sub.G(3) comprises peaks and troughs (as schematically indicated above the envelope extraction unit (EXU)). The group envelope may e.g. be determined as an average of each envelope of the sub-band group in question, or using frequency weighted averaging, where bands are weighted, e.g. by their importance for speech comprehension. The signal processing unit (SPU) further comprises an enhancement unit (EHU) for providing respective enhancement functions f.sub.e(j), j=1, N.sub.SBG (here N.sub.SBG=3). Each enhancement function f.sub.e(1), f.sub.e(2), f.sub.e(3) enhances the peaks and/or attenuates the troughs in respective ones of the group envelope signals E.sub.G(1), E.sub.G(2), E.sub.G(3). Thereby enhanced group envelope signals can be determined EH.sub.G(j)=E.sub.G(j)*f.sub.e(j), j=1, . . . , N.sub.SBG (here N.sub.SBG=3) (as schematically indicated above the enhancement unit (EHU)). In an embodiment, the enhancement functions f.sub.e(1), f.sub.e(2), f.sub.e(3), which are (or may be) different from sub-band group to sub-band group, may also be different from frequency sub-band to frequency sub-band within a sub-band group, e.g. in dependence of a parameter defining a difference between the group envelope of the sub-band group in question and the envelope of the frequency sub-band in question. In other words, f.sub.e(j)=f.sub.e(j,p), where p is a parameter, e.g. related to correlation between group and band envelopes. The frequency sub-band parameter p may thus depend on frequency sub-band index k (and may be time dependent as well, p=p(k,n)). In an embodiment, the enhancement functions f.sub.e(j) for different frequency sub-bands (k.sub.j) of a given sub-band group j are scaled versions of f.sub.e(j) (e.g. dependent on a parameter of the individual sub-bands k.sub.j). Respective multiplication units ‘X’ are configured to multiply the frequency sub-band signals (Y.sub.SBG1(k,m), Y.sub.SBG2(k,m) Y.sub.SBG3(k,m) in FIG. 6) in each individual sub-band-group (SBG1, SBG2, SBG3 in FIG. 6) by a respective one of the enhancement functions f.sub.e(1), f.sub.e(2), f.sub.e(3) (or individualized versions thereof f.sub.e(j,p(k,n))) to provide enhanced frequency sub-band signals (ES.sub.SBG1(k,m), ES.sub.SBG2(k,m), ES.sub.SBG3 (k,m) in FIG. 6). The individual enhanced frequency sub-band signals ES(k,m), k=1, N (constituted by the enhanced sub-band signals ES.sub.SBG1(k,m), ES.sub.SBG2(k,m), ES.sub.SBG3(k,m) in FIG. 6), have been modified by the enhancement of comodulation between frequency sub-bands to thereby improve a user's ability to separate a target signal from noise. In the embodiment of a signal processing unit (SPU) of FIG. 6, the enhanced frequency sub-band signals are processed by (further) processing unit (FPU), e.g. for applying a frequency and/or level dependent gain to the enhanced frequency sub-band signals to provided (further) processed signals Z(k,m), k=1, . . . , N. Other processing algorithms may additionally (or alternatively) be applied in the processing unit (FPU), such as feedback cancellation, noise reduction, etc. In an embodiment, the input unit may comprise more than one microphone, e.g. 2 or more. In an embodiment, the hearing device comprises a multi input beamformer filtering unit for providing a spatially filtered signal. The scheme of providing comodulation in frequency sub-bands of a number of sub-band groups may be applied to each microphone input signal separately and/or to a spatially filtered (beamformed) signal.

[0230] FIG. 7A shows an embodiment of a binaural hearing aid system comprising left and right hearing devices in communication with an auxiliary device, and FIG. 7B shows the auxiliary device functioning as a user interface for the binaural hearing aid system.

[0231] FIG. 7A shows an embodiment of a binaural hearing system comprising left (second) and right (first) hearing devices (HA.sub.l, HA.sub.r) in communication with a portable (handheld) auxiliary device (AD) functioning as a user interface (UI) for the binaural hearing aid system. In an embodiment, the binaural hearing system comprises the auxiliary device (Aux, and the user interface UI). In the embodiment of FIG. 7A, wireless links denoted IA-WL (e.g. an inductive link between the left and right hearing devices) and WL-RF (e.g. RF-links (e.g. Bluetooth) between the auxiliary device Aux and the left HA.sub.l, and between the auxiliary device Aux and the right HA.sub.r, hearing device, respectively) are indicated (implemented in the devices by corresponding antenna and transceiver circuitry, indicated in FIG. 7A in the left and right hearing devices as RF-IA-Rx/Tx-l and RE-IA-Rx/Tx-r, respectively). In the acoustic situation illustrated by FIG. 7A, a dominant sound source, e.g. a voice of a person, denoted Target Sound, is located to the right of the user (U) and a noise sound field, possibly comprising competing voice/speech signals and/or natural or artificial noise, denoted Noise, is indicated around the user.

[0232] The user interface (UI) of the auxiliary device (Aux) is shown in FIG. 7B. The user interface comprises a display (e.g. a touch sensitive display) displaying a screen of a Hearing instrument Remote control APP for controlling the hearing system and a number of predefined actions regarding functionality of the binaural hearing system (or of a bilateral hearing aid system or of single hearing aid). In the exemplified (part of the) APP, a user (U) has the option of influencing a mode of operation via the selection of one of a number of predefined (or configurable) programs, each optimized for specific acoustic situations (in box Select program). The exemplary acoustic situations are: Multienvironment, Conversation, Music, Tinnitus, and Comodulation, each illustrated as an activation element, which is selected one at a time by clicking on the element. Each exemplary acoustic situation is associated with the activation of specific algorithms and specific processing parameters (programs) of the left and right hearing devices. In the example of FIG. 7B, the acoustic situation Comodulation has been chosen, (as indicated by bold italic highlight of the corresponding activation element on the screen). The acoustic situation Comodulation refers to a specific mode of operation of the hearing system, where a target (speech) sound source present in the acoustic environment of the user (as indicated in FIG. 7A by the element Target Sound) together with one or more noise sources (or competing voice sources). In the exemplified remote control APP-screen of FIG. 7B, the user has the option of helping identifying the target sound source (cf. box Comodulation enhancement. Select target signal). The user has the option of clicking on the smiley icon custom-character representing a target source and is encouraged to press (hold down) the icon for a period of time, where the target sound is present in the environment of the user. Thereby, the hearing aid(s) are guided in the task of identifying spectral characteristics of the target signal (cf. (1) in FIG. 1A, 1B, including the frequency sub-bands where the target signal is present) and to apply the appropriate comodulation (characteristic of the ‘noise’, cf. (2) in FIG. 1A, 1B) in neighbouring frequency sub-bands. Alternatively, this task may be executed automatically, e.g. by the left and right hearing devices individually, or in common, and/or in collaboration with the auxiliary device (e.g. using one or more microphone signals of an auxiliary device). The noise components (denoted ‘Noise’ in FIG. 7A) can be ‘artificial’ noise from traffic, car noise, mechanical devices (fans, air condition, etc.), but may also include (competing) voices from other persons than the target source.

[0233] The auxiliary device Aux comprising the user interface UI is adapted for being held in a hand of a user (U), and hence convenient for displaying information about the hearing aid system and/or for the user to influence its function. In an audio streaming mode of operation of the hearing system, audio signals (e.g. from a telephone conversation or a music or other sounds) may be transferred from the auxiliary device to the left and right hearing aids (using wireless links WL-RF, and optionally IA-WL), cf. signals ADCD.sub.l and ADCD.sub.r in FIG. 7A). In a remote control mode of operation (as illustrated in FIG. 7B), control data and/or information data (and/or audio data) may be exchanged between the auxiliary device and the left and right hearing devices (using wireless links WL-RF, and optionally IA-WL), cf. signals ADCD.sub.l and ADCD.sub.r in FIG. 7A.

[0234] The wireless communication link(s) (WL-RF, IA-WL in FIG. 7A) between the hearing devices and the auxiliary device and between the left and right hearing devices may be based on any appropriate technology with a view to the necessary bandwidth and available part of the frequency spectrum. In an embodiment, the wireless communication link (WL-RF) between the hearing devices and the auxiliary device is based on far-field (e.g. radiated fields) communication e.g. according to Bluetooth or Bluetooth Low Energy or similar standard or proprietary scheme. In an embodiment, the wireless communication link (IA-WL) between the left and right hearing devices is based on near-field (e.g. inductive) communication.

[0235] It is intended that the structural features of the devices described above, either in the detailed description and/or in the claims, may be combined with steps of the method, when appropriately substituted by a corresponding process.

[0236] As used, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well (i.e. to have the meaning “at least one”), unless expressly stated otherwise. It will be further understood that the terms “includes,” “comprises,” “including,” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will also be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element but an intervening elements may also be present, unless expressly stated otherwise. Furthermore. “connected” or “coupled” as used herein may include wirelessly connected or coupled. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. The steps of any disclosed method is not limited to the exact order stated herein, unless expressly stated otherwise.

[0237] It should be appreciated that reference throughout this specification to “one embodiment” or “an embodiment” or “an aspect” or features included as “may” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the disclosure. Furthermore, the particular features, structures or characteristics may be combined as suitable in one or more embodiments of the disclosure. The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects.

[0238] The claims are not intended to be limited to the aspects shown herein, but is to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more.

[0239] Accordingly, the scope should be judged in terms of the claims that follow.

REFERENCES

[0240] [Hall et al., 1984] Hall, J. W., Haggard, M. P., Fernandes, M. A. (1984). “Detection in noise by spectro-temporal pattern analysis”, J. Acoust. Soc. Am. 76, 50-56. [0241] [Nelken et al., 1999] Nelken I., Rotman Y., and Yosef O. B. (1999). “Responses of auditory cortex neurons to structural features of natural sounds,” Nature 397, 154-157.