HEARING DEVICE OR SYSTEM COMPRISING A COMMUNICATION INTERFACE
20220337960 · 2022-10-20
Assignee
Inventors
Cpc classification
H04R25/40
ELECTRICITY
H04R25/50
ELECTRICITY
H04R25/554
ELECTRICITY
International classification
Abstract
A hearing device, e.g. a hearing aid, comprises a) at least one input transducer for converting sound in the environment of the hearing device to respective at least one acoustically received electric input signal or signals representing said sound; b) a wireless receiver for receiving an audio signal from a wireless transmitter of a sound capturing device for picking up sound in said environment and providing a wirelessly received electric input signal representing said sound; and c) a processor configured c1) to receive said at least one acoustically received electric input signal or signals, or a processed version thereof; c2) to receive said wirelessly received electric input signal; and c3) to provide a processed signal. The processor comprises a signal predictor for estimating future values of said wirelessly received electric input signal in dependence of a multitude of past values of said signal, thereby providing a predicted signal. The hearing device further comprises d) an output transducer for presenting output stimuli perceivable as sound to the user in dependence of said processed signal from said processor, or a further processed version thereof. The processor is configured to provide said processed signal in dependence of the predicted signal or a processed version thereof 1) alone, or 2) mixed with said at least one acoustically received electric input signal or signals, or a processed version thereof. A hearing device comprising an earpiece and a separate audio processing device is further disclosed. The invention may e.g. be used in hearing devices in wireless communication with audio capture devices in an immediate environment of the user wearing the hearing device.
Claims
1. A hearing device comprising at least one input transducer for converting sound in the environment of the hearing device to respective at least one acoustically received electric input signal or signals representing said sound; a wireless receiver for receiving an audio signal from a wireless transmitter of a sound capturing device for picking up sound in said environment and providing a wirelessly received electric input signal representing said sound; a processor configured to receive said at least one acoustically received electric input signal or signals, or a processed version thereof; to receive said wirelessly received electric input signal; and to provide a processed signal, the processor comprising a signal predictor for estimating future values of said wirelessly received electric input signal in dependence of a multitude of past values of said signal, thereby providing a predicted signal; an output transducer for presenting output stimuli perceivable as sound to the user in dependence of said processed signal from said processor, or a further processed version thereof, wherein the processor is configured to provide said processed signal in dependence of the predicted signal or a processed version thereof alone, or mixed with said at least one acoustically received electric input signal or signals, or a processed version thereof.
2. A hearing device according to claim 1 wherein the processor comprises a delay estimator configured to estimate a time-difference-of-arrival of sound from a given sound source in said environment at said processor between said acoustically received electric input signal or signals, or a processed version thereof, and said wirelessly received electric input signal.
3. A hearing device according to claim 1 comprising a wireless transmitter for transmitting data to another device.
4. A hearing device according to claim 1 wherein the processor comprises a selection controller configured to include said estimated predicted signal or pails thereof in said processed signal in dependence of a sound quality measure.
5. A hearing device according to claim 1 comprising a transform unit, or respective transform units, for providing said at least one acoustically received electric input signal or signals, or a processed version thereof, and/or said wirelessly received electric input signal in a transform domain.
6. A hearing device according to claim 5 wherein said transform units are configured to provide said signals in the frequency domain.
7. A hearing device according to claim 6 wherein the processor is configured to include said estimated future values of said wirelessly received electric input signal in the processed signal only in a limited part of an operating frequency range of the hearing device.
8. A hearing device according to claim 7 wherein the processed signal comprises future values of said wirelessly received electric input signal only in frequency bands or time-frequency regions that fulfil a sound quality criterion.
9. A hearing device according to claim 1 comprising a beamformer configured to provide a beamformed signal based on said at least one acoustically received electric input signal or signals and said predicted signal.
10. A hearing device according to claim 1 configured to apply spatial cues to the predicted signal before being presented to the user.
11. A hearing device according to claim 2 configured to only activate the signal predictor in case the time-difference-of-arrival is larger than a minimum value.
12. A hearing device comprising at least one earpiece configured to be worn at or in an ear of a user; and a separate audio processing device; the at least one earpiece comprising an input transducer for converting sound in the environment of the hearing device to an acoustically received electric input signal representing said sound; a wireless transmitter for transmitting said acoustically received electric input signal, or a part thereof, to said audio processing device; a wireless receiver for receiving a first processed signal from said audio processing device, at least in a normal mode of operation of the hearing device; and an output transducer for converting a final processed signal to stimuli perceived by, said user as sound, the audio processing device comprising a wireless receiver for receiving said acoustically received electric input signal, or a part thereof, from the earpiece, and to provide a received signal representative thereof; a computing device for processing said received signal, or a signal originating therefrom, and to provide a first processed signal; a transmitter for transmitting said first processed signal to said earpiece; wherein said earpiece or said audio processing device comprises a signal predictor for estimating future values of said received signal, or a processed version thereof, in dependence of a multitude of past values of said signal, thereby providing a predicted signal; wherein said signal predictor is configured to fully or partially compensate for a processing delay incurred by one or more, such as all of said transmission of the acoustically received electric input signal from the hearing device to the audio processing device, said processing in the audio processing device, and said transmission of the predicted signal or a processed version thereof to said earpiece and its reception therein; wherein the final processed signal, at least in a normal mode of operation of the hearing device, is constituted by or comprises at least a part of said predicted signal.
13. A hearing device according to claim 12 wherein the audio processing device comprises said signal predictor.
14. A hearing device according to claim 12 wherein the earpiece comprises an earpiece-computing device configured process said acoustically received electric input signal and/or to said first processed signal received from the audio processing device, and to provide said final processed signal, and wherein the earpiece computing device, at least in a normal mode of operation of the hearing device, is configured to mix the acoustically received electric input signal, or the modified signal, with a predicted signal received from the audio processing device and to provide the mixture as the final processed signal to the output transducer.
15. A hearing device according claim 14 wherein the earpiece computing device, in an earpiece-mode of operation, where said first processed signal is not received from the audio processing device, or is received in an inferior quality, is configured to provide the final processed signal to the output transducer in dependence of the acoustically received input signal.
16. A hearing device or system comprising at least one earpiece configured to be worn at or in an ear of a user and to receive an acoustic signal and to present a final processed signal to the user; and a separate audio processing device in communication with the at least one earpiece; wherein the earpiece is configured to transmit said acoustic signal to the audio processing device; and wherein the audio processing device comprises a signal predictor for estimating future values of the acoustical signal received by the at least one earpiece, or a processed version thereof, in dependence of a multitude of past values of said signal, thereby providing a predicted signal; and wherein the predictor is configured to compensate for or reduce the delay incurred by the processing being conducted in the external processing device.
17. A hearing device or system according to claim 16 wherein the audio processing device is configured to transmit the predicted signal or a processed version thereof to said earpiece; and wherein the earpiece is configured to determine said final processed signal in dependence of said predicted signal.
18. A hearing device or system according to claim 16 wherein the signal predictor is configured to fully or partially compensate for a processing delay incurred by one or more, such as all of a) a transmission of the acoustically received electric input signal from the hearing device to the audio processing device, b) a processing in the audio processing device providing a predicted signal, and c) a transmission of the predicted signal or a processed version thereof to said earpiece and its reception therein.
19. A hearing device according to claim 1 comprising a hearing instrument, a headset, an earphone, an ear protection device or a combination thereof.
20. A hearing device according to claim 12 comprising a hearing instrument, a headset, an earphone, an ear protection device or a combination thereof.
Description
BRIEF DESCRIPTION OF DRAWINGS
[0113] The aspects of the disclosure may be best understood from the following detailed description taken in conjunction with the accompanying figures. The figures are schematic and simplified for clarity, and they just show details to improve the understanding of the claims, while other details are left out. Throughout, the same reference numerals are used for identical or corresponding parts. The individual features of each aspect may each be combined with any or all features of the other aspects. These and other aspects, features and/or technical effect will be apparent from and elucidated with reference to the illustrations described hereinafter in which:
[0114]
[0115]
[0116]
[0117]
[0118]
[0119]
[0120]
[0121]
[0122]
[0123]
[0124]
[0125]
[0126] The figures are schematic and simplified for clarity, and they just show details which are essential to the understanding of the disclosure, while other details are left out. Throughout, the same reference signs are used for identical or corresponding parts.
[0127] Further scope of applicability of the present disclosure will become apparent from the detailed description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the disclosure, are given by way of illustration only. Other embodiments may become apparent to those skilled in the art from the following detailed description.
DETAILED DESCRIPTION OF EMBODIMENTS
[0128] The detailed description set forth below in connection with the appended drawings is intended as a description of various configurations. The detailed description includes specific details for the purpose of providing a thorough understanding of various concepts. However, it will be apparent to those skilled in the art that these concepts may be practiced without these specific details. Several aspects of the apparatus and methods are described by various blocks, functional units, modules, components, circuits, steps, processes, algorithms, etc. (collectively referred to as “elements”). Depending upon particular application, design constraints or other reasons, these elements may be implemented using electronic hardware, computer program, or any combination thereof.
[0129] The electronic hardware may include micro-electronic-mechanical systems (MEMS), integrated circuits (e.g. application specific), microprocessors, microcontrollers, digital signal processors (DSPs), field programmable gate arrays (FPGAs), programmable logic devices (PLDs), gated logic, discrete hardware circuits, printed circuit boards (PCB) (e.g. flexible PCBs), and other suitable hardware configured to perform the various functionality described throughout this disclosure, e.g. sensors, e.g. for sensing and/or registering physical properties of the environment, the device, the user, etc. Computer program shall be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.
[0130] The present application relates to the field of hearing systems, e.g. hearing aids or headsets. It relates in particular to a situation where a wearer of the hearing system receives acoustic as well as electromagnetically transmitted versions of sound from a sound environment around the wearer of the hearing system. Mutual timing of the arrival of the two representations matter (in particular if they differ in propagation/processing time). A too large difference in time of arrival (e.g. more than 10-20 ms) of the same sound ‘content’ of the two representations at the user's ear leads to confusion and disturbance, rather than improved perception (e.g. speech intelligibility) by the wearer.
[0131] Consider a situation, where a user is in a very noisy environment—the target talker is speaking, but the SNR at the user is low, because he or she is located at a distance from the target talker and the user cannot understand the target talker. However, a wireless microphone is located on the table, very close to the target talker and can pick up an essentially noise-free version of the target speech signal. Unfortunately, the essentially noise-free signal picked up by the wireless microphone is delayed by T.sub.D (e.g. 30 ms) relative to the direct sound (for example) when it arrives at the user (and thus the hearing system worn by the user), and it cannot be presented to the user for the reasons described above.
[0132] However, the hearing system may use the received (essentially clean, but delayed) signal to predict the clean signal T.sub.D (e.g. 30 ms) into the future—the prediction will obviously not be perfect, but parts of the predicted signal (in particular low-frequency parts) will be a good representation of the actual clean signal T.sub.D (e.g. 30 ms) in the future. This predicted part can be usefully presented to the user, either directly, or combined with the microphone signals of the hearing system.
[0133] Further, a gain varying in time and frequency may be extracted from the predicted signal and applied to the hearing aid microphone signals. The gain may e.g. be depending on the level of the signal of interest, such that only the time-frequency regions of the external signal with high amount of energy are preserved (and the low-energy (or low SNR) regions are attenuated).
[0134] More specifically, consider the situation depicted in
[0135] It should be noted that this is not always the case. It takes approximately 3 ms for sound to travel 1 meter. If e.g. the sound source is 5 meters away, the transmission delay through air is 15 ms, so in fact the wirelessly transmitted sound may arrive prior to the sound picked up by the microphones. It may thus be advantageous knowing the TDOA, and only apply prediction, when T>0 or T>5 ms or T>10 ms (where T=T.sub.1=T.sub.2, where T.sub.1 and T.sub.2 are the time of arrival at the hearing device of the wirelessly received signal and (‘corresponding’) the acoustically received signal, cf. below).
[0136] In the wireless microphone (M.sub.ex), the speech signal is encoded and transmitted (WTS) to the hearing aid user (U), where it is received T.sub.1 ms later (e.g. in one or both hearing aids (HD1, HD2) or a in separate processing device in communication with the hearing aid(s)). Meanwhile, the acoustic speech signal (ATS) emitted from the target talker (TT) is received at the microphones of the hearing aid user T.sub.2 ms later. Hence, the wirelessly transmitted signal (WTS) is delayed by T=T.sub.1−T.sub.2 ms compared to the acoustic signal (ATS) received at the hearing aid user (U). There may be differences between the time of arrival of the acoustic signals (equal to the interaural time difference, ITD) (and theoretically also between the time of arrival of the wirelessly transmitted signal) at the two hearing aids (HD1, HD2).
[0137] In practice, the time-difference-of-arrival (TDOA) T may be estimated by a similarity measurement between the relevant signals, e.g. by correlating the acoustic signal and the wirelessly received signal of a given hearing aid to determine the time difference T. Alternatively or additionally, the time-difference-of-arrival T may be estimated by other means, e.g. via ultra wide band (UWB) technology.
[0138] In a binaural hearing aid setup comprising left and right hearing instruments (cf. e.g. HD1, HD2 in
[0139] If T is too large, e.g. larger than 10 ms, the wirelessly received signal cannot be used for real-time presentation to the user. In that case, we propose to use signal samples that are available in the hearing aid system s(n−T−K+1), . . . , s(n−T) to predict future samples s(n) (relative to the signal available in the hearing aid system (HD1, HD2)), where K−1 is the number of past samples of the wirelessly received signal used for the prediction (s(n−T) is the last wirelessly received sample, so that K is the number of samples used for the prediction). The prediction may be performed in the time domain or in other domains, e.g. the (time-) frequency domain.
[0140]
[0141] Different use cases of the concept of the present disclosure may be envisioned, e.g. the following situations:
1) The wireless microphone is an external wireless microphone—e.g. a microphone unit clipped on to a target speaker, a table microphone, a microphone in a smart-phone, etc. (cf. e.g. M.sub.ex in
2) The “wireless microphone” is in fact the opposite hearing device (HDx, x=1 or 2 as the case may be), and the target signal is an external sound source: the sound signal is picked up by the left (HD2 for example) and sent to the right hearing device (HD1 for example) using signal prediction to reduce latency, so that the (parts of) the microphone signal from the left (HD2) may replace/or be combined with the right (HD1) microphone signals (see below for a more detailed description).
3) The “wireless microphone” is not a microphone. We consider the situation, where we would like to export computations to an external (processing) device, e.g. a smart-phone: A sound signal is picked up by the hearing device-users microphones, sent from the hearing device to an external device for computations (potentially using signal prediction to reduce latency), and sent back from the external device to the hearing device (potentially using signal prediction to reduce latency).
[0142] Prediction of future samples s(n) based on past samples s(n−T−K+1), . . . , s(n−T), as illustrated in
[0143] Prediction of future samples s(n) based on past samples s(n−T−K+1), . . . , s(n−T) is a well-known problem with well-known existing solutions. For example, prediction of future samples s(n) may be based on linear prediction, see e.g. [1], where an estimate z(n) of s(n) is formed as a linear combination of past samples, i.e.,
z(n)=Σ.sub.k=0.sup.P−1a.sub.ks(n−T−k), (1)
where a.sub.k, k=0, . . . , P−1 are time-varying, signal-dependent coefficients derived from past samples s(n−T−K+1), . . . , s(n−T) [1], and where P denotes the order of the linear predictor.
[0144] Many other ways of predicting s(n) exist. For example, an estimate z(n) of s(n) could be computed using non-linear methods, such as deep neural networks (DNNs), i.e.,
z(n)=G(s(n−T−K+1), . . . ,s(n−T);Θ,T), (2)
where Θ denotes the set of parameters of the DNN and G(., Θ, T) represents the network. In this situation, the network G(., Θ, T) would be trained off-line, before deployment to predict signal samples separated by T samples, using a training set of clean speech signals, cf. e.g. chapter 5 in [2].
[0145] More generally, an estimate z(n) of s(n) could be computed using a DNN of the form,
z(n)=G(s(n−T−K+1), . . . ,s(n−T),x(n);Θ), (3)
where x(n) represents a microphone signal captured at the hearing aid (i.e., a noisy version of what the network tries to predict). In this situation, we removed the network dependency on T, because it can be estimated internally in the DNN by comparing the wireless received samples s(n−T−K+1), . . . , s(n−T) with the local microphone signal x(n), for example by correlating the two sequences. This configuration which has access to an up-to-date, but potentially very noisy signal x(n) is particularly well-suited for prediction of transients/speech onsets, which may otherwise be challenging.
[0146] In yet another generalized version of the predictor,
z(n)=G(s(n−T−K+1), . . . ,s(n−T),x.sub.1(n), . . . ,x.sub.M(n);Θ), (4)
the estimate z(n) is a function of the (out-dated, but potentially relatively noise-free) received wireless signal and multiple local microphone signals x.sub.1(n), . . . , x.sub.M(n), (which are up-to-date, but potentially very noisy). This latter configuration has as a special case the situation, where z(n) is computed (partly) as a function of a beamformed signal y(n), computed at the hearing aid using the local microphone signals,
y(n)=H(x.sub.1(n), . . . ,x.sub.M(n)), (5)
where H(.) represents a beamforming operation. Yet other prediction methods exist.
[0147] Obviously, prediction is not limited to time domain signals s(n) as described above. For example, (linear) prediction could also take place in the time-frequency domain or in other domains (e.g. cosine, wavelet, Laplace, etc.).
[0148] In general, one can write the predicted signal as
z(n)=s(n)+e(n), (6)
where e(n) is the estimation error. If e(n) is considered as a noise term, the prediction process may be seen as simply ‘trading’ a delayed (outdated), essentially noise-free signal s(n−T) with an up-to-date, but generally noisy signal z(n)=s(n)+e(n). The more accurate prediction, the smaller the noise (prediction error). The signal-to-noise ratio (SNR) ξ(n) in the predicted signal may be estimated from the available past samples, for example as
{circumflex over (ξ)}(n)=Σ.sub.n′=n−T−K+1.sup.n−Ts.sup.2(n′)/Σ.sub.n′=n−T−K+1.sup.n−Te.sup.2(n′)|.sup.2, (7)
where the sum is taken over available past samples. Alternatively, the SNR may be computed offline as a long-term average SNR to be expected for a particular value of T.
[0149] The SNR may also be estimated in the time-frequency domain (m,q),
{circumflex over (ξ)}(m,q)=Σ.sub.m′=m−T′−K′+1.sup.m−T′|s(m′q)|.sup.2/Σ.sub.m′=m−T′−K′+1.sup.m−T′|e(m′,q)|.sup.2, (8)
where s(m, q) and e (m, q) denote time-frequency representations (for example, short-time Fourier transforms) of signals s(n) and e(n), and T′ and K′ time-frequency analogues of T and K, where m is a time index (e.g. a time-frame index) and q is a frequency index (e.g. a frequency band index).
[0150] The predicted signal z(n) may be used in several ways in the hearing device.
[0151] For example, if the SNR ξ(n) is sufficiently high, one could simply substitute the noisy signal x(n) picked up at a microphone of the hearing device with the predicted signal z(n), for example in signal regions where the SNR ξ(n) in the predicted signal would be higher than the SNR in the microphone signal x(n) (as estimated by an SNR estimation algorithm on board the hearing device).
[0152] Alternatively, rather than substituting signal samples z(n), one could perform the substitution in frequency bands. For example, one could decompose and substitute z(n) in frequency channels (e.g. low frequencies), for which it is known that the predicted signal is generally of better quality than the hearing aid microphone signal More generally, substitution could even be performed in the time-frequency domain according to an estimate of the SNR in time-frequency tiles, cf. eq. (8) above.
[0153] Alternatively, the signal z(n) may be combined with one or more of the microphone signals of the hearing device in various beamforming schemes in order to produce a final noise-reduced signal. In this situation, the signal z(n) is simply considered yet another microphone signal with a noisy realization of the target signal.
[0154] The description above assumed the predictor to be part of the receiver (i.e., the hearing system, e.g. a hearing device). However, it is also possible to do the prediction in the wireless microphone—assuming it has processing capabilities and can be informed of the time-difference-of-arrival T. In other words, the predicted signal z(n) is formed in the wireless microphone and transmitted to the hearing system (e.g. hearing device(s)), potentially together with side information such as the estimated SNR {circumflex over (ξ)}(n).
[0155] The description above has assumed the wireless microphone is a single microphone that captures the essentially noise-free signal s(n). However, it could also consist of a microphone array (i.e., more than one microphone). In this case, a beamforming system could be implemented in the wireless device, and the output of the beamformer play the role of the essentially noise-free signal s(n). Further, the external (sound capturing) device may e.g. be constituted by or comprise a table microphone array capable of extracting at least one noise free signal.
[0156]
[0157] In an analogue to digital (AD) process, a digital sample x(n) has a length in time of 1/f.sub.s e.g. 50 μs, for f.sub.s=20 kHz. A number of (audio) samples N.sub.s are e.g. arranged in a time frame, as schematically illustrated in the lower part of
[0158]
[0159] A time frame of an electric signal may e.g. comprise a number N.sub.s of consecutive samples, e.g. 64, (written as vector x.sub.m) of the digitized electric signal representing sound, m being a time index, cf. e.g.
[0160] The electric input signal(s) representing sound may be provided as a number of frequency sub-band signals. The frequency sub-bands signals may e.g. be provided by an analysis filter bank, e.g. based on a number of band-pass filters, or on a Fourier transform algorithm as indicated above (e.g. by consecutively extracting respective magnitude spectra from the Fourier transformed data).
[0161] As indicated in
[0162]
[0163]
[0164]
[0165]
[0166] The at least one earpiece (EP) comprises an input transducer (here a microphone (M) for converting sound in the environment of the hearing device to an acoustically received electric input signal y(n) representing the sound. The earpiece further comprises a wireless transmitter (Tx) for transmitting the acoustically received electric input signal y(n), or a part (e.g. a filtered part, e.g. a lowpass filtered part) thereof, to the audio processing device (APD). The earpiece (EP) further comprises a wireless receiver for receiving a predicted signal from said audio processing device, at least in a normal mode of operation of the hearing device. The wireless transmitter and receiver may be provided as antenna and transceiver circuitry for establishing an audio communication link (WL) according to a standardized of proprietary (short range) protocol. The earpiece (EP) further comprises an output transducer (here a loudspeaker (SPK)) for converting a (final) processed signal s′.sub.out(n) to stimuli perceived by the user as sound. The processed signal (s′.sub.out(n)) may, at least in a normal mode of operation of the hearing device, be constituted by or comprise at least a part of the predicted signal (provided by the audio processing device, (or by the earpiece as in
[0167] The audio processing device (APD) comprises a wireless receiver (Rx) for receiving the acoustically received electric input signal y(n), or a part thereof, from the earpiece (EP), and is configured to provide a received signal y(n′) representative thereof. The audio processing device (APD) (e.g. the computing device (CPD.sub.apd)) further comprises a processor part (HAP) for applying a processing algorithm (e.g. including a neural network) to said received signal (y(n′)), or to a signal originating therefrom, e.g. a transformed version thereof (Y), and to provide a modified signal (Y′). The processor part (HAP) may e.g. be configured to compensate for a hearing impairment of the user (e.g. by applying a compressive amplification algorithm, e.g. providing a frequency and/or level dependent gain (or attenuation) to be applied to the input signal (y(n′), or Y). The audio processing device (APD) (e.g. the computing device (CPD.sub.apd)) further comprises a signal predictor (PRED) for estimating future values of the modified signal (y′, Y′) in dependence of a multitude of past values of the signal, thereby providing a predicted signal (z, Z). The signal predictor (PRED) may comprise a prediction algorithm (either working in the time domain or in a transform domain, e.g. the time-frequency domain) configured to predict future values of an input signal based on past values of the input signal, and knowledge of a prediction time, e.g. a processing delay (cf. input T) between the first future value(s) and the latest past value(s) of the input signal (cf. e.g.
[0168] The signal predictor (PRED) is configured to fully or partially compensate for a processing delay incurred by a) the transmission of the acoustically received electric input signal (y(n)) from the earpiece (EP) to the audio processing device (APD), b) the processing in the audio processing device (APD) (through its audio signal processing path from receiver (Rx) to transmitter (Tx)), and c) the transmission of the predicted signal (z(n) or a processed version thereof to said earpiece (EP) and its reception therein (as signal s′.sub.out(n)). This is achieved by providing an estimate T of the total processing delay (T=T.sub.link+T.sub.apd) as input to the prediction algorithm (PRED).
[0169] In the embodiment of
[0170]
[0171]
[0172] Regarding the embodiments of
[0173] The term ‘or a processed version thereof’ may e.g. cover such extracted features from an original audio signal. The term ‘or a processed version thereof’ may e.g. also cover an original audio signal that has been subject to a processing algorithm that applies gain or attenuation to the original audio signal and this results in a modified audio signal (preferably enhanced in some sense, e.g. noise reduced relative to a target signal).
[0174] It is intended that the structural features of the devices described above, either in the detailed description and/or in the claims, may be combined with steps of the method, when appropriately substituted by a corresponding process.
[0175] As used, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well (i.e. to have the meaning “at least one”), unless expressly stated otherwise. It will be further understood that the terms “includes,” “comprises,” “including,” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will also be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element but an intervening element may also be present, unless expressly stated otherwise. Furthermore, “connected” or “coupled” as used herein may include wirelessly connected or coupled. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. The steps of any disclosed method is not limited to the exact order stated herein, unless expressly stated otherwise.
[0176] It should be appreciated that reference throughout this specification to “one embodiment” or “an embodiment” or “an aspect” or features included as “may” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the disclosure. Furthermore, the particular features, structures or characteristics may be combined as suitable in one or more embodiments of the disclosure. The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects.
[0177] The claims are not intended to be limited to the aspects shown herein but are to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more.
REFERENCES
[0178] [1] Deller, Hansen and Proakis, “Discrete-Time Processing of Speech Signals,” IEEE Press, 2000. [0179] [2] Goodfellow, Bengio and Courville, “Deep Learning,” MIT Press, 2016. [0180] EP3681175A1 (Oticon) 15 Jul. 2020 [0181] US20190378531A1 (Oticon) 12 Dec. 2019