SYSTEMS AND METHODS FOR ESTIMATING HEARING LOSS USING WIDEBAND ACOUSTIC IMMITTANCE

20230240562 · 2023-08-03

    Inventors

    Cpc classification

    International classification

    Abstract

    Estimating conductive hearing loss from wideband acoustic immittance can be accomplished by use of an analog-electric model of an ear canal and inner ear utilized to convert acoustic measurements into output data. The output data can be used to create conductive hearing loss methods for diagnostically estimating hearing loss based upon the acoustic measurements. The model includes a number of inputs from the acoustic measurements that are fit and converted to the output data that can be compared to measured data for hearing loss to train a system to quickly and easily diagnose an estimated hearing loss, such as via a diagnostic tool. The system can also be trained to identify an ear type based upon the transformed acoustic measurement data, and the type of ear can be used to provide additional hearing loss estimation.

    Claims

    1. A method of estimating hearing loss, comprising: obtaining an acoustic measurement from an ear canal; modeling the acoustic measurement with an electric-analog model to obtain a model output; transforming the model output to a measured absorbance to determine an averaged absorbance over a frequency range converted to decibels; and training a machine learning network, wherein the training comprises: acquiring measured hearing loss data; fitting the parameters of the model such that the transformed model output correlates to the measured hearing data; and identifying one or more classifiers of the transformed model output that provides an estimate of the hearing loss.

    2. The method of claim 1, wherein the acoustic measurement comprises an impedance-based measurement.

    3. The method of claim 2, wherein the impedance based measurement comprises a wideband acoustic immittance.

    4. The method of claim 1, wherein the step of modeling the acoustic measurement further comprises a transmission line representing the ear canal terminated by the network.

    5. The method of claim 4, wherein the network of the model comprises three parallel branches, with each branch comprising a stiffness, damping, and mass component, wherein the network represents mechanics of a tympanic membrane coupled to ossicles of an ear.

    6. The method of claim 1, wherein the averaged absorbance over a frequency range is compared to a pure tone average to obtain the estimated hearing loss.

    7. A method of estimating hearing loss, comprising: obtaining an acoustic measurement from an ear canal; modeling the acoustic measurement with an electric-analog model to obtain model outputs including absorbance levels and ossicular loss levels in decibels; using a machine-learned network to identify ear type, wherein the machine-learned network comprises: comparing ossicular loss level outputs to a pure tone average to identify classifiers; and separating the classifiers into at least first and second ear types; based upon the identified ear type, transforming the model outputs to estimate the hearing loss.

    8. The method of claim 7, wherein the classifiers comprise: a. a first identifier wherein the ossicular loss level outputs that are lower than the pure tone average; or b. a second identifier wherein the ossicular loss level outputs that are larger than the pure tone average by a factor of at least two.

    9. The method of claim 8, wherein the step of transforming the model outputs comprises combining the absorbance levels and ossicular loss levels to obtain an estimated hearing loss for the first identifiers.

    10. The method of claim 9, wherein the step of transforming the model outputs comprises dividing the ossicular loss levels by two to obtain an estimated hearing loss for the second identifiers.

    11. The method of claim 10, wherein the estimated hearing loss is determined by comparing the results to a known pure tone average.

    12. The method of claim 7, wherein the acoustic measurement comprises an impedance-based measurement.

    13. The method of claim 12, wherein the impedance based measurement comprises a wideband acoustic immittance.

    14. The method of claim 7, wherein the step of modeling the acoustic measurement further comprises a transmission line representing the ear canal terminated by the network.

    15. The method of claim 14, wherein the network of the model comprises three parallel branches, with each branch comprising a stiffness, damping, and mass component, wherein the network represents mechanics of a tympanic membrane coupled to ossicles of an ear.

    16. A system for estimating hearing loss, comprising: a device for obtaining an acoustic measurement from an ear canal; the device including a computer readable medium configured to: obtain the acoustic measurement from the ear canal; model the acoustic measurement with an electric-analog model to obtain model outputs and training a machine-learned network, wherein the training comprises: identifying values of the model output that correlate with measured or assessed data; based upon machine-learned training, identify a classifier indicating an ear type, said machine-learned training including the steps of comparing ossicular loss level outputs to a pure tone average; and based upon the identified ear type, transform the model outputs to estimate the hearing loss.

    17. The system of claim 16, wherein the classifiers comprise: a. a first identifier wherein the ossicular loss level outputs that are lower than the pure tone average; or b. a second identifier wherein the ossicular loss level outputs that are larger than the pure tone average by a factor of at least two.

    18. The system of claim 16, wherein the electric-analog model comprises a transmission line representing the ear canal terminated by the network.

    19. The system of claim 18, wherein the network of the model comprises three parallel branches, with each branch comprising a stiffness, damping, and mass component, wherein the network represents mechanics of a tympanic membrane coupled to ossicles of an ear.

    20. The system of claim 19, wherein the transmission line of the model represents an ear canal comprising a plurality of concatenated truncated-cone sections.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0053] Several embodiments in which the invention can be practiced are illustrated and described in detail, wherein like reference characters represent like components throughout the several views. The drawings are presented for exemplary purposes and may not be to scale unless otherwise indicated.

    [0054] FIG. 1 is a graph showing mean absorbance at tympanometric peak pressure (TPP) across the four effusion volume groups: healthy normal control, OME full, OME partial, and OME clear.

    [0055] FIG. 2 is a graph showing average audiometric thresholds grouped by volume of effusion in children with otitis media.

    [0056] FIG. 3 is a schematic of an analog circuit model of ear canal acoustics and middle ear mechanics.

    [0057] FIGS. 4A-4D are graphs showing measured and modeled mean absorbance based on effusion volume groups.

    [0058] FIGS. 4E-4H are graphs illustrating differences between measured and modeled data for each effusion volume group.

    [0059] FIG. 5A is a graph showing the effect of effusion volume on middle-ear admittance for each of the four effusion volume groups.

    [0060] FIG. 5B is a graph showing the influence on the phase by the groups of effusion volume.

    [0061] FIGS. 6A-6F show effect of effusion volume on the admittance of each branch of the middle-ear model.

    [0062] FIG. 7 is a graph showing absorbance levels of the effusion volume groups as a function of pure tone average.

    [0063] FIG. 8 is a graph showing ossicular loss levels of the effusion volume groups as a function of pure tone average.

    [0064] FIG. 9 is a graph showing the predicted hearing loss as a value of decibels as a function of pure tone average based upon aspects of the present disclosure.

    [0065] FIG. 10 shows graphs including umbo admittance modeled as the sum of two branch admittances, Y1+Y3.

    [0066] FIG. 11 shows graphs including umbo velocity.

    [0067] FIG. 12 is a schematic diagram showing steps of the development of a middle-ear conductive hearing loss algorithm.

    [0068] FIG. 13 is a schematic diagram showing steps of the application of a system or method for middle-ear conductive hearing loss estimation.

    [0069] FIG. 14 is a schematic diagram showing steps of the development of a middle-ear fluid volume determining algorithm.

    [0070] FIG. 15 is a schematic diagram showing the application of a system or method for diagnosing/estimating the amount of middle-ear fluid volume.

    [0071] An artisan of ordinary skill need not view, within isolated figure(s), the near infinite number of distinct permutations of features described in the following detailed description to facilitate an understanding of the invention.

    DETAILED DESCRIPTION OF THE INVENTION

    [0072] The present disclosure is not to be limited to that described herein. Mechanical, electrical, chemical, procedural, and/or other changes can be made without departing from the spirit and scope of the invention. No features shown or described are essential to permit basic operation of the invention unless otherwise indicated.

    [0073] Unless defined otherwise, all technical and scientific terms used above have the same meaning as commonly understood by one of ordinary skill in the art to which embodiments of the invention pertain.

    [0074] The terms “a,” “an,” and “the” include both singular and plural referents.

    [0075] The term “or” is synonymous with “and/or” and means any one member or combination of members of a particular list.

    [0076] The terms “invention” or “present invention” are not intended to refer to any single embodiment of the particular invention but encompass all possible embodiments as described in the specification and the claims.

    [0077] The term “about” as used herein refer to slight variations in numerical quantities with respect to any quantifiable variable. Inadvertent error can occur, for example, through use of typical measuring techniques or equipment or from differences in the manufacture, source, or purity of components.

    [0078] The term “substantially” refers to a great or significant extent. “Substantially” can thus refer to a plurality, majority, and/or a supermajority of said quantifiable variable, given proper context.

    [0079] The term “generally” encompasses both “about” and “substantially.”

    [0080] The term “configured” describes structure capable of performing a task or adopting a particular configuration. The term “configured” can be used interchangeably with other similar phrases, such as constructed, arranged, adapted, manufactured, and the like.

    [0081] Terms characterizing sequential order, a position, and/or an orientation are not limiting and are only referenced according to the views presented.

    [0082] The “scope” of the invention is defined by the appended claims, along with the full scope of equivalents to which such claims are entitled. The scope of the invention is further qualified as including any possible modification to any of the aspects and/or embodiments disclosed herein which would result in other embodiments, combinations, subcombinations, or the like that would be obvious to those skilled in the art.

    [0083] The term “pure tone average” (PTA) refers to the average of hearing threshold levels at a set of specified frequencies: 500, 1000, 2000 and 4000 Hz. This value gives a snapshot of an individual's hearing level in each ear. As speech sounds are more densely represented in the mid frequencies, the outlying frequencies are not included in the PTA calculation to allow for more consistent comparisons. If your PTA is <25 dB, your overall hearing would be considered to be within normal limits. With a PTA of 95 dB, your hearing would be considered in the profound range.

    [0084] Wideband Acoustic Immittance.

    [0085] As described herein, multiple transfer functions or measurements or can be calculated from each WAI response. Reflectance Γ(f) is defined as the complex ratio of reverse pressure to forward pressure. In a uniform tube, reflectance is related to the input impedance at the point of measurement by the equation:

    [00001] Γ ( f ) = Z ec ( f ) - Z 0 Z ec ( f ) + Z 0 . ( 1 )

    [0086] where f is frequency, Z.sub.ec(f) is the frequency-dependent impedance looking into the ear canal, and Z.sub.0 is the characteristic acoustic impedance of the ear canal at the point of measurement. The characteristic impedance is calculated as:


    Z.sub.0=ρc/A.sub.ec  (2),

    [0087] where ρ is the density of air, c is the speed of sound in air, and A.sub.ec is the cross-sectional area of the ear-canal at the measurement point. The cross-sectional area used in this study was 44.18 mm2 for all ears, which corresponds to a diameter of 7.5 mm.

    [0088] Absorbance is the recommended WAI measure form for clinical assessment. Absorbance custom-character is defined as:


    custom-character(f)=1−|Γ(f)|.sup.2  (3)

    [0089] Absorbance measurements averaged across each of the four OME groups (full, partial, clear, and normal) are shown in FIG. 1.

    [0090] To stabilize the fitting of model parameters, noisy data below 0.2 kHz was replaced by extrapolation of measured ear-canal impedance. This smoothing procedure consisted of extending the real part of the ear-canal impedance at 0.2 kHz with the same value to all frequencies below 0.2 kHz and extending the imaginary part with its magnitude inversely proportional to frequency, which would be consistent with it being due to a compliance.

    [0091] Ear-Canal and Middle-Ear Model

    [0092] The analog circuit model that is used to represent ear-canal acoustic responses is shown in FIG. 3. This model comprises a nonuniform transmission line that is terminated by a network of stiffness (K), damping (R), and mass (M) components. The transmission line represents the ear canal, while the terminating circuit represents the middle ear. The nonuniform transmission line is modeled as six concatenated truncated-cone sections. The transmission line is implemented numerically as a lossless two-port transmission matrix.

    [0093] As shown in FIG. 3, the ear-canal is modeled as a non-uniform transmission line comprising six concatenated segments. The middle-ear input impedance Z.sub.me is modeled as three parallel branches that each contain stiffness K, damping R, and mass M components.

    [0094] The middle-ear network contains three branches that together represent mechanics of the tympanic membrane coupled to the ossicles (i.e., malleus, incus, and stapes). Middle-ear input impedance Z.sub.me is calculated as the parallel impedance of these three branches.

    [00002] Z me = 1 1 / Z 1 + 1 / Z 2 + 1 / Z 3 . ( 4 )

    [0095] In this equation, Z.sub.1=jωM.sub.1+R.sub.1+K.sub.1/jω, Z.sub.2=jωM.sub.2+R.sub.2+K.sub.2/jω, and Z.sub.3=jωM.sub.3+R.sub.3+K.sub.3/jω, where ω=2πf, and f represents frequency. The specific association of middle-ear components in the model with middle-ear anatomy is not required for estimation of CHL.

    [0096] The model ear-canal input impedance Z.sub.ec is related to its middle-ear impedance Z.sub.me by elements of the ear-canal transmission matrix:

    [00003] Z ec = 𝒜 ec Z me - ec - 𝒞 ec Z me + 𝒟 ec . ( 5 )

    [0097] Model parameters were selected by a “simplex” search algorithm that minimizes a cost function that is calculated as a weighted sum of deviations between the model and measurements of ear-canal impedance and absorbance. Search constraints, which comprised mostly lower and upper bounds on the model parameters, were implemented by adding a constraint penalty to the cost function.

    [0098] Middle-Ear Admittance

    [0099] An alternative to Eq. (4), is to calculate Z.sub.me from ear-canal impedance Z.sub.ec by, in effect, inverting Eq. (5):

    [00004] Z me = 𝒟 ec Z ec + ec 𝒞 ec Z ec + 𝒜 ec . ( 6 )

    [0100] The matrix elements in Eq. (6) are the same as those in Eq. (5) because the transmission line is a reciprocal network, so the transmission matrix has a determinant that is equal to one, custom-character.sub.eccustom-character.sub.ec−custom-character.sub.eccustom-character.sub.ec=1. When Z.sub.ec is measured instead of calculated from the model, then Z.sub.me becomes an estimate of middle-ear impedance. Admittance is the reciprocal of impedance, so Y.sub.me=1/Z.sub.me. This notation may be extended as Y.sub.me,m and Y.sub.me,e to distinguish between the modeled and estimated versions of middle-ear admittance, respectively.

    [0101] Conductive Hearing Loss Estimate

    [0102] Two different approaches to modeling CHL will be considered. The first approach models CHL as being due to incomplete sound absorption (i.e., reflection) at the tympanic membrane and is calculated as the average absorbance custom-character over the frequency range from f.sub.1=2 kHz to f 2=4 kHz converted to decibels:

    [00005] 𝔸 _ = 1 f 2 - f 1 f 1 f 2 𝔸 ( f ) df , ( 7 ) CHL 1 = - 10 .Math. log 10 𝔸 _ . ( 8 )

    [0103] In Eq. (8), the minus sign is needed to make CHL.sub.1 positive because custom-character is always less than one.

    [0104] The second approach models CHL as being due to a pressure drop across the ossicular chain, which is presumed to be in series with a cochlear load, and is calculated as the middle-ear model damping component R.sub.1 divided by the cochlear load R.sub.0=2×10.sup.7 (mks acoustic ohm) converted to decibels:


    CHL.sub.2=20.Math.log.sub.10(R.sub.1/R.sub.0).  (9)

    estimating CHL will be derived below taking advantages of features that were observed when Eqs. (8-9) were applied to the OME data.

    [0105] Diagnostic Prediction Performance.

    [0106] The four OME conditions (full, partial, clear, normal) may be predicted from acoustic measurements by utilizing binomial logistic regression (mnrfit, MATLAB) and hierarchical classification. A binomial logistic regression (often referred to simply as logistic regression) predicts the probability that an observation falls into one of two categories of a dichotomous dependent variable based on one or more independent variables that can be either continuous or categorical. A hierarchical classifier is a classifier that maps input data into defined subsumptive output categories. The classification occurs first on a low-level with highly specific pieces of input data. The classifications of the individual pieces of data are then combined systematically and classified on a higher level iteratively until one output is produced. This final output is the overall classification of the data. Depending on application-specific details, this output can be one of a set of pre-defined outputs, one of a set of on-line learned outputs, or even a new novel classification that hasn't been seen before. Generally, such systems rely on relatively simple individual units of the hierarchy that have only one universal function to do the classification. In a sense, these machines rely on the power of the hierarchical structure itself instead of the computational abilities of the individual components.

    [0107] Prior to the regression, a principal component analysis reduced the decision variable, which was either absorbance or the magnitude of a middle-ear admittance, to fewer components. The hierarchical classification procedure first predicted whether effusion was present or absent. Then wet ears were classified as being either full or partial and dry ears as being either clear or normal. Test performance was quantified for each of the binomial categories by calculating the area under a receiver operating characteristic curve (AUC).

    [0108] A fourth regression analysis was performed to predict whether CHL.sub.2 over-estimates PTA by a factor of two, as will be understood as disclosed herein.

    [0109] Ear-Canal and Middle-Ear Model.

    [0110] Model results were obtained by fitting model parameter values to measured ear-canal impedance Z.sub.ec and absorbance. Separate model results were obtained at two different static pressures: 0 daPa (0 Pa) and tympanic peak pressure (TPP).

    [0111] In the upper panels of FIG. 4, the model absorbances averaged across each group (red dashed) are superimposed over the average measured absorbance at TPP (blue solid). Individual differences between measured and modeled absorbance are shown in the lower panels for each individual ear included in that group, while the mean of those differences shown in black. On average, the agreement between the model and the measurements at TPP is excellent. Individual differences are mainly due to artifacts and “fine structure” in the measured absorbance.

    [0112] The WAI measurement at TPP (i.e., absorbance) was selected for FIG. 4 to facilitate comparison with a similar figure known. Such comparison demonstrates that the current version of the middle-ear model fits the OME data at least as well as a previous version of the model that did not include M.sub.3. The graphs of FIG. 4A-4H show results measured at 0 Pa because CHL predictions were better. The top panels, FIGS. 4A-4D, show measured (blue solid) and modeled (red dashed) mean absorbance. The bottom panels, FIGS. 4E-4H show individual differences between measured and modeled data for each ear included in the mean above. The gray lines represent each individual difference, while the black line represents the mean difference. The columns represent the four OME conditions: normal, clear, partial, and full. For future reference, the nine middle-ear model parameters averaged across each OME group at 0 Pa are listed in Table I.

    [0113] Table I shows middle-ear model parameters at 0 Pa averaged across each OME condition. For convenience, units of damping (R) are cgs acoustic ohm (dyn.Math.s/cm.sup.5). Conversion to mks acoustic ohm (Pa.Math.s/m.sup.3) requires multiplying tabled values by 10.sup.5. Units of mass (M) and stiffness (K) are multiplied and divided by 1 kHz, respectively.

    TABLE-US-00001 TABLE I Middle-ear model parameters at 0 Pa averaged across each OME condition. OME condition M.sub.1 R.sub.1 K.sub.1 M.sub.2 R.sub.2 K.sub.2 M.sub.3 R.sub.3 K.sub.3 full 114 10792 21066 69 4274 171064 464 4324 14655 partial 90 5041 16720 68 2510 66403 389 2738 11209 clear 61 1319 11153 45 674 33447 361 3866 12194 normal 65 686 11377 34 873 25390 295 2832 9094

    [0114] The effect of effusion volume on middle-ear admittance at 0 Pa is shown in FIG. 5 by superimposing the average admittance for each of the four OME conditions. At frequencies below 0.8 kHz, where the admittance is dominated by stiffness, the partial and full conditions show an apparent increase in stiffness. Above 0.8 kHz, the main effect on admittance of increasing effusion volume is a magnitude decrease (upper panel). Concurrent flattening of the admittance phase (lower panel) above 1 kHz suggests that the magnitude decrease is likely due to increased damping. As shown, FIG. 5 shows the influence of effusion volume on middle-ear admittance at 0 Pa. These admittances were calculated from the parameter values listed in Table I.

    [0115] The effect of effusion volume on the admittance of each branch of the middle-ear model is shown in FIG. 6, which shows the effect of effusion volume on model-branch admittance at 0 daPa. These admittances were calculated from the parameter values listed in Table I. Because middle-ear admittance is calculated as the sum of the three branch admittances (Y.sub.me=Y.sub.1+Y.sub.2+Y.sub.3), FIG. 6 demonstrates a decomposition of tympanic motion into three distinct vibration modes. The main effect of effusion volume on Y.sub.1 is an increase in damping. The main effect of Y.sub.2 is an increase in both stiffness and damping. The effect on Y.sub.3 is relatively small compared to the effects on Y.sub.1 and Y.sub.2.

    [0116] Conductive Hearing Loss Estimate.

    [0117] Individual absorbance level CHL.sub.1 (left panel) and ossicular loss (right panel) CHL.sub.2 are plotted in FIGS. 7 and 8 as a function of pure tone average (PTA) for each of the 34 ears with audiograms. Both CHL.sub.1 and CHL.sub.2 are well-correlated with PTA (87% and 81% respectively). However, note that the range of absorbance loss is less than 10 dB, which is much less than the nearly 50 dB range of PTA. Although CHL.sub.2 spans a larger (40 dB) range, it also falls short of explaining the entire PTA range. As included in FIGS. 7 and 8, the correlation between each CHL estimate and PTA is labeled as ρ.

    [0118] Note that CHL.sub.2 (FIG. 8) appears to form two clusters that are separated by the two reference lines, which represent CHL slopes relative to PTA of one and two. The CHL estimates in the lower cluster (23/34) are consistently lower than PTA by as much as 10 dB. The CHL estimates in the upper cluster (11/34) are larger by than PTA a factor of two. This observation suggests a scheme for improving the CHL estimate. For ears in the lower cluster, CHL.sub.1 (absorbance loss) will be added to CHL.sub.2 (ossicular loss). For ears in the upper cluster, CHL.sub.2 will be divided by two. Cluster assignment depends on whether CHL.sub.2 is closer to PTA or PTA times 2.

    [00006] CHL 3 = { CHL 1 + CHL 2 , if lower cluster CHL 2 / 2 , if upper cluster . ( 10 )

    [0119] The results of applying Eq. (10) to the OME data are shown in FIG. 9, which shows a prediction of conductive hearing loss CHL.sub.3 as a function of PTA. The is labeled as ρ and the estimation error is calculated as the mean absolute difference in dB. The correlation of CHL.sub.3 with PTA is 95% and the prediction error, quantified as the mean absolute difference, is 3.2 dB. It is noted that 3.3 dB is less than the +/−5 dB test retest reliability of pure tone behavioral audiometry. If we look at these results at TPP instead of 0 Pa (not shown), the correlation of CHL.sub.3 with PTA is reduced to 88% and the prediction error increases to 5.3 dB, which demonstrates the superiority of measurements at 0 Pa for our CHL estimation methods.

    [0120] Although Eq. (10) is empirically motivated, its application requires prior knowledge of whether an ear belongs to the lower or upper cluster. One way to address this requirement in practice would be to use a logistic regression to predict the appropriate cluster assignment from the WAI measurement, such as described in herein.

    [0121] Diagnostic Prediction Performance.

    [0122] Diagnostic performance is quantified as the validation AUC (area under the curve) for each binomial hierarchical decision. In addition to the three hierarchical binomial decisions required for classification of OME group (effusion present or absent, effusion present ears as full/partial, effusion absent ears as clear/normal), a fourth binomial decision (ear type) is required to predict (at least without any a priori knowledge) whether the CHL2 cluster is on the lower or upper cluster.

    [0123] For convenience, the present disclosure will refer to the lower and upper clusters by the designations LWR and UPR, respectively. To focus the training portion of the regression on these two clearly separated clusters, only 13 ears (8 LWR and 5 UPR ears), specifically those with a PTA in the 5 to 25 dB HL range, were included in the LWR/UPR regression analysis. Ears with a PTA below 5 dB were not included as there is no clear separation in that region of LRW and UPR ears. Ears with a PTA above 359 25 dB were not included as they are all LWR ears.

    [0124] Diagnostic performance was optimized by selection of (1) a decision variable (custom-character, |Y.sub.me,m∥Y.sub.1|, |Y.sub.2|, |Y.sub.3|, |Y.sub.1+Y.sub.2|, |Y.sub.1+Y.sub.3|, |Y.sub.2+Y.sub.3|), and (2) a static pressure (0 Pa or TPP) to achieve the largest possible AUC. For all regressions, the number of principal components was 3 and the regularization parameter (to avoid overfitting) was λ=0.5. The four rows in Table II indicate the optimal combination of decision variable, static pressure, and number of principal components for each of the four binomial-decision types.

    [0125] Table II below provides diagnostic prediction performance. The bold numbers indicate the maximum validation AUC for each of the four binomial-decision types.

    TABLE-US-00002 TABLE II Diagnostic prediction performance. Decision Static Wet/ Full/ Clear/ LWR/ Variable Pressure Dry Partial Normal UPR custom-character TPP 0.997 0.981 0.692 0.809 |Y.sub.1 + Y.sub.2| 0 Pa 0.988 0.999 0.720 0.669 |Y.sub.1 + Y.sub.3| TPP 0.984 0.869 0.822 0.638 |Y.sub.2 + Y.sub.3| TPP 0.917 0.875 0.650 0.992

    [0126] As disclosed herein, aspects and/or embodiments of the disclosure include the use of modeling, such as machine learning to perform steps and/or identify classifiers for the various steps of the processes provided. Examples of the processes for the machine learning and algorithms used for the aspects are shown generally in FIGS. 12-15.

    [0127] For example, in FIG. 14 a first box shown includes the acquisition of ear-canal acoustic impedance measurements, such as in the form of the WAI using a tool. As shown by the arrow, the next step involves a middle-ear feature extraction. This can be in the form of the electric-analog model, which is shown in FIG. 3. As noted herein, the model is used to transform the data from the acoustic measurements to usable data, such as data involving absorbance, power reflectance, admittance, and impedance, among others. The model is used to fit the data to mimic the ear canal and middle-ear mechanics, which will provide information akin to that of the actual measurements of an ear-canal and middle-ear.

    [0128] A middle-ear fluid-volume (MEFV) assessment is done by measuring the fluid volume of a number of patients to acquire data on the patients, such as types of data that could be used to identify the amount of fluid volume. Model results were obtained by fitting model parameter values to measured ear-canal impedance and absorbance. As has been disclosed herein, the analog model and fitted parameter values that were used provides excellent overlap and agreement between the modeled values and the measured values.

    [0129] Next steps include the estimation of effusion volume in an ear. This is done, for example, with machine learning of the assessed data and the acoustic data, wherein a regression model is utilized to train a network to determine classifiers. In the instant case, this may be in the form of admittance, which will identify classifiers that identify the listed ear types, including full effusion, partial effusion, ears clear of effusion, and normal ears. The logistic regression analysis will be used to identify middle-ear fluid-volume regression multipliers, which is an estimation of the fluid volume in an ear. To do so, the fitted parameters of the analog model are compared to assessed measurements to identify which parameters are key for determining/estimating MEFV. This results in regression multipliers, which are used as a basis for the estimation of MEFV.

    [0130] In FIG. 15, a middle-ear fluid-volume estimation is provided that can be used in a clinical setting. Once the MEFV regression multipliers have been identified using the machine learning and logistic regression analysis shown in FIG. 14, said multipliers could be used in a diagnostic tool to quickly and accurately determine and estimate an amount of middle-ear fluid volume (MEFV). As shown in FIG. 15, the ear-canal acoustic impedance measurement is again used to acquire acoustic data. The middle-ear model feature extraction includes the electric-analog model to fit the acoustic data (i.e., the WAI data) to an actual ear-canal and middle-ear mechanics. The MEFV regression multipliers that have been trained are applied to the feature extraction information that has been modeled, and a score computation is developed. The machine learned classifiers are applied to the fit data from the WAI instrument, and a MEFV prediction is provided. Therefore, the modeling of the ear-canal and middle-ear mechanics can transform the acquired acoustic data, such as the WAI data, to be useful for clinical settings. The WAI tool can include the machine learned neural network that has been trained using the analog model of the ear and can provide quick and accurate predictions as to the middle-ear fluid-volume amounts.

    [0131] This information will allow for a quicker and more objective source of information (i.e., the presence and amount of fluid in the middle ear) that can be used by clinicians to assess a patient's issues and to better diagnose the potential issues. In addition, as the WAI data acquisition is generally considered non-invasive (e.g., a probe tip is quickly inserted into an ear canal), the ability to assess and diagnose patients of all ages and conditions can be done in a quick, efficient, and accurate manner which improves over the common manner today, which is perform a number of invasive and/or highly subjective tests to verify an amount of fluid or other issues in the ear.

    [0132] Additional algorithms are shown schematically in FIGS. 12 and 13. For example, as shown in FIG. 12, a middle-ear conductive loss (MECL) algorithm is developed. As shown schematically, this includes first acquiring ear-canal acoustic impedance measurements, such as in the form of the WAI using a tool. Again, the data is transformed and fit to an ear-canal and middle-ear model, such as using the analogy model and steps related to FIG. 3 herein. This puts the steps at the middle-ear model feature extraction step.

    [0133] The MECL assessment is done by measuring the hearing loss of a number of patients to acquire data on the patients, such as types of data that could be used to identify and quantify the hearing loss. Model results were obtained by fitting model parameter values to measured ear-canal impedance and absorbance. As has been disclosed herein, the analog model and fitted parameter values that were used provides excellent overlap and agreement between the modeled values and the measured values.

    [0134] Again, a logistic regression analysis (e.g., a machine learning network) can be used to identify MECL regression multipliers that correlate the middle-ear model feature extraction data, which includes the fitted parameters of the analog model, with the assessed/measured data from patients to identify regression multipliers, also known as classifiers. In the present disclosure, the analysis was done two ways. The first approach (CHL.sub.1) models CHL as being due to incomplete sound absorption (i.e., reflection) at the tympanic membrane and is calculated as the average absorbance from 2-4 kHz converted to decibels (dB). The second approach (CHL.sub.2) models CHL as being due to a pressure drop across the ossicular chain, which is presumed to be in series with a cochlear load and is calculated as the middle-ear model damping component R1 divided by the cochlear load R.sub.0=2×10.sup.7 (mks acoustic ohm) converted to dB.

    [0135] However, as identified herein, a third approach included the identification of the classification of ear types, identified as the lower cluster (LWR) in FIG. 8, and those identified as the upper cluster (UPR) in FIG. 8. Therefore, the algorithm of FIG. 12 may be used to train a machine learning system to identify the classification of ear types, which are LWR and UPR.

    [0136] FIG. 13 then shows a clinical application of the estimation of MECL at a clinic or in real life application. As shown in the figure, a WAI tool is used to acquire acoustic data, such as ear-canal acoustic impedance measurements. The analog model is used to fit the parameters of the ear-canal and middle-ear mechanics.

    [0137] The regression multipliers (also referred to as classifiers) that were learned in FIG. 12, are applied to the data from the WAI tool. This includes the ossicular damping, which was identified as CH L.sub.2 in the disclosure. The use of the ear types is applied to the data as provided for CHL.sub.3, wherein the data is either added (CHL.sub.1+CHL.sub.2) when the ear types are in the LWR cluster, or wherein the data is divided by two (CHL.sub.2/2) when the ear types are in the UPR cluster. The resulting data of CHL.sub.3 can then be plotted as a function of PTA to provide a quick and accurate estimation of the conductive hearing loss of the patient.

    [0138] Subjects with OME and PTA measurements in both ears almost always had either two LWR ears or two UPR ears. The only exception had partial fluid in the UPR ear and full fluid in the LWR ear, which might suggest that the full-fluid condition interferes with UPR status. Indeed, none of the ears classified as being full had UPR status. The strong within-subject correlation of UPR status suggests that this feature may have a biological origin.

    [0139] Ear-Canal and Middle-Ear Model.

    [0140] As has been noted herein, a key aspect of the modeling approach includes the (1) addition of a third mass component M.sub.3 to the middle-ear model from a previous version and (2) adjustment of the model parameter-search constraints (MPSC) to reduce the CHL estimation error. An important benefit of adding M.sub.3 was to remove the influence of OME away from the third branch of the middle-ear model and toward the first branch, which caused R.sub.1 to become more sensitive to effusion volume, thereby making R.sub.1 a better predictor of CHL. The MPSC, which comprise mainly lower and upper bounds on middle-ear model parameters, influence the parameter-search toward more desirable model fits.

    [0141] Although anatomical interpretation of middle-ear model components was not required for obtaining either CHL estimates or effusion-volume predictions, such interpretations are desirable for understanding underlying mechanisms. The strong correlation between the CHL estimates and the PTA supports the association of first branch with motion of the ossicular chain. On the other hand, the addition of M.sub.3 causes Y.sub.1 to no longer provide a good representation of umbo velocity (as it had in a previous model version) primarily because its resonance no longer shifts to lower frequencies with increasing effusion volume. In the present model, the trends observed in umbo velocity with increasing fluid volume are now better represented by Y.sub.um=Y.sub.1+Y.sub.3. Comparison with measurements (see FIGS. 10 and 11) indicate that Y.sub.um at 0 Pa (FIG. 10) simulates umbo velocity (FIG. 11, curves numbered 0-4) better than either Y.sub.1 alone (see FIGS. 6A-6F) and better than Y.sub.me=Y.sub.1+Y.sub.2+Y.sub.3 (see FIGS. 5A-5B). This observation suggests that the mode of TM motion represented by the third branch of the middle-ear model likely contributes to umbo velocity. In contrast, the second branch likely does not contribute to umbo velocity. So, the second branch apparently represents a mode of TM motion that does not contribute to motion of the ossicular chain.

    [0142] The fact that the third branch was not included in an initial prediction of CHL makes its contribution to motion of the stapes unclear. Further insights into the association of model components with middle-ear anatomy may be acquired by considering the influence of static pressure on admittance.

    [0143] The parameter-search methods currently used to fit the model to the measurements have been improved by selection of initial values from a set of values that are known to span the region of the parameter space of greatest interest. This method could be adapted to provide model fits to individual WAI measurements in a clinical setting in less than 10 seconds.

    [0144] Conductive hearing loss estimate.

    [0145] Having a non-invasive and objective acoustic estimate of CHL in children with OME is clinically useful because obtaining pure tone thresholds and a PTA is difficult in young children and infants, where OME is most common.

    [0146] The current approach disclosed herein to combining objective and non-invasive WAI measures with an electrical analog model of outer and middle ear mechanics results in CHL estimates within 3.2 dB of PTA. This is significant given that +/−5 dB is considered the test-retest reliability of behavioral audiometry. In addition, this method could be easily implemented as an algorithm in software that could provide a CHL estimate in seconds after a WAI measurement was made, demonstrating clinical viability of this approach. It should be noted that the low prediction error in the present analysis may be partially due to the fact that adjustment of parameter-search constraints included consideration of CHL estimation.

    [0147] Diagnostic Prediction Performance.

    [0148] The diagnostic prediction performance for the wet/dry decision (see, e.g., Table II, first row) is better than previously known. This improvement is due to the inclusion of regularization in the logistic-regression method, which reduces overfitting the training data.

    [0149] The full/partial and clear/normal predictions have been improved by a combination of factors, including improvements to the middle-ear model and consideration of different static pressures. Current predictions utilizing the model and other aspects of the present disclosure are good enough to be clinically useful, which is an improvement over the need to have more invasive testing, which is difficult with certain age demographics. The machine-learning techniques disclosed herein will continue to improve the system and methods disclosed, as will testing on an extended bandwidth (e.g., from 8 kHz to 11 kHz).

    [0150] Although the binomial classifier that has been developed for detection of ears in the UPR cluster has been based on a limited number of ears, it performed well on the entire set of ears with PTA by correctly identifying eight of the 11 ears in the UPR cluster. The three ears that were not detected by this classifier did not contribute much to the average prediction error because they all had PTA<5. Thus, the average CHL prediction error when using acoustic classification of UPR status was 3.7 dB, which is only 0.5 dB greater than the ideal error of 3.2 dB. The success of the UPR classifier on the training data suggests that predictions errors less than 4 dB may be achievable in a clinical setting. This is less than what test-retest reliability for the gold standard hearing assessment. behavioral pure tone audiometry.

    [0151] The success of the CHL and effusion-volume estimation methods described herein are due in part by the nine-parameter model of middle-ear mechanics that decomposes WAI measurements into three modes of motion. The first mode appears to be strongly correlated to PTA, which supports the anatomical interpretation that this mode involves the ossicular chain and contributes to stapes motion. Comparisons with temporal-bone measurements suggest that the first and third modes both contribute to umbo motion while the second mode is likely to represent TM motion that is independent of umbo motion.

    [0152] As noted herein, aspects of the disclosure include the development of systems and processes for the estimation of MECL and/or MEFV, such as in clinical applications. It is envisioned that the systems and/or processes, including the algorithms and machine learning networks, could be implemented with devices, such as WAI devices that are used in clinical settings. For example, the WAI tool could be used to non-invasively acquire the data from a patient in the clinic or other setting, and the tool or an associated network (e.g., a computer, server, processor, or the like) could quickly receive the acoustic data and provide the output information, which can include MECL and/or MEFV.

    [0153] In addition, it should be appreciated that the aspects of the disclosure, including the development of the analog model, can be used for additional estimations of ear conditions and/or pathologies. For example, upon determination and fitting of parameters of ear mechanics via the analog model, said information could be correlated and compared with measured and assessed data from patients that include any number and/or type of ear condition, issue, and/or pathology. The machine learning network could be utilized to train the network to identify classifiers associated with any number of ear conditions, issues, and/or pathologies that could be identified using the WAI data to quickly and accurately provide results in a clinical setting that would negate the need for invasive testing on a patient.

    [0154] In other words, the modeling and learning of the features of the present disclosure provide a near infinite basis for being able to assess, estimate, diagnose, provide treatment, or otherwise provide useful information related to ear conditions, issues, and/or pathologies for any number of patients, including those that may not be ideal for invasive and/or extensive testing, which is currently required. The results of the modeling and machine learning could be implemented into tools or otherwise associated with the tools, such as by receiving the data from the tools in a wired or wireless manner, to quickly provide the results and estimations for the conditions, features, issues, and/or pathologies of the patient's ear.

    [0155] As will be understood, aspects and/or embodiments disclosed herein will utilize processors, memory, instructions, and the like, and will include a machine learning model or models to identify classifiers of aspects of ear conditions and/or pathologies. Machine learning (ML) is the study of computer algorithms that can improve automatically through experience and by the use of data. It is seen as a part of artificial intelligence. Machine learning algorithms build a model based on sample data, known as “training data”, in order to make predictions or decisions without being explicitly programmed to do so.

    [0156] While it is envisioned that generally any type of ML (e.g., supervised learning, unsupervised learning, semi-supervised learning, or reinforcement learning) can be utilized by any of the aspects and/or embodiments of the present disclosure utilize supervised learning. Supervised learning (SL) is the machine learning task of learning a function that maps an input to an output based on example input-output pairs. It infers a function from labeled training data consisting of a set of training examples. In supervised learning, each example is a pair consisting of an input object (typically a vector) and a desired output value (also called the supervisory signal). A supervised learning algorithm analyzes the training data and produces an inferred function, which can be used for mapping new examples. An optimal scenario will allow for the algorithm to correctly determine the class labels for unseen instances. This requires the learning algorithm to generalize from the training data to unseen situations in a “reasonable” way (see inductive bias). This statistical quality of an algorithm is measured through the so-called generalization error.

    [0157] To solve a given problem of supervised learning, one has to perform the following steps: (1) Determine the type of training examples. Before doing anything else, the user should decide what kind of data is to be used as a training set. (2) Gather a training set. The training set needs to be representative of the real-world use of the function. Thus, a set of input objects is gathered, and corresponding outputs are also gathered, either from human experts or from measurements. (3) Determine the input feature representation of the learned function. The accuracy of the learned function depends strongly on how the input object is represented. Typically, the input object is transformed into a feature vector, which contains a number of features that are descriptive of the object. The number of features should not be too large, because of the curse of dimensionality; but should contain enough information to accurately predict the output. (4) Determine the structure of the learned function and corresponding learning algorithm. For example, the engineer may choose to use support-vector machines, regression analysis, or decision trees. (5) Complete the design. Run the learning algorithm on the gathered training set. Some supervised learning algorithms require the user to determine certain control parameters. These parameters may be adjusted by optimizing performance on a subset (called a validation set) of the training set, or via cross-validation. (6) Evaluate the accuracy of the learned function. After parameter adjustment and learning, the performance of the resulting function should be measured on a test set that is separate from the training set.

    [0158] As will be understood, while generally any type of SL can be utilized, the example provided herein utilized three different classification algorithms to train the model, namely the support vector machine (SVM), k-Nearest Neighbors (k-NN), and classification ensemble (ENS).

    [0159] Support-vector machines (SVMs, also support-vector networks) are supervised learning models with associated learning algorithms that analyze data for classification and regression analysis. Given a set of training examples, each marked as belonging to one of two categories, an SVM training algorithm builds a model that assigns new examples to one category or the other, making it a non-probabilistic binary linear classifier (although methods such as Platt scaling exist to use SVM in a probabilistic classification setting). SVM maps training examples to points in space so as to maximize the width of the gap between the two categories. New examples are then mapped into that same space and predicted to belong to a category based on which side of the gap they fall.

    [0160] The k-nearest neighbors algorithm (k-NN) is a non-parametric classification method. k-NN is a type of classification where the function is only approximated locally, and all computation is deferred until function evaluation. Since this algorithm relies on distance for classification, if the features represent different physical units or come in vastly different scales then normalizing the training data can improve its accuracy dramatically.

    [0161] Classification ensemble may also be referred to as ensemble learning. Ensemble methods use multiple learning algorithms to obtain better predictive performance than could be obtained from any of the constituent learning algorithms alone. Unlike a statistical ensemble in statistical mechanics, which is usually infinite, a machine learning ensemble consists of only a concrete finite set of alternative models, but typically allows for much more flexible structure to exist among those alternatives.

    [0162] The trained model and the associated machine learning and application of the model will utilize processors, modules, memories, databases, networks, and potentially user interfaces to show the results and allow changes to be made.

    [0163] In communications and computing, a computer readable medium is a medium capable of storing data in a format readable by a mechanical device. The term “non-transitory” is used herein to refer to computer readable media (“CRM”) that store data for short periods or in the presence of power such as a memory device. It is envisioned that the WAI tools and/or associated processors for evaluating the data from the WAI tools could implement CRM.

    [0164] One or more embodiments described herein can be implemented using programmatic modules, engines, or components. A programmatic module, engine, or component can include a program, a sub-routine, a portion of a program, or a software component or a hardware component capable of performing one or more stated tasks or functions. A module or component can exist on a hardware component independently of other modules or components. Alternatively, a module or component can be a shared element or process of other modules, programs, or machines.

    [0165] The system will preferably include an intelligent control (i.e., a controller) and components for establishing communications. Examples of such a controller may be processing units alone or other subcomponents of computing devices. The controller can also include other components and can be implemented partially or entirely on a semiconductor (e.g., a field-programmable gate array (“FPGA”)) chip, such as a chip developed through a register transfer level (“RTL”) design process.

    [0166] A processing unit, also called a processor, is an electronic circuit which performs operations on some external data source, usually memory or some other data stream. Non-limiting examples of processors include a microprocessor, a microcontroller, an arithmetic logic unit (“ALU”), and most notably, a central processing unit (“CPU”). A CPU, also called a central processor or main processor, is the electronic circuitry within a computer that carries out the instructions of a computer program by performing the basic arithmetic, logic, controlling, and input/output (“I/O”) operations specified by the instructions. Processing units are common in tablets, telephones, handheld devices, laptops, user displays, smart devices (TV, speaker, watch, etc.), and other computing devices.

    [0167] As noted, the WAI tool itself could include a processor, and/or the tool could be connected, either wired or wirelessly to a separate system encompassing the processing unit.

    [0168] The memory includes, in some embodiments, a program storage area and/or data storage area. The memory can comprise read-only memory (“ROM”, an example of non-volatile memory, meaning it does not lose data when it is not connected to a power source) or random access memory (“RAM”, an example of volatile memory, meaning it will lose its data when not connected to a power source). Examples of volatile memory include static RAM (“SRAM”), dynamic RAM (“DRAM”), synchronous DRAM (“SDRAM”), etc. Examples of non-volatile memory include electrically erasable programmable read only memory (“EEPROM”), flash memory, hard disks, SD cards, etc. In some embodiments, the processing unit, such as a processor, a microprocessor, or a microcontroller, is connected to the memory and executes software instructions that are capable of being stored in a RAM of the memory (e.g., during execution), a ROM of the memory (e.g., on a generally permanent basis), or another non-transitory computer readable medium such as another memory or a disc.

    [0169] In the instant case, the memory could include the machine learned classifiers and analog models, so as to fit the parameters of the model and to quickly and accurately identify the results based on the trained classifiers.

    [0170] Generally, the non-transitory computer readable medium operates under control of an operating system stored in the memory. The non-transitory computer readable medium implements a compiler which allows a software application written in a programming language such as COBOL, C++, FORTRAN, or any other known programming language to be translated into code readable by the central processing unit. After completion, the central processing unit accesses and manipulates data stored in the memory of the non-transitory computer readable medium using the relationships and logic dictated by the software application and generated using the compiler.

    [0171] In one embodiment, the software application and the compiler are tangibly embodied in the computer-readable medium. When the instructions are read and executed by the non-transitory computer readable medium, the non-transitory computer readable medium performs the steps necessary to implement and/or use the present invention. A software application, operating instructions, and/or firmware (semi-permanent software programmed into read-only memory) may also be tangibly embodied in the memory and/or data communication devices, thereby making the software application a product or article of manufacture according to the present invention.

    [0172] The database is a structured set of data typically held in a computer. The database, as well as data and information contained therein, need not reside in a single physical or electronic location. For example, the database may reside, at least in part, on a local storage device, in an external hard drive, on a database server connected to a network, on a cloud-based storage system, in a distributed ledger (such as those commonly used with blockchain technology), or the like.

    [0173] It is envisioned that the machine learned model and any of the training of the same could include cloud computing. Cloud computing is a model of service delivery for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, network bandwidth, servers, processing, memory, storage, applications, virtual machines, and services) that can be rapidly provisioned and released with minimal management effort or interaction with a provider of the service.

    [0174] A cloud computing environment is service oriented with a focus on statelessness, low coupling, modularity, and semantic interoperability. At the heart of cloud computing is an infrastructure comprising a network of interconnected nodes.

    [0175] As noted, the training model could be implemented on a user interface. The interface could also be a point on introduction of data, such as training data or test data to compare to the trained model for analysis. The results of the comparison could then be shown on a user interface.

    [0176] A user interface is how the user interacts with a machine. The user interface can be a digital interface, a command-line interface, a graphical user interface (“GUI”), oral interface, virtual reality interface, or any other way a user can interact with a machine (user-machine interface). For example, the user interface (“UI”) can include a combination of digital and analog input and/or output devices or any other type of UI input/output device required to achieve a desired level of control and monitoring for a device. Examples of input and/or output devices include computer mice, keyboards, touchscreens, knobs, dials, switches, buttons, speakers, microphones, LIDAR, RADAR, etc. Input(s) received from the UI can then be sent to a microcontroller to control operational aspects of a device.

    [0177] The user interface module can include a display, which can act as an input and/or output device. More particularly, the display can be a liquid crystal display (“LCD”), a light-emitting diode (“LED”) display, an organic LED (“OLED”) display, an electroluminescent display (“ELD”), a surface-conduction electron emitter display (“SED”), a field-emission display (“FED”), a thin-film transistor (“TFT”) LCD, a bistable cholesteric reflective display (i.e., e-paper), etc. The user interface also can be configured with a microcontroller to display conditions or data associated with the main device in real-time or substantially real-time.

    [0178] Any components of the system could be connected via network or other communication protocol to transfer information, communicate with other systems, or provide other connectivity. In some embodiments, the network is, by way of example only, a wide area network (“WAN”) such as a TCP/IP based network or a cellular network, a local area network (“LAN”), a neighborhood area network (“NAN”), a home area network (“HAN”), or a personal area network (“PAN”) employing any of a variety of communication protocols, such as Wi-Fi, Bluetooth, ZigBee, near field communication (“NFC”), etc., although other types of networks are possible and are contemplated herein. The network typically allows communication between the communications module and the central location during moments of low-quality connections. Communications through the network can be protected using one or more encryption techniques, such as those techniques provided by the Advanced Encryption Standard (AES), which superseded the Data Encryption Standard (DES), the IEEE 802.1 standard for port-based network security, pre-shared key, Extensible Authentication Protocol (“EAP”), Wired Equivalent Privacy (“WEP”), Temporal Key Integrity Protocol (“TKIP”), Wi-Fi Protected Access (“WPA”), and the like.

    [0179] Therefore, it has been shown that the model results provide a good description of the measured data. The presence of OME causes an increase in stiffness below 0.8 kHz and an increase in damping above 1 kHz, with larger changes resulting from larger volume effusions.

    [0180] In addition, WAI, combined with the model provided herein, can estimate behavioral audiometric thresholds or CHL within a clinically meaningful margin of error. This is a significant finding given the challenges associated with behavioral audiometric testing in pediatric populations where OME is most common.

    [0181] From the foregoing, it can be seen that the invention accomplishes at least all of the stated objectives.