SYSTEMS AND METHODS FOR EVALUATING HEARING HEALTH
20200315498 ยท 2020-10-08
Assignee
Inventors
Cpc classification
A61B5/7232
HUMAN NECESSITIES
G10L25/18
PHYSICS
H04B1/665
ELECTRICITY
A61B5/7278
HUMAN NECESSITIES
International classification
A61B5/00
HUMAN NECESSITIES
G10L25/18
PHYSICS
Abstract
Systems and methods are provided for evaluating hearing health of a given user. An input audio signal is transformed into the frequency domain and a first and second hearing profile are applied to the input audio sample. The first hearing profile represents a healthy hearing standard and the second hearing profile is the given user's hearing profile. Using the hearing profiles, first and second perceptually relevant information (PRI) values are generated for the input audio sample. The first and second PRI values are analyzed against each other to generate a PRI index value for the given user, where the PRI index value is a hearing health index value for the given user. The given user's hearing profile may additionally be applied to differently processed audio samples to evaluate the amount of perceptual rescue offered by various digital signal processing algorithms.
Claims
1. A method comprising: transforming an input audio sample into the frequency domain; applying a first hearing profile to the input audio sample to generate a first perceptually relevant information (PRI) value for the input audio sample; applying a second hearing profile to the input audio sample to generate a second perceptually relevant information value for the input audio sample; and analyzing the first and second PRI values to generate a PRI index value for a given user.
2. The method of claim 1, further comprising: processing the input audio sample using a parameterized processing function to generate a processed audio sample; applying the second hearing profile to the processed audio sample to calculate a third perceptually relevant information value for the processed audio sample; and analyzing the first, second, and third PRI values to determine a perceptual rescue effected by the parameterized processing function.
3. The method of claim 1, further comprising measuring the perceptual rescue of a plurality of parameterized processing functions by: separately processing the input audio sample using the plurality of parameterized processing functions to generate a plurality of processed audio samples; applying the second hearing profile to each processed audio sample to calculate a perceptually relevant information value for each parameterized processing function, wherein the second hearing profile is obtained from the given user and comprises masking threshold and hearing threshold information; and for each parameterized processing function, analyzing the calculated perceptually relevant information value to generate a corresponding PRI index value uniquely associated with the given user and the parameterized processing function.
4. The method of claim 1, wherein: the first hearing profile is obtained from a healthy human listener and the second hearing profile is obtained from the given user; and the first hearing profile and the second hearing profile comprise one or more of masking threshold information and hearing threshold information.
5. The method of claim 4, further comprising: determining one or more portions of the masking threshold information from masking threshold curves, the masking threshold curves corresponding to the first hearing profile and the second hearing profile; and determining one or more portions of the hearing threshold information from audiogram data.
6. The method of claim 4, wherein: the first PRI value represents a healthy standard PRI value; the second PRI value represents the given user's unique PRI value; and generating the PRI index value for the given user comprises calculating a ratio value based on the first PRI value and the second PRI value
7. The method of claim 1, wherein one or more of the first hearing profile and the second hearing profile are derived from at least one of a suprathreshold test, a psychophysical tuning curve, a masked threshold test, a threshold test and an audiogram.
8. The method of claim 1, in which the second hearing profile is the given user's hearing profile and is estimated from demographic information of the given user.
9. The method of claim 2, wherein: the perceptual rescue effected by the parameterized processing function is optimized by determining at least one processing parameter by a sequential determination of subsets of the at least one processing parameter, each subset determined so as to optimize the given user's perceptually relevant information for the input audio signal.
10. The method of claim 9, wherein: the parameterized processing function operates on one or more subband signals of the input audio signal; and the perceptual rescue effected by the parameterized processing function is optimized by: selecting a subset of the subbands so that a masking interaction between the selected subset of subbands is minimized; determining at least one first processing parameter for the selected subbands; and determining at least one second processing parameter for an unselected subband based on the first processing parameters of adjacent subbands.
11. The method of claim 10, wherein the at least one second processing parameter for an unselected subband is determined based on an interpolation of the first processing parameters of adjacent subbands.
12. The method of claim 10, wherein the at least one first processing parameter is determined sequentially on a subband by subband basis.
13. The method of claim 9, wherein the parameterized processing function operates on one or more subband signals of the input audio signal, the method further comprising: selecting a subset of adjacent subbands; tying the corresponding values of the at least one processing parameter for each subband of the selected subset of adjacent subbands; and performing a joint determination of the tied processing parameter values by minimizing the given user's perceptually relevant information for the selected subset of adjacent subbands.
14. The method of claim 13, further comprising: selecting a reduced subset of adjacent subbands from the selected subset of adjacent subbands; tying the corresponding values of the at least one processing parameter for each selected subband of the reduced subset of subbands; performing a joint determination of the tied parameter values by minimizing the given user's perceptually relevant information for the reduced subset of subbands; repeating the previous steps until a first single subband is selected as the reduced subset of adjacent subbands; and determining a first processing parameter of the first single subband.
15. The method of claim 14, further comprising: selecting a second subset of adjacent subbands; successively reducing the selected second subset of adjacent subbands until a second single subband is selected; determining a second processing parameter of the second single subband; and performing a joint processing of the first processing parameter of the first single subband and the second processing parameter of the second single subband.
16. The method of claim 15, wherein the joint processing of the first and second processing parameter comprises at least one of: joint optimization of the processing parameters for the derived single subbands; smoothing of the parameters for the derived single subbands; and applying constraints on the deviation of corresponding values of the processing parameters for the derived single subbands.
17. The method of claim 1 in which perceptually relevant information is calculated by calculating perceptual entropy.
18. An audio processing device comprising: at least one processor; and at least one memory storing instructions, which when executed causes the at least one processor to: transform an input audio sample into the frequency domain; apply a first hearing profile to the input audio sample to generate a first perceptually relevant information (PRI) value for the input audio sample; apply a second hearing profile to the input audio sample to generate a second perceptually relevant information value for the input audio sample; and analyze the first and second PRI values to generate a PRI index value for a given user.
19. The audio processing device of claim 18, wherein the instructions further cause the at least one processor to: process the input audio sample using a parameterized processing function to generate a processed audio sample; apply the second hearing profile to the processed audio sample to generate a third perceptually relevant information value for the processed audio sample; and analyze the first, second, and third PRI values to determine a perceptual rescue effected by the parameterized processing function.
20. The audio processing device of claim 18, wherein the instructions further cause the at least one processor to measure the perceptual rescue of a plurality of parameterized processing functions by: separately processing the input audio sample using the plurality of parameterized processing functions to generate a plurality of processed audio samples; applying the second hearing profile to each processed audio sample to calculate a perceptually relevant information value for each parameterized processing function, wherein the second hearing profile is obtained from the given user and comprises masking threshold and hearing threshold information; and for each parameterized processing function, analyzing the calculated perceptually relevant information value to generate a corresponding PRI index value uniquely associated with the given user and the parameterized processing function.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0042] In order to describe the manner in which the above-recited and other advantages and features of the disclosure can be obtained, a more particular description of the principles briefly described above will be rendered by reference to specific embodiments thereof, which are illustrated in the appended drawings. Understand that these drawings depict only example embodiments of the disclosure and are not therefore to be considered to be limiting of its scope, the principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings in which:
[0043]
[0044]
[0045]
[0046]
[0047]
[0048]
[0049]
[0050]
[0051]
[0052]
[0053]
[0054]
[0055]
[0056]
[0057]
[0058]
[0059]
[0060]
[0061]
DETAILED DESCRIPTION
[0062] Various example embodiments of the present disclosure are discussed in detail below. While specific implementations are discussed, it is appreciated that these implementations are described for illustration purposes only. One of ordinary skill in the art will recognize that other components and configurations may be used without parting from the spirit and scope of the disclosure.
[0063] The present disclosure relates to creating improved systems and methods for evaluating hearing health as well as creating means to compare the perceptual rescue effects of different DSP systems using perceptually relevant information (PRI) calculations. PRI is the information rate (bits/second) that can be transferred to a receiver for a given piece of audio content after factoring in what information will be lost because it is below the hearing threshold of the listener, or lost due to masking from other components of the signal within a given time frame. This is the result of a sequence of signal processing steps that are well defined for the ideal listener. In general terms, PRI is calculated from absolute thresholds of hearing (the minimum sound intensity at a particular frequency that a listener is able to detect) as well as the masking patterns for the particular listener.
[0064] Masking is a phenomenon that occurs across all sensory modalities where one stimulus component prevents detection of another. The effects of masking are present in the typical day-to-day hearing experience as individuals are rarely in a situation of complete silence with just a single pure tone occupying the sonic environment. To counter masking and allow the listener to perceive as much information within their surroundings as possible, the auditory system processes sound in way that provides a high bandwidth of information to the brain. The basilar membrane running along the center of the cochlea, which interfaces with the structures responsible for neural encoding of mechanical vibrations, is frequency selective. To this extent, the basilar membrane acts to spectrally decompose incoming sonic information whereby energy concentrated in different frequency regions is represented to the brain along different auditory fibers. The basilar membrane can be modelled as a filter bank with near logarithmic spacing of filter bands. This allows a listener to extract information from one frequency band, even if there is strong simultaneous energy occurring in a remote frequency region. For example, an individual will be able to hear both the low frequency rumble of a car approaching whilst listening to someone speak at a higher frequency. High energy maskers are required to mask signals when the masker and signal have different frequency content, but low intensity maskers can mask signals when their frequency content is similar.
[0065] The characteristics of auditory filters can be measured, for example, by playing a continuous tone at the center frequency of the filter of interest, and then measuring the masker intensity required to render the probe tone inaudible to a given listener as a function of relative frequency difference between masker and probe components. A psychophysical tuning curve (PTC), consisting of a frequency selectivity contour extracted via behavioral testing, provides useful data to determine a given individual's masking contours. In one embodiment of the test, a masking band of noise is gradually swept across frequency, from below the probe frequency to above the probe frequency. The given individual (i.e. the listener) then responds when he or she can hear the probe and stops responding when the probe can no longer be heard. This gives a jagged trace that can then be interpolated to estimate the underlying characteristics of the auditory filter of interest. It is appreciated that other methodologies known in the art may be employed to obtain user masking contour curves without departing from the scope of the present disclosure. For instance, an inverse paradigm may be used in which a probe tone is swept across frequency while a masking band of noise is fixed at a center frequency (known as a masked threshold test or MT test).
[0066] Patterns begin to emerge when testing listeners with different hearing capabilities using the PTC test. For example, as seen in
[0067]
[0068]
[0069]
[0070] PRI can be calculated according to a variety of methods, as would be appreciated by one of ordinary skill in the art. One such method, called perceptual entropy, was developed by James D. Johnston at Bell Labs, and generally comprises: transforming a sampled window of audio signal into the frequency domain, obtaining masking thresholds using psychoacoustic rules by performing critical band analysis, determining noise-like or tone-like regions of the audio signal, applying thresholding rules for the signal and then accounting for absolute hearing thresholds. Following this, the number of bits required to quantize the spectrum without introducing perceptible quantization error is determined. For instance, Painter & Spanias disclose the following formulation for perceptual entropy in units of bits/s, which is closely related to ISO/IEC MPEG-1 psychoacoustic model 2 [Painter & Spanias, Perceptual Coding of Digital Audio, Proc. Of IEEE, Vol. 88, No. 4 (2000); see also generally Moving Picture Expert Group standards https://mpeg.chiariglione.org/standards]:
Where:
[0071] i=index of critical band; [0072] bl.sub.i and bh.sub.i=upper and lower bounds of band i; [0073] k.sub.i=number of transform components in band i; [0074] T.sub.i=masking threshold in band i; [0075] nint =rounding to the nearest integer; [0076] Re()=real transform spectral components; and [0077] Im()=imaginary transform spectral components.
[0078]
[0079] The approach described above with respect to
[0080]
[0081] In other words, the multiband dynamics processor 702 is configured to process the audio sample 701 such that audio sample 701 has an increased PRI for the particular listener (i.e., the user), taking into account the individual listener's personal hearing profile 704. To this end, parameterization of the multiband dynamics processor 702 is adapted to increase the PRI of the processed audio sample over the unprocessed audio sample. As mentioned previously, the parameters of the multiband dynamics processor 702 are determined by an optimization process 711 that uses PRI as its optimization criterion. The above approach for processing an audio signal based on optimizing PRI and taking into account a listener's hearing characteristics may be based not only on multiband dynamic processors, but also on any kind of parameterized audio processing function that can be applied to the audio sample 701 and its parameters determined so as to optimize PRI of the audio sample. A [PRI.sub.optimized] value 712 is then outputted, which then can be analyzed against [PRI.sub.user] 713 and [PRI.sub.healthy] 714 to generate an improved PRI index value 715 for the user.
[0082] The parameters of the audio processing function may be determined for an entire audio file, for a corpus of audio files, or separately for portions of an audio file (e.g. for specific frames of the audio file). The audio file(s) may be analyzed before being processed, played or encoded. Processed and/or encoded audio files may be stored for later usage by the particular listener (e.g. in the listeners audio archive). For example, an audio file (or portions thereof) encoded based on the listener's hearing profile may be stored or transmitted to a far-end device such as an audio communication device (e.g. telephone handset) of the remote party. Alternatively or additionally, an audio file (or portions thereof) processed using a multiband dynamic processor that is parameterized according to the listener's hearing profile may be stored or transmitted.
[0083] Various optimization methods are possible to maximize the PRI of an audio sample, depending on the type of the applied audio processing function. For example, a subband dynamic compressor may be parameterized by compression threshold, attack time, gain and compression ratio for each subband, and these parameters may be determined by the optimization process. In some cases, the effect of the multiband dynamics processor on the audio signal is nonlinear and an appropriate optimization technique is required to account for the nonlinearity. The number of parameters that need to be determined may become large, e.g., if the audio signal is processed in many subbands and a plurality of parameters needs to be determined for each subband of the many subbands. In such cases, it may not be practicable to optimize all parameters simultaneously and a sequential approach to parameter optimization may be applied. Various approaches for sequential optimization are disclosed below. Although these sequential optimization procedures do not necessarily result in the optimum parameters, the obtained parameter values result in increased PRI over the unprocessed audio sample and thereby improve the user's listening experience.
[0084] A brute force approach to multi-dimensional optimization of processing parameters may be based on trial and error and successive refinement of a search grid. First, a broad search range is determined based on some a priori expectation on where an optimal solution might be located in the parameter space. Constraints on reasonable parameter values may be applied to limit the search range. Then, a search grid or lattice having a coarse step size is established in each dimension of the parameter space. One should note that the step size need not be constant but may, in some embodiments, differ across one or more of the processing parameters. For example, a compression threshold may be searched between 50 and 90 dB, in steps of 10 dB; simultaneously, a compression ratio between 0.1 and 0.9 may be searched in steps of 0.1. Thus, the search grid of this example has 59=45 points. PRI is determined for each parameter combination associated with a search point of the search grid, and the maximum PRI for the search grid is determined. The search may then be repeated in a next iteration, starting with the parameters that previously yielded the best (i.e., maximum) result of the prior iteration, and using a reduced range and step size. For example, a compression threshold of 70 dB and a compression rate of 0.4 were determined to have maximum PRI in the first search grid. Then, a new search range for thresholds between 60 dB and 80 dB and for ratios between 0.3 and 0.5 may be set for the next, second, iteration. The step sizes for the next optimization may be determined to 2 dB for the threshold and 0.05 for the ratio, and the combination of parameters having maximum PRI for the search grid of the second iteration determined. Further iterations may be performed for refinement. Other and additional parameters of the signal processing function may be considered as well. In the case of a multiband compressor, parameters for each subband must be determined. Simultaneously searching optimum parameters for a larger number of subbands may, however, take a long time or even become unfeasible. Thus, the present disclosure suggests various ways of structuring the optimization in a sequential manner to perform the parameter optimization in a shorter time without losing an unacceptable amount of precision in the search. The disclosed approaches are not limited to the above brute force search but may be applied to other optimization techniques as well.
[0085] As illustrated in
[0086] Another optimization approach is illustrated in
[0087] For example, in
[0088] In some embodiments, this optimization could be continued and taken a step further, as illustrated in
[0089]
[0090] In the following discussion of
[0091]
[0092]
[0093] While the masker signal tone 1405 is still sweeping upwards in frequency, the intensity 1402 of masker signal 1405 is then increased again, until the test subject no longer hears signal tone 1403. In this manner, the masker signal intensity 1402 oscillates around the hearing level 1401 (as indicated by the solid line) of the test subject with regard to the masker signal frequency and the signal tone. This hearing level 1401 is well established and well known for people having no hearing loss. Any deviations from this curve indicate a hearing loss (see for example
[0094] In
[0095]
[0096]
[0097]
[0098] In some embodiments computing system 1800 is a distributed system in which the functions described in this disclosure can be distributed within a datacenter, multiple datacenters, a peer network, etc. In some embodiments, one or more of the described system components represents many such components each performing some or all of the function for which the component is described. In some embodiments, the components can be physical or virtual devices.
[0099] Example system 1800 includes at least one processing unit (CPU or processor) 1810 and connection 1805 that couples various system components including system memory 1815, such as read only memory (ROM) 1820 and random access memory (RAM) 1825 to processor 1810. Computing system 1800 can include a cache of high-speed memory connected directly with, in close proximity to, or integrated as part of processor 1810.
[0100] Processor 1810 can include any general-purpose processor and a hardware service or software service, such as services 1832, 1834, and 1836 stored in storage device 1830, configured to control processor 1810 as well as a special-purpose processor where software instructions are incorporated into the actual processor design. Processor 1810 may essentially be a completely self-contained computing system, containing multiple cores or processors, a bus, memory controller, cache, etc. A multi-core processor may be symmetric or asymmetric.
[0101] To enable user interaction, computing system 1800 includes an input device 1845, which can represent any number of input mechanisms, such as a microphone for speech, a touch-sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech, etc. In some examples, the input device can also include audio signals, such as through an audio jack or the like. Computing system 1800 can also include output device 1835, which can be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input/output to communicate with computing system 1800. Computing system 1800 can include communications interface 1840, which can generally govern and manage the user input and system output. In some examples, communication interface 1840 can be configured to receive one or more audio signals via one or more networks (e.g., Bluetooth, Internet, etc.). There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
[0102] Storage device 1830 can be a non-volatile memory device and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs), read only memory (ROM), and/or some combination of these devices.
[0103] The storage device 1830 can include software services, servers, services, etc., that when the code that defines such software is executed by the processor 1810, it causes the system to perform a function. In some embodiments, a hardware service that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as processor 1810, connection 1805, output device 1835, etc., to carry out the function.
[0104] The presented technology offers a novel way of evaluating hearing health as well as an effective means of comparing the perceptual rescue offered by digital signal processing algorithms. It is to be understood that the present invention contemplates numerous variations, options, and alternatives. The present invention is not to be limited to the specific embodiments and examples set forth herein.
[0105] It should be further noted that the description and drawings merely illustrate the principles of the proposed device. Those skilled in the art will be able to implement various arrangements that, although not explicitly described or shown herein, embody the principles of the invention and are included within its spirit and scope. Furthermore, ail examples and embodiment outlined in the present document are principally intended expressly to be only for explanatory purposes to help the reader in understanding the principles of the proposed device. Furthermore, all statements herein providing principles, aspects, and embodiments of the invention, as well as specific examples thereof, are intended to encompass equivalents thereof.