SPEECH DISCRIMINATION TEST SYSTEM AND DEVICE
20210353182 · 2021-11-18
Inventors
- Harvey A. Dillon (Chatswood, AU)
- Teresa Yuk Ching Ching (Macquarie University, NSW, AU)
- Sanna Yuk Lan Hou (Macquarie University, NSW, AU)
- Nicola Raewyn Chong-White (Macquarie University, NSW, AU)
Cpc classification
A61B5/0002
HUMAN NECESSITIES
G16H50/20
PHYSICS
G16H80/00
PHYSICS
International classification
A61B5/00
HUMAN NECESSITIES
Abstract
The invention relates to a test to measure the ability of a subject to discriminate between speech sounds. The speech sounds may be selected from the worlds most widely spoken languages to enable the test to be carried out irrespective of the language spoken by the subject. Speech sounds may be presented in sequences, such as triplets. The subject would then be required to detect which speech sound in each sequence is different from the others. The test may be provided on a computer or similar device, in an embodiment a tablet computer, enabling the subject to conduct a self-assessment. The test may be performed on hearing aid users to determine the likelihood of a hearing aid user obtaining better test results after receiving one or two cochlear implants.
Claims
1. A method of testing speech comprehension of a subject, the method comprising: storing an inventory of speech sounds on a speech discrimination testing system, the speech discrimination testing system comprising a data processor, a memory in data communication with the data processor, and at least one transducer for presentation of speech sounds to a subject; selecting speech sounds from the inventory of speech sounds to be presented to the subject in a sequence of speech sounds; presenting the sequence of speech sounds to the subject, wherein each speech sound in the sequence of speech sounds is presented to the subject more than once, apart from a lone speech sound which is presented to the subject only once; and requesting the subject identify which speech sound in the sequence of speech sounds was the lone speech sound; and receiving, at the speech discrimination testing system, the subject's identification of which of the presented speech sounds was the lone speech sound.
2. (canceled)
3. The method according to claim 1, wherein the speech discrimination testing system further comprises a display for presenting visual images to the subject, and the method further comprises: presenting one or more visual images to the subject in association with each speech sound in the sequence of speech sounds; and requesting the subject to identify which speech sound in the sequence of speech sounds was the lone speech sound, by identifying the visual image associated with the lone speech sound.
4. The method according to claim 3, wherein the step of presenting visual images to the subject comprises presenting a sequence of visual images synchronized with the presentation of the speech sounds in the sequence of speech sounds such that each speech sound in the sequence of speech sounds is associated with a presented visual image.
5. (canceled)
6. (canceled)
7. (canceled)
8. The method according to claim 1, wherein speech sounds stored in the inventory of speech sounds are used in a majority of the most commonly spoken languages.
9. The method according to claim 1, wherein speech sounds stored in the inventory of speech sounds and selected for presentation to a follow a vowel-consonant-vowel format, a consonant-vowel-consonant format, a consonant-vowel format, or a vowel-consonant format.
10. The method according claim 9, wherein vowels used in the speech sounds are selected from a group consisting of: [a], [i] and [o]; and the consonants used in the speech sounds are selected from a group consisting of: [j], [k], [l], [m], [n], [p], [t] and [s].
11. The method according to claim 9, wherein the speech sounds presented in a sequence vary from one another by substitution of either one vowel, or one consonant.
12. The method according to claim 11, in which speech sounds are presented within a sequence as a consonant pair.
13. The method according to claim 1, wherein more than one sequence of speech sounds is presented such that the subject is required to identify lone speech sounds within each presented sequence.
14. The method according to claim 1, further comprising emitting noise via the at least one transducer while the speech sounds are presented, to provide a signal-to noise-ratio such that the subject is required to discriminate between presented speech sounds while the noise is emitted.
15. The method according to claim 14, wherein a signal-to-noise ratio at which speech sounds are presented against emitted noise is adjusted during the test.
16. The method according to claim 15, wherein the signal-to-noise ratio is adjusted from sequence to sequence to account for inherent difficulty in discriminating speech sounds presented in each sequence.
17. The method according to claim 16, wherein the signal-to-noise ratio is adjusted from sequence to sequence to ensure that each sequence provides substantially the same likelihood of identifying a correct lone speech sound.
18. (canceled)
19. The method according to claim 15, wherein the signal-to-noise ratio is adjusted based on prior identifications received from the subject, to identify a signal-to-noise ratio at which the subject may correctly discriminate speech sounds at a pre-determined ratio of correct responses to incorrect responses.
20. (canceled)
21. (canceled)
22. (canceled)
23. (canceled)
24. A speech discrimination testing system comprising: a data processor; a memory in data communication with the data processor configured to store an inventory of speech sounds for presentation to a subject; and at least one transducer for presentation of speech sounds to a subject wherein the speech discrimination testing system is configured to: select speech sounds from the stored inventory of speech sounds to be presented to the subject in a sequence; present selected speech sounds via the at least one transducer to the subject in a sequence such that each speech sound in the sequence is presented to the subject more than once, apart from one lone speech sound which is presented to the subject only once; and receive the subject's identification of the lone speech sound.
25. (canceled)
26. The speech discrimination testing system according to claim 24, further comprising a display for presenting visual images to the subject, wherein the speech discrimination testing system is further configured to present visual images to the subject in association with presented speech sounds; and enable the subject to identify which speech sound in the sequence of speech sounds was the lone speech sound, by identifying the visual image associated with that speech sound.
27. The speech discrimination testing system according to claim 26, wherein the speech discrimination testing system is configured to present visual images to the subject in a sequence of visual images synchronized with the presentation of speech sounds such that each presented speech sound is associated with a presented visual image.
28. (canceled)
29. (canceled)
30. (canceled)
31. The speech discrimination testing system according to claim 24, wherein speech sounds of the inventory of speech sounds are used in a majority of the most commonly spoken languages.
32. The speech discrimination testing system according to of claim 24, wherein speech sounds of the inventory of speech sounds to a follow a vowel-consonant-vowel format, a consonant-vowel-consonant format, a consonant-vowel format, or a vowel-consonant format.
33. The speech discrimination testing system according claim 32, wherein vowels used in the speech sounds are selected from a group consisting of: [a], [i] and [o]; and the consonants used in the speech sounds are selected from a group consisting of: [j], [k], [l], [m], [n], [p], [t] and [s].
34. The speech discrimination testing system according to claim 32, wherein the speech discrimination testing system is configured to present speech sounds in a sequence so as to vary from one another by substitution of either one vowel, or one consonant.
35. The speech discrimination testing system according to claim 34, wherein the speech discrimination testing system is configured to present speech sounds within a sequence as a consonant pair.
36. The speech discrimination testing system according to claim 24, wherein the speech discrimination testing system is configured to present more than one sequence of speech sounds such that the subject is required to identify the lone speech sound within each presented sequence.
37. The speech discrimination testing system according to of claim 24, wherein the speech discrimination testing system is configured to emit noise via the at least one transducer while the speech sounds are presented, to provide a signal-to noise-ratio such that the subject is required to discriminate between presented speech sounds while the noise is emitted.
38. The speech discrimination testing system according to claim 37, wherein the speech discrimination testing system is configured to adjust a signal-to-noise ratio at which speech sounds are presented against emitted noise while speech sounds are presented to the subject.
39. The speech discrimination testing system according to claim 37, wherein the speech discrimination testing system is configured to adjust the signal-to-noise ratio from sequence to sequence to account for inherent difficulty in discriminating speech sounds presented in each sequence.
40. The speech discrimination testing system according to claim 39, wherein speech discrimination testing system is configured to adjust the signal-to-noise ratio such that all sequences presented have approximately the same likelihood of being correctly discriminated.
41. The speech discrimination testing system according to claim 37, wherein the speech discrimination testing system is configured to adjust the signal-to-noise ratio based on responses received from the subject, to identify a signal-to-noise ratio at which the subject may correctly discriminate speech sounds at a pre-determined ratio of correct responses to incorrect responses.
42. (canceled)
43. The speech discrimination testing system according to claim 24, wherein the speech discrimination testing system is configured to estimate whether, or a likelihood that, the subject's identification of the lone speech sound would be improved through cochlear implantation.
44. The speech discrimination testing system according to claim 43 wherein the speech discrimination testing system is configured to estimate whether, or the likelihood that, the subject's identification of the lone speech sound would be improved through cochlear implantation taking into account how long the subject has experienced hearing loss.
45. (canceled)
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0084] It will be beneficial to further describe the invention with respect to the accompanying drawings, which demonstrate a preferred embodiment of method and device for testing the ability to discriminate speech sounds. Other embodiments of the invention are possible, and consequently, the particularity of the accompanying drawings is not to be understood as superseding the generality of the preceding description of the invention.
[0085]
[0086]
[0087]
[0088]
[0089]
[0090]
[0091]
[0092]
[0093]
[0094]
[0095]
[0096]
[0097]
[0098]
[0099]
[0100]
[0101]
[0102]
DETAILED DESCRIPTION OF THE INVENTION
[0103] In an embodiment the inventors have provided a test based on the ability of a subject to discriminate between speech sounds, rather than on a subject's ability to identify a speech sound, thereby allowing for a test method and device that may be effectively applied across language barriers. Such a test may never require the subject to identify a speech sound (such as by spelling it out or identifying it as a written word or a word of a language), other than to discriminate which sound is different to another that has been presented.
[0104] In an embodiment the invention is based on measuring the ability of a subject to correctly discriminate the odd speech sound present in sequence of other speech sounds. For example, pairs of speech sounds may be presented in sequential triplets, in which a first speech sound is presented once and a second speech sound is presented twice (such that the first speech sound is the ‘odd one out’). For example, if two speech sounds presented are [aja] and [ata], the subject may be presented with a sequential triplet of: [aja], [aja], [ata]. The subject would then be required to specify which of the three speech sounds was different to the other two. In this way, the subject is not required to identify either sound but is only required to recognise that one is different to the other. In other embodiments the speech sounds need not be presented as triplets, but rather as quadruplets, quintuplets, etc., and more than two different speech sounds may be presented.
Speech Sounds for Presentation to a Subject
[0105] In embodiments of the invention non-word speech sounds may be used as stimuli for a speech discrimination test so as to minimise the undue influence of language and subject learning on the ability of a subject to characterise a sound, and otherwise maximise the language neutrality of the test. This is because, if the presented speech sounds are commonly words of a particular language, then speakers of that language would have the advantage of being familiar with, and hence better able to discriminate those words over other words, phrases or speech sounds that they are less familiar with.
[0106] Nonsense speech sounds of various forms may be used according to embodiments of the invention. Nonsense speech sounds may in an embodiment take a consonant-vowel-consonant [CVC] format or a vowel-consonant-vowel [VCV] format. In an embodiment vowel-consonant-vowel [VCV] sounds may be used to maximise the availability of acoustic cues for discrimination by the subject, through perception of the medial consonant.
[0107] As part of their development of the invention, the inventors conducted a literature review on phonemic inventories occurring in the most widely spoken languages, being Mandarin, English, Spanish, Arabic, Hindi-urdu, Bengali, Russian, Portuguese, Japanese, German, Javanese, Korean, French, Turkish, Vietnamese, Telugu, Cantonese, Italian, Polish, Ukranian, Thai, Gujarati, Malay, Malayalam, Tamil, Marathi, Burmese, Romanian, Pashto, Dutch, Finnish, Greek, Indonesian, Norwegian, Hebrew, Croatian, Danish, Hungarian, Swedish, and Serbian. This was followed by a literature review on confusion matrices in consonant identification tasks, in noise and in quiet, for adult listeners: having typical hearing, using hearing aids, and using cochlear implants. To demonstrate the process undertaken by the inventors, a phonetic inventory matrix for the above languages respect of vowel sounds is included in
[0108] Consonants or vowels which are not common to a popular spoken language, and thereby may not be readily distinguishable by a transducer of that popular language, may be avoided for presentation according to certain embodiments of the invention. For example, speakers of the Hindi language are known to commonly have difficulty discriminating between [v] and [w] sounds, since the two are allophones in that language. Thus, in certain embodiments of the invention, [v] and [w] sounds may be avoided for presentation.
[0109] Accordingly, in certain embodiments of the invention, consonants for presentation to a subject may be reduced to a shortlist which avoids consonants that may not be readily distinguishable by speakers of each the most common languages. Such a shortlist may consist of the consonants [p] [t] [k] [m] [n] [s] [l] and [j]. In an embodiment these consonants may in turn be combined with the vowels [a], [i] and [o] to form [CVC] or [VCV] speech sounds which are believed to be generally distinguishable by speakers of the most commonly spoken languages.
[0110] Pursuant to the above, it is believed that speakers of the most commonly spoken languages should generally be able to readily discriminate between each of the following speech sounds: [apa], [ata], [aka], [ama], [ana], [asa], [ala], [aja], [ipi], [iti], [iki], [imi], [ini], [isi], [ili], [iji], [opo], [oto], [oko], [omo], [ono], [oso], [olo], and [ojo]. In an embodiment, speech sounds are presented in [VCV] format with the same vowel used twice in a given speech sound. Further, when presenting a pair of speech sounds in a sequence to a subject, a pair of speech sounds presented may share the same vowel, but differ in the selection of a consonant, so as to provide a ‘consonant pair’ of speech sounds presented to the subject. Following this embodiment, suitable speech sound pairs, or consonant pairs, would for example include: [asa] and [ala], or [imi] and [isi]. Presenting the above exemplified [VCV] speech sounds in consonant pairs allows for 84 different combinations of consonant pairs for presentation to a subject in accordance with this embodiment.
[0111] In a further embodiment, speech sounds presented together in a sequence need not utilise the same vowel or consonant twice. Speech sounds such as [ato] and [ito], or [ima] and [omi] may therefore be used together, as well as speech sound such as [tom], [tok], [pot] and [son]. In further embodiments, speech sounds need not be presented as consonant pairs. In broader embodiments any speech sounds may be used as part of the test to enable a subject to discriminate between speech sounds.
Hearing Analysis Method and System
[0112] In certain embodiments enabling self-administration, a hearing discrimination test may be implemented via a system comprising: a desktop or laptop computer, a tablet computer such as an iPad, a smart phone, a home assistant device such as a ‘Google Home’, or another device or combination of devices as further exemplified below. In each of the devices exemplified above, the device comprises a data processor and a memory in data communication with the data processor. In certain embodiments, as further described below, the memory may comprise a hard drive device contained inside or outside a particular device. In other embodiments, the memory may be provided remotely such as via the ‘cloud’. Other embodiments are also described below.
[0113] Speech sounds may be presented in sequential triplets, or longer sequences, from a transducer such as a loudspeaker integral to or separate from a device such as those described above. In an embodiment, speech sounds may be presented via headphones, which may be connected via cable, Bluetooth, Wi-Fi or other avenue to another device. In another embodiment the speech sounds may be presented via a larger loudspeaker such as desktop speaker or home stereo system. In another embodiment speech sounds may be presented to the user directly via a hearing aid. In another embodiment, speech sounds may be presented to the user via a bone conductor transducer or other suitable form of sound transducer. Any device which is capable of presenting speech sounds to the user may be employed. Upon presentation of the speech sounds, the subject is presented with options to select which speech sound within a presented sequence is different from those others presented, as further described below.
[0114]
[0115]
[0116] The memory unit (14) may in some embodiments also store a speech discrimination test (18). More specifically, the memory unit (14) may store a computer-readable version of the speech discrimination test (18) that can be executed by the computer (11). During execution, a portion of the speech discrimination test (18) may be temporarily loaded from, for example, the hard disk and into the main memory components of the memory unit (14). As noted elsewhere, in addition to the stand-alone arrangement described above, it is also possible to execute the speech discrimination test (18) from a network. For example, the speech discrimination test (18) may be stored on a server computer (not expressly shown) that is accessible to several client computers. This arrangement has an advantage in that updates to the speech discrimination test (18) may be quickly and easily implemented. Other environments for executing the speech discrimination test (18) may also be used without departing from the scope of the invention.
[0117] The source code for the speech discrimination test (18) may be written in any suitable programming language (e.g.: C, C++, BASIC, Java). In addition, the speech discrimination test (18) can be implemented using a number of different programming methodologies (e.g., top-down, object oriented).
[0118] In one embodiment, the methodology of the speech discrimination test (18) involves a plurality of individual modules or object class modules with subroutines, properties and functions that can be called to perform specific tasks. The modules or subroutines can be called from a main routine, which would control the general sequence and flow of the speech discrimination test (18) and from within other modules or subroutines, which would control specific functions or tasks in either an isolated or cooperative manner. This is further exemplified below in respect of
[0119] Depending on particular embodiments, other components (16) of the system (10) that may be present include a keyboard, mouse or touchpad, microphone, printer and the like.
[0120] In
[0121] Visual images may be presented to the subject via any suitable device or mechanism. In an embodiment, as described above, visual images may be presented via a touchscreen. This further enables the subject to select the ‘odd’ speech sound through interaction with the same touch screen. In other embodiments, visual images may be presented by for example a monitor, television, projector, or by another suitable mechanism such as a mechanical display (e.g. pop-up buttons). In other embodiments a screen might not be provided but, for example, a set of lights is provided which light up so that a given light, or colour of light, is associated with a given speech sound. In other embodiments a display may not be provided and the subject may simply be required to select speech sounds based on order, i.e. the subject may select the ‘first’ speech sound as the lone speech sound. Using this approach a blind subject may still have the ability to complete the speech discrimination test through either speech recognition software as further described below, by pressing on an associated ‘first sound’ button, ‘second sound’ button or ‘third sound’ button (based on the number of sounds presented) on a device; or by pressing a single button once, twice or three times as appropriate (again based on the number of sounds presented). A touch pad may for example be used as an alternative to a button.
[0122] In embodiments not involving a touchscreen a mechanism, component or device allowing the subject to select the ‘odd’ speech sound may be provided, such as a computer mouse, touchpad, keyboard, button arrangement, remote control, a motion detector (for example recognising a number of hand waves or particular facial movements) or sound detector (for example speech recognition software may allow the subject to simply state ‘first sound’, ‘second sound’ or similar which would be recognised by the software). Using a voice recognition approach, and without requiring a display to provide for visual representations, other devices such as a home assistant (e.g. a ‘Google Home’) or similar could be used to implement the test.
[0123] Where visual images are provided, rather than having a visual image appear as the speech sound is presented, in other embodiments the visual image may appear just before or just after the speech sound is presented, or the visual image may be highlighted in some way (such as by the visual image becoming larger, brighter, shimmering, moving, or changing colour), just before, during, or just after the speech sound is presented. Any mechanism which ensures that the subject can associate a presented visual image with a presented speech sound may be used according to embodiments of the invention.
[0124]
Development of a Hearing Discrimination Test System and Method—Stage I
[0125] A game-based method and system was developed to test the suitability of embodiments of the invention to test hearing and speech comprehension. In this game-based embodiment, speech sounds were selected to be presented as triplets of different [VCV] consonant pairs, and the method and device were designed to be self-administrable by adults.
[0126] Recordings of each speech sound were produced by a native Australian English female talker in an anechoic chamber. The recordings were reviewed by a panel, and two of the six recordings for each speech sound were selected for use in the test. All recordings were high-pass filtered at 250 Hz to minimise the impact of variations in frequency response at low frequencies for transducer outputs of a range of tablet computers and smartphones. The recordings were equalised in overall RMS level.
[0127] Preliminary testing was conducted with 50 subjects having hearing impairment. Criteria for subject selection involved hearing impaired adults: with at least one year of experience wearing cochlear implants, with no other co-morbidities affecting their communication ability, and who speak English as a first language. The subjects selected were users of either a single cochlear implant, a cochlear implant combined with a hearing aid, or bilateral cochlear implants. All subjects were tested when using a cochlear implant in one ear alone.
[0128] [VCV] speech sounds were presented during testing as consonant pairs (i.e. pairs sharing the same consonant but different vowels, e.g. [aka] and [ata]) at 65 dB SPL in speech-shaped noise. Signal-to-noise ratios were set for different presentation conditions: primarily Quiet, 7 dB signal to noise ratio (‘SNR’), and 0 dB SNR, but at several other SNRs for some of the subjects.
[0129]
[0130] The mean scores per consonant pair (expressed as proportion correct as a function of signal-to-noise ratio) are shown in
[0131] The reduced set of consonant pairs, each adjusted with their own sound-to-noise ratio, then became the set of stimuli to be used in the next stage of developing a speech comprehension testing method and device according to embodiments of the invention.
Development and Testing of a Hearing Analysis System and Method—Stage II
[0132] A testing method and system implemented by the inventors on a laptop computer was evaluated to examine whether it would be suitable for objectively testing speech comprehension.
[0133] 81 adult subjects participated in Phase II of the project. These included 41 adults wearing hearing aids with hearing loss of 4 frequency average hearing level (4FAHL) >40 dB HL, and 40 adults wearing ‘Nucleus’ cochlear implants with at least one year of cochlear implant experience. Table 1 gives the characteristics of participant subjects.
TABLE-US-00001 TABLE 1 characteristics of subjects No. of Age at Assessment Duration of use Device participants (Yrs) Age at first CI of first CI CI 40 Mean = 68.0 Mean = 60.0 Mean =8.1 Unilateral = 7 SD = 13.8 SD = 18.1 SD = 7.6 Bilateral = 10 Median = 71.3 Median = 62.7 Median = 6.0 CI + HA = 23 Range = (25.8-92.4) Range = (4.8-89.1) Range = (1.1-32.0) HA 41 Mean = 68.4 SD = 12.3 Median = 71.1 Range = (35.7-89.8)
[0134] For subjects using hearing aids alone, the distribution of hearing level, expressed as four-frequency average hearing loss (‘4FAHL’) across 0.5, 1, 2 and 4 kHz in the better ear is shown in
[0135] Prior to assessments, the hearing devices of the participants were checked. For users of hearing aids, otoscopy and tympanometry were performed to exclude cases of cerumen build up and/or middle ear dysfunction. Behavioural pure tone thresholds were measured in both ears using standard pure tone audiometry if an audiogram within 12 months of the test date was not available. Participants provided demographic information by completing a written questionnaire.
[0136] While subjects were wearing their devices (hearing aids or cochlear implants) at their personal settings, they were measured using a speech discrimination test (described as ‘LIT’ in several of the Figures) according an embodiment of the invention, using a laptop computer. The test was self-administered, with minimal input from an assessor. After written instructions were presented on screen, the subject was directed to adjust the overall loudness of presentation by moving a slider to set a comfortable loudness level. This was followed by a practice run after which the subject completed a test run. After every 20 trials of speech sound triplets, the subject could either take a brief break or press a button to continue testing until all 81 triplet sequences (as described above) were completed. The SNR for each triplet was adjusted in accordance with the findings for Stage I described above. Each subject completed two runs of 81 triplet sequences.
[0137] While not incorporated as part of the test discussed above, the background noise of the surrounding environment (e.g. household) may be monitored and factored into the volume at which the test is presented to the subject. For example, if the subject is taking the test at home, and there are high levels of background noise in the vicinity, a device such as a computer may monitor the noise levels and automatically adjust the level at which the test is presented, as opposed to having the subject manually adjust loudness levels for comfort.
[0138] The below Table 2, and
TABLE-US-00002 TABLE 2 Summary scores of hearing aid users Results (%) Mean 72.4 SD 13.4 25.sup.th percentile 64.5 50.sup.th percentile 75.5 75.sup.th percentile 83 Min-Max 42 to 94
[0139] Similarly, the below Table 3, and
TABLE-US-00003 TABLE 3 Summary scores of cochlear implant users Results (%) Mean 80.1 SD 14.1 25.sup.th percentile 75.8 50.sup.th percentile 82.8 75.sup.th percentile 90.0 Min-Max 36.5 to 99
[0140] As demonstrated by Table 2 and Table 3, cochlear implant users generally obtained better test results during speech discrimination testing, noting that the mean proportion of correct responses for hearing aid users was 72.4%, while cochlear implant users obtained a mean proportion of 80.1%.
[0141] While according to the embodiment analysed during Stage II testing, the metric extracted from subjects relates to the percentage of correct responses, it is nevertheless possible to utilise other metrics according to other embodiments of the invention. For example, by providing an adaptive signal-to-noise ratio during testing based on responses from the subject, it is possible to extract a signal-to-noise ratio at which a subject provides the correct response during the test at a pre-determined rate, such as for example 60%, 70%, or 80% of the time. In this way, a high signal to noise ratio would be indicative of poor speech discrimination abilities, since a high signal to noise ratio would be required to elicit a high proportion of correct responses. Alternatively, the signal-to-noise ratio need not be adjusted at all from sequence to sequence and in other embodiments it may be unnecessary to emit any noise at all during testing.
[0142] Product-Moment correlation analyses revealed large and highly significant correlations between the first and second runs of the speech discrimination test for users of cochlear implants (r=0.907, p<0.001) and for users of hearing aids (r=0.876, p<0.001). This indicated that the test used reliably differentiates between subjects with different degrees of ability to discriminate between speech sounds, so as to reliably identify those having limited ability to discriminate speech sounds.
Using System and Method to Estimate Benefits of Cochlear Implantation
[0143] The inventors performed analyses to determine whether a method and system for testing speech discrimination according to embodiments of the invention may be used to estimate whether the user of a hearing aid may benefit from cochlear implantation.
[0144] The results were used to adjust the test scores to remove the effect of duration of hearing loss. Then, the probability of a person using hearing aids to score higher with cochlear implants during testing was estimated as the proportion of cochlear implant users whose adjusted score was higher than the non-implanted person's adjusted score.
[0145] Further investigations on the impact by item-number on estimated probabilities were carried out using simulated data from the estimated distributions of performance with cochlear implants.
[0146] The four rows depict different values for duration of hearing loss. The left panels depict estimates calculated using the least squares method, and the right panels depict estimates calculated using the lasso estimation method. Although
[0147] While development of certain embodiments of the software have been undertaken with having regard to determining whether a subject would obtain better test results upon receiving cochlear implantation, it is to be understood that applications of the invention are not so limited. More simply, the test may be used to test the speech discrimination or speech comprehension abilities of a subject. Other applications include the ongoing monitoring of the speech discrimination or speech comprehension skills of a subject through repeated testing to, for example, identify instances of hearing degradation, or improvement in speech discrimination skills following training in the use of a device. In another application, the test may be used by those who do not currently use a hearing aid, cochlear implant or similar so as to determine whether a person has poor speech comprehension or discrimination abilities and may benefit from use of a hearing device such as a hearing aid, or otherwise may benefit from visiting a hearing professional for further analysis.
Program Modules
[0148] Having described the development of certain embodiments of the invention as set out above, now described with reference to
[0149] After initial power up, the main program module performs an equipment check at step B1 to ensure all components (for example transducers, screens,) of the system are functioning properly. Such a check may involve, for example, comparing the initial calibration data of the equipment with current measurements. In some embodiments, the various components of the equipment may be pre-calibrated together as a unit during manufacture or assembly. The calibration data may then be stored in a storage medium that is connected or attached to or sent together with the equipment. A determination is made at step B3 as to whether the equipment check passed, that is, whether the equipment is within a predetermined percentage of the initial calibration data. If the equipment check fails, then the main program module issues an equipment failure warning at step B4 and returns to the first step B1 to re-check the equipment.
[0150] If the equipment check passes, then the main program module proceeds to obtain the subject's information at step B5. This can be done, for example, by prompting the patient to manually enter his or her information (for example: name, address, date of birth, years of hearing difficulties, years of using a hearing aid, etc.), or by loading the information from a previously stored patient file. Here, as throughout the description, manual prompting may be done visually by displaying the instructions as text on a screen, or by audio instructions via a transducer, or by a combination of both. At step B6, the main program module allows the subject to select which test to be performed, for example a speech discrimination in which: no noise is emitted during testing (‘Test A’), in which noise levels during testing are modified from sequence to sequence for ensure that presented consonant pairs are equally difficult to discriminate (‘Test B’), or to modify noise levels during testing based on previous responses to as to identify an SNR at which the subject correctly identifies ‘lone’ speech sounds at a predetermined ratio (‘Test C’). In the flow chart shown it is possible to select multiple tests to be performed one after the other. For simplicity, the flow chart exemplifies a main program module in which only options Test A and Test B are presented to the subject.
[0151] After the above selection, the main program module makes a determination as to whether Test A was selected at step B7. If Test A was selected, then at step B8, the main program module presents the subject with Test A according, in certain embodiments, to another module or sub-routine. If Test A was not selected, or otherwise upon completion of Test A, the main program module moves to step B9.
[0152] At step B9, the main program module makes a determination as to whether Test B was selected. If Test B was selected, then at step B10, the main program module presents the subject with Test B according, in certain embodiments, to another module or sub-routine. If Test B was not selected, or otherwise upon completion of Test B, the main program module moves to step B11.
[0153] At step B11, the main program module alerts the subject that he or she has completed the selected hearing tests. At step B12 the data processor may analyse data obtained testing, however this may alternatively be performed as part of the module or sub-routine of a completed test. At step B13 results and relevant analyses arising from the selected test(s) are presented to the subject and the subject is presented with a list of options at step B14 which includes: forwarding the results to a nearby hearing professional as step B15, printing the results as step B16, or undertaking a new test session, in which main program module returns the subject to step B2.
[0154] Now described with reference to
[0155] The test commences at step C1 in which the subject has selected Test B from the main program module. At step C2 suitable volume levels for performing the test are determined. This may be manually chosen by the subject upon being presented with sound and a choice to raise or lower presented sound so as to provide comfortable levels Otherwise it may be automatically selected by taking account of background noise levels as detected by a microphone where the test is being performed (so as to, for example, present the test at a predetermined level above background noise).
[0156] The test then commences at step C3 in which a selection is made of a consonant pair, and a noise/SNR level to present to the subject. This information may be stored in a memory of the system. The selection may be made randomly from stored consonant pairs, or in a set sequence. To avoid familiarity with the test, in an embodiment the selection is made randomly. At C4 a first speech sound is presented to the subject associated with a first visual representation. This step is repeated for a second speech sound and associated visual representation at C5, and for a third speech sound and associated visual representation at C6. At C7 the subject is presented with three visual representations and required to select which of the visual representations is associated with a ‘lone’ speech sound.
[0157] At C8 a time limit may be set on the amount of time provided to the subject to select a visual representation such that, if the subject does not respond in a given time, the subject is returned to step C4 so as to rehear the already-presented consonant pair triplet. In certain embodiments this step is removed since it provides the subject an opportunity to rehear a triplet that presented difficulties, which may not be desirable in certain embodiments. Otherwise, once the subject selects a visual representation the module is taken to step C9, where a determination is made as to whether the subject has correctly identified the visual representation associated with the lone speech sound. If the lone speech sound is correctly identified, it is added to the correct number count at C10. If the lone speech sound is not correctly identified, it is added to the incorrect number count at C11. In an embodiment, data pertaining to the consonant pair that obtained an incorrect response is stored to enable future improvement to the test (such as adjusting SNR levels for a given consonant pair if continued testing demonstrates that the paired speech sounds are more difficult to discriminate than previously understood).
[0158] Either way, following a correct response at C10 or an incorrect response at C11, the module is led to C12 in which a determination is made as to whether all consonant pair triplets have been presented to the subject. If not all consonant pair triplets have been presented to the subject, the module returns to C3 where a determination is made as to which consonant pair triplet to present to the subject next (while, in an embodiment, ensuring that consonant pair triplets are not repeatedly presented). If all consonant pair triplets have been presented then the test is stopped at C15 and the number of correct responses is recorded. In certain embodiments the recorded results may be retrieved later, which may be particularly useful where the test is applied to record ongoing speech discrimination skills of a subject through repeated testing over time.
[0159] Summary
[0160] Based on the above, the inventors have developed a speech discrimination test which may, in an embodiment, be used to determine if a hearing-impaired subject is likely to obtain better test results upon cochlear implantation, as well as for other applications. In an embodiment, since the test may not require the subject to recognise speech sounds as words, a speech discrimination test may be language-neutral, as well as neutral regarding the syntactic and semantic knowledge of the subject. In an embodiment, speech sounds utilised in a speech discrimination test may be specifically selected to enable speakers of most languages to undergo the test.
[0161] In an embodiment, the test may be subject-implemented without necessitating the involvement of a hearing professional, such as a physician. Rather, the results of testing, in an embodiment including the likelihood that cochlear implantation would improve any further test results, may be automatically sent to a hearing professional and/or to the person taking the test, such as via email, for further consideration.
[0162] Modifications and variations as would be deemed obvious to the person skilled in the art are included within the ambit of the present invention as defined in the claims.