PRECISELY CONTROLLED MICROPHONE ACOUSTIC ATTENUATOR WITH PROTECTIVE MICROPHONE ENCLOSURE
20230345158 · 2023-10-26
Inventors
- Scott A. Moser (La Jolla, CA, US)
- Carlos Montoya (San Diego, CA, US)
- Peter L. Madaffari (Camden, ME, US)
- Vijay Parsa (London, CA)
- Niamh C. Finnegan (Bristol, GB)
Cpc classification
H04R2231/003
ELECTRICITY
G10L13/027
PHYSICS
International classification
G10L13/027
PHYSICS
Abstract
An attenuator is disclosed that enables a microphone's relatively undistorted pick up of a voice that is generated in close proximity to the microphone. This attenuator is a key component of a groundbreaking assistive device or handset/headset that empowers individuals with speech impairments to effectively communicate and reintegrate into society. By leveraging advanced acoustic hardware, intelligent voice algorithms, and a comprehensive image-or-vocabulary-to-impaired-voice database, the handset/headset facilitates understanding the user's impaired speech by harvesting understood terminology and outputting the same in a communicative context. Thus, the handset/headset enables seamless public interaction, independence, and improved quality of life to the user with speech impediments and ensures equal participation and inclusion in everyday verbal interactions.
Claims
1. A method of picking up at least one speech sound comprising the steps of: enclosing a microphone in an acoustic attenuator that has a diaphragm assembly occupying an aperture of at least one volume of space for attenuating an acoustic speech sound; generating the speech sound in a chamber of a handset, wherein the chamber contains the acoustic attenuator; screening the plosive energy of the speech sound; attenuating the speech sound via passively reducing the sound level of the acoustic speech sound; picking up the acoustic speech sound via the microphone; converting the acoustic speech sound into an electric signal.
2. The method of claim 1 wherein speech sound is an impeded speech sound and wherein the method further comprises the step of sending the electric signal to a computerized device featuring computer readable memory with installed software for rehabilitating impeded speech and a processor for executing the installed software for rehabilitating speech.
3. The method of claim 2 wherein: the software features: an Impaired Speech Analysis and Comparison module configured to receive the electric signal and recognize the electrical signal as an unintelligible rendition of a word or phrase; a Generation of Intelligible Output module configured to rehabilitate the unintelligible rendition of the word or phrase via generating an intelligible output as a different electric signal; and, the method further comprises the steps of: executing the Impaired Speech module and the Generation of Intelligible Output module; and, outputting the different electric signal from the computerized device.
4. The method of claim 3 wherein the step of outputting the different electric signal further comprises outputting the different electric signal via a speaker wherein the different electric signal is used to generate a voice.
5. The method of claim 4 wherein speaker is located externally of said chamber of the handset.
6. The method of claim 3 wherein the step of outputting the different electric signal further comprises outputting the different electric signal via a graphical user interface wherein the different electric signal is text.
7. A handset comprising: a handle that defines a component housing a chamber disposed at a first end of the handle, said chamber featuring a mouth opening for engaging a mouth of a user, and a passive acoustic attenuator for receiving sounds, said passive acoustic attenuator comprising: an enclosed volume comprising at least one aperture, a microphone, and, a diaphragm assembly occupying said aperture, wherein the diaphragm assembly passively and reactively reduces the sound level coming into the microphone; an each chamber disposed at a second end of the handle with an ear opening for engaging the ear of a user; a speaker that is either configured to communicate sound into the ear chamber or externally of the handset; a near-field communication chip for (a) receiving electric signals from a computerized device and transmitting said electric signals to the speaker and (b) receiving an electric signal that represents sounds received by the microphone of the passive acoustic attenuator and transmitting the electric signal that represents sounds received by the microphone to the computerized device, said near-field communication chip disposed within the telephone component housing, said near-field communication chip electronically coupled to said speaker and said microphone.
8. The handset of claim 7 wherein the sounds received by the microphone represent impaired or impeded speech and wherein the computerized device features computer readable memory with installed software for rehabilitating impeded speech and a processor for executing the installed software for rehabilitating speech.
9. The handset of claim 8 wherein the software features: an Impaired Speech Analysis and Comparison module configured to receive the electric signal that represents sounds received by the microphone of the passive acoustic attenuator and configured to recognize the electrical signal that represents sound received by the microphone of the passive acoustic attenuator as the user's unintelligible rendition of a word or phrase; a Generation of Intelligible Output module configured to rehabilitate the user's unintelligible rendition of the word or phrase and configured to generate an intelligible output as an electric signal.
10. The handset of claim 9 wherein the software is executed such that the intelligible output as an electric signal is transmitted to the near-field communication chip of the handset for output by the speaker.
11. The handset of claim 9 wherein the software is executed such that the intelligible output as an electric signal is transmitted to a graphical user interface and displayed as text.
12. A passive acoustical attenuator for a microphone, said acoustical attenuator combining attenuation to lower a sound level of a sound introduced into the microphone with physical protection for the microphone, said acoustical attenuator defined by an enclosed volume of space bounded by a sound inlet at the proximate end, containing a diaphragm structure and bounded at the distal end by a sound outlet sealed to a microphone, wherein the sound entering at the proximate inlet is reduced in level according to the divider effect of acoustical compliances of the diaphragm and the enclosed volume of space that is approximately constant over a wide acoustical range of speech; wherein the passive acoustical attenuator for a microphone is disposed within a chamber of a handset and a speech sound enters the proximate inlet and is reduced in level according to the diver effect of acoustical compliances of the diaphragm and the enclosed volume of space and then converted to an electric signal.
13. The passive acoustical attenuator for a microphone of claim 12 wherein speech sound is an impeded speech sound and wherein the electric signal is sent to a computerized device featuring computer readable memory with installed software for rehabilitating impeded speech and a processor for executing the installed software for rehabilitating speech.
14. The passive acoustical attenuator for a microphone of claim 13 wherein: the software features: an Impaired Speech Analysis and Comparison module configured to receive the electric signal and recognize the electrical signal as an unintelligible rendition of a word or phrase; a Generation of Intelligible Output module configured to rehabilitate the unintelligible rendition of the word or phrase via generating an intelligible output as a different electric signal.
15. The passive acoustical attenuator for a microphone of claim 14 wherein the Impaired Speech and the Generation of Intelligible Output modules are executed by the processor and the different electric signal is output from the computerized device.
16. The passive acoustical attenuator for a microphone of claim 15 wherein the different electric signal is output via a speaker wherein the different electric signal is used to generate a voice.
17. The passive acoustical attenuator for a microphone of claim 16 wherein speaker is located externally of said chamber of the handset.
18. The passive acoustical attenuator for a microphone of claim 17 wherein the handset is configured to reduce the plosive raw voice of the impaired speech.
19. The passive acoustical attenuator for a microphone of claim 18 where the handset includes a low durometer voice air flow flap for exhale speech and inhale life air intake as needed for plosive words require more air flow for pronunciation.
20. The passive acoustical attenuator for a microphone of claim 18 wherein the air flow flap is a two-way vent such that the area for voice air intake and exhaust may be open for normal speech and air inhalation but closed off during expression of plosive words.
Description
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
[0066] The manner in which these objectives and other desirable characteristics can be obtained is explained in the following description and attached figures in which:
[0067]
[0068]
[0069]
[0070]
[0071]
[0072]
[0073]
[0074]
[0075]
[0076]
[0077]
[0078]
[0079]
[0080]
[0081]
[0082]
[0083]
[0084]
[0085]
[0086]
[0087]
[0088]
[0089]
[0090]
[0091]
[0092]
[0093]
[0094]
[0095]
[0096]
[0097]
[0098]
[0099]
[0100] In the drawings, the following reference numerals correspond with the associated components of the acoustic attenuator: [0101] 1—acoustic attenuator; [0102] 2—attenuator sound inlet; [0103] 3—attenuator collar; [0104] 4—attenuator shell; [0105] 5—microphone adapter ring; [0106] 6—attenuator sound exit; [0107] 7—enclosed volume; [0108] 20—attenuator diaphragm assembly; [0109] 21—diaphragm pocket; [0110] 22—stepped shoulder; [0111] 23—slot; [0112] 24—flange; [0113] 25—diaphragm film; [0114] 30—microphone; [0115] 31—microphone sound inlet; [0116] 32—microphone wiring; [0117] 33—microphone diaphragm; [0118] 34—microphone coil; [0119] 35—microphone magnet; [0120] 40—circular collar.
[0121] It is to be noted, however, that the appended figures illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments that will be appreciated by those reasonably skilled in the relevant arts. Also, figures are not necessarily made to scale but are representative.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
[0122] Generally disclosed is an assistive device for rehabilitating or treatment of speech impediments where the device includes a precisely controlled microphone acoustic attenuator with protective microphone enclosure. In use, the attenuator may be disposed in a telephone handset and be used for voice to text dictation. In the preferred use, the attenuator with protective microphone enclosure may be used to assist users with impaired speech to communicate more effectively. The details of a preferred embodiment of an attenuator are described in connection with the figures.
[0123]
[0124]
[0125]
[0126] Still referring to
[0127]
[0128]
[0129]
[0130]
[0131]
[0132]
[0133]
[0134]
[0135]
[0136] The effects of the microphone acoustical capacitances must be considered when computing the attenuation unless the microphone diaphragm's capacitance is much lower than the attenuator volume's capacitance. If this is not true or if the exact calculation is wanted, Cmic may be measured with an acoustic compliance test system, which a person of ordinary skill in the art of microphone design or acoustical test measurements can design and build. However, the acoustical capacitance of a diaphragm, like the diaphragm film 25, is difficult to pre-calculate because it depends on the diaphragm's material, geometry, and tensioning. A preferred diaphragm film 25 made of mylar is the same material used for subminiature diaphragms in electret microphones and as the insulator in electrical capacitors. Mylar is readily available in various thicknesses applicable to subminiature systems, and when metalized it forms a barrier to problematic vapors that could potentially harm the microphone or its components. The addition of the metallization layer and the additional processes of forming, clamping, or tensioning make the formula for computing the capacitance difficult to generate from a theoretical model. However, the acoustical capacitance of a diaphragm, Cadia, is generally proportional to the area and thickness of the diaphragm.
[0137] In practice, an appropriate diaphragm design procedure would be to first select the diaphragm thickness that gave the best protective properties and the diaphragm area that seemed applicable. Next, acoustic capacitance would be measured with acoustic capacitance test equipment. The capacitance value would then be used to vary the diaphragm's area to achieve the desired capacitance so that, when used with a known fixed volume, the desired attenuation would be reached. Alternately, the attenuator's acoustic volume could be varied to achieve the desired attenuation. Accordingly, the design process is very flexible.
[0138] Specifically,
[0139] In
[0140]
[0141]
[0142] If, however, we take into account a higher driver level so that 70 dB SPL average is recorded at 36″, but assume peak readings 15 dB higher, we get a maximum drive of 85 dB SPL. The numbers are then for each line at 100 Hz: =>85 dB SPL=>115 dB SPL=>171 dB SPL=>141 dB SPL. The side channel of the Quietphone® does help, but an Acoustic Attenuator of 30 dB or more is obviously called for. With the Quietphone® side channel and the attenuator, the level would be 141−30=111 dB, which is close to a conventional miniature microphone's limit. Without the side channel into the same enclosed volume, the level is 171−30=141 dB, resulting in severe distortion.
[0143]
[0144] The simplest improvement is electrical equalization. The shape of the attenuation does differ between the two microphone models, but for the examples of the particular model, the shapes are fairly constant, so an equalization network should give a consistent performance. It is true that the overload margin for the preamplifier is decreased, but the acoustic energy for speech is predominantly in the central portion of the curve and may not be a problem. However, there are methods to improve the shape of the attenuation curve that precede the microphone.
[0145] Returning to
Suitably, the first three cause the attenuation reduction at the low and high frequencies. Rdavt bypasses the attenuator diaphragm and should be as small as possible to have acoustic impedance as high as possible. Lda causes a peaking in the response shape within the pass band of the attenuator and should be as small as possible to shift the peak above the upper end of the pass band. Rda controls damping of the peak at the attenuator and should be set to flatten that peak. The last three can be set to minimize the attenuation's degradation, and the values need to be selected essentially are as in the preceding paragraph for the respective element. Unfortunately, the only way to do this is to design the microphone or select the microphone so that those criteria are met. Designing the microphone results in a more expensive microphone. Selecting the microphone is more cost efficient given the large number of microphone manufacturers, each with very broad product lines.
[0152] Returning to
[0153] Returning again to
[0154] The attenuator's level of attenuation can be checked before the microphone is cemented to the attenuator because the small leaks between the attenuator and the microphone will not affect the attenuation at or above 1 kHz when the vent hole is sealed with tape. The attenuator may be removed using its flange and replaced, even if the cement is strong enough to retain the microphone to the attenuator, although in a preferable embodiment the cement bond is breakable. When the bond is not breakable, a vent hole can be created in the attenuator's face and covered by tape while the assembly is checked and possibly replaced; as discussed, the tape sufficiently seals the vent hole to not affect attenuation. After the result is satisfactory, the vent hole can be covered over with a suitable viscous cement. Suitably, if the attenuator diaphragm is damaged after the assembly and after the vent hole is sealed, the diaphragm can be replaced by peeling back the viscous cement layer and replacing the diaphragm. Furthermore, the attenuator's volume can be ensured to be accurate if positive stops are used.
[0155] Additionally, adjusting the length of the chamber forming Cvol can also vary the attenuation. For example, in
[0156] Furthermore, adjusting the attenuation also adjusts the microphone's sensitivity. The adjustment could be used to achieve better uniformity from microphone to microphone because the base microphones' sensitivity normally varies by +/−3 dB to +/−4 dB according to industry specifications. For multi-inlet microphones, especially directional and noise canceling microphones, it is necessary to provide an acoustic attenuator for each sound inlet. It is necessary that the attenuator does not alter the level or phase of the input signals presented at each sound inlet. This is possible to achieve by matching the attenuators as they are built and then testing them to ensure good amplitude and phase match; a selection process to form a matched set is reasonable.
[0157]
[0158]
[0159]
[0160] Still referring to
[0161] Finally,
[0162] Thus,
[0163] 1. User Initialization: [0164] Prompt the user to initialize the software upon starting the handset/headset. [0165] Collect user-specific data such as name, preferred language, and any specific speech impairments or vocabulary limitations.
[0166] 2. Vocabulary Database Initialization: [0167] Provide the user with a list of vocabulary words, pictures or phrases to annunciate aloud while being recorded. [0168] Store the user's incorrect renditions of the words, along with the correct computerized renditions, in a database. [0169] Associate each recorded rendition with the corresponding vocabulary word.
[0170] 3. Online or Local Database Storage: [0171] Provide options for storing the vocabulary database either online or locally in the computer memory of the handset/headset. [0172] Ensure secure and accessible storage of the database for real-time comparison and retrieval.
[0173] 4. Real-Time Speech Recognition and Translation: [0174] Prompt the user to speak into the handset's chamber, ensuring even pressure and using an attenuator if desired. [0175] Convert the user's raw voice into an electrical signal using a microphone and analog matching circuit. [0176] Process the electrical signal through audio signal conditioning, amplification, noise reduction, and digital filtering within the Bluetooth (BT) chip. [0177] Transmit the processed audio signal wirelessly to a connected cell phone, computer, or other BT-capable device.
[0178] 5. Impaired Speech Analysis and Comparison: [0179] Receive the wireless audio signal on the connected device. [0180] Utilize the installed impaired speech assistance software to analyze the user's speech in real-time. [0181] Compare the user's unintelligible renditions with the stored vocabulary database, focusing on the incorrect renditions.
[0182] 6. Generation of Intelligible Output: [0183] Retrieve the corresponding correct renditions from the database based on the matched incorrect renditions. [0184] Generate intelligible output in the desired format: [0185] Convert the incorrect renditions to text, displaying them in a notepad-like section of the software for written communication. [0186] Generate a pronounced voice analog signal, transmitting it wirelessly back to the handset/headset for immediate playback through an outside ambient speaker on the device. [0187] Optionally, transmit the intelligible voice signal through the wireless communication system (e.g., cell phone call) to enable live communication with others.
[0188] 7. Real-Time Communication and Interaction: [0189] Enable seamless communication between the user and listeners through the translated output. [0190] Allow for ongoing interaction and conversation, with the software continuously processing the user's speech and providing accurate associations.
This logic flow outlines a few key steps and functionalities of the software installed in the computer memory of the handset and cooperating phone or computerized device, facilitating real-time speech recognition, comparison, and associations to enable effective communication for individuals with speech impairments.
[0191]
[0192]
[0193] The processed voice signal is then directed to a processor device that runs the software responsible for generating the desired output. The software analyzes and interprets the input speech to rehabilitate it, improving its intelligibility. The output can take several forms depending on the user's preference and communication needs. Firstly, the software can convert the rehabilitated speech into text, allowing it to be displayed as written words. This text output can be utilized in various applications, such as notepads, emails, or any other text-based communication platform. Secondly, the rehabilitated speech can be transmitted as part of a regular phone call. The processed voice signal is sent to the recipient on the other end, enabling real-time conversation through traditional phone communication. Lastly, the software can direct the rehabilitated speech to a speaker within the handset. This allows for in-person communication, where the user's rehabilitated speech is emitted audibly, enabling effective interaction with individuals in the immediate vicinity. In some embodiments, the output computer-generated speech or text can be in the user's or any other language.
[0194] By providing multiple output options via a touch screen on the graphical user interface of the computerized device (e.g., cell phone, computer, tablet, etc.), the hardware and software system of the invention ensures that individuals with speech impairments have versatile means of communication, tailored to their specific needs and preferences.
[0195]
Use Case: Quietphone® (FIGS. 18-24) for Assisting Impaired Speech
[0196] The Quietphone® (
[0197] 1. User Interaction: [0198] The user utilizes the Quietphone® or an open microphone on their cell phone to speak into the device. [0199] The microphone within the device converts the mechanical voice signal into an electrical signal, which comprises voltage and current representing the voice.
[0200] 2. Signal Conditioning: [0201] The electrical voice signal passes through an analog matching circuit that applies analog signal conditioning and filtering to optimize the signal quality.
[0202] 3. Bluetooth Connectivity: [0203] The conditioned signal enters the QCC3044 Bluetooth Chip (BT) or a similar BT communications chip. [0204] Inside the BT chip, the signal undergoes another phase of audio signal conditioning, this time in the digital domain. [0205] This conditioning stage allows for signal amplification, equalization, automatic gain control, and various other audio enhancements such as noise reduction and digital filtering.
[0206] 4. Wireless Transmission: [0207] Once the audio signal is fully processed, it is transmitted wirelessly within the BT frequency range. [0208] The wireless signal, containing the user's voice, is sent to the connected cell phone or other BT-capable devices where the impaired speech assistance software and user voice recognition database reside.
[0209] 5. Impaired Speech Assistance Software: [0210] The cell phone (or computer) receives the BT signal, and the dedicated impaired speech aid application (APP) processes the raw voice audio signal. [0211] Using the user's previously trained voice database, the APP generates accurate and appropriate words or phrases that correspond to the user's intended message with a high level of accuracy.
[0212] 6. Output Formats: [0213] The generated output can be converted into different formats: [0214] Written text can be displayed in a notepad-like section of the APP, used in emails, or any other application on the cell phone/tablet/computer. [0215] A computer-generated pronounced voice analog signal is sent back to the Quietphone® via BT. [0216] The signal is converted into an electrical signal that travels to a second speaker within the Quietphone®. [0217] The intelligible audio signal is emitted through the speaker, allowing the user to engage in real-time conversations with those around them. [0218] Conditioning and amplification stages between the BT chip and the speaker optimize the audio quality. [0219] A computer-generated pronounced voice analog signal is transmitted through wireless communication systems, such as a cell phone call or computer call, allowing listeners (not the user) to hear the intelligible computer-generated voice in real-time communication.
[0220] 7. Seamless Communication: [0221] In both scenarios, the APP takes approximately 150 milliseconds to process the user's partially intelligible or non-intelligible raw words and match them to clear and concise words. [0222] This enables seamless communication between the user and listeners, ensuring efficient and effective interactions in real-time.
By utilizing the Quietphone® and the impaired speech assistance software, individuals with impaired speech can overcome communication barriers, enabling them to express themselves more clearly and engage in meaningful conversations with others.
Use Case: Quietphone® (FIGS. 18-24) for Quietphone® Calls
[0223] Suitably, the Quietphone® (
[0224] 1. User Interaction: [0225] The user utilizes the Quietphone® or an open microphone on their cell phone to speak into the device. [0226] The microphone within the device converts the mechanical voice signal into an electrical signal, which comprises voltage and current representing the voice.
[0227] 2. Signal Conditioning: [0228] The electrical voice signal passes through an analog matching circuit that applies analog signal conditioning and filtering to optimize the signal quality. [0229] Bluetooth Connectivity: [0230] The conditioned signal enters the QCC3044 Bluetooth Chip (BT) or a similar BT communications chip. [0231] Inside the BT chip, the signal undergoes another phase of audio signal conditioning, this time in the digital domain. [0232] This conditioning stage allows for signal amplification, equalization, automatic gain control, and various other audio enhancements such as noise reduction and digital filtering.
[0233] 3. Wireless Transmission: [0234] After the signal is fully processed, it is transmitted wirelessly within the BT frequency range. [0235] The wireless signal carrying the user's voice is sent to the connected cell phone or other BT-capable devices.
[0236] 4. Cellular or Internet Communication: [0237] The cell phone (or connected device) receives the BT signal containing the user's voice and transmits it wirelessly, either through cellular frequency bands (600 MHz to 6 GHz) or over the internet, depending on the communication mode being used (cellular network or internet-based program).
[0238] 5. Bidirectional Communication: [0239] Simultaneously, the cell phone receives data from the other party involved in the call, typically another cell phone. [0240] This data signal is received by the cell phone and then transmitted back to the Quietphone® via BT.
[0241] 6. Audio Playback: [0242] The wireless signal reaches the BT chip inside the Quietphone®, where it is converted back into an electrical signal. [0243] The electrical signal is then directed to the Quietphone®'s speaker, located near the user's ear. [0244] Conditioning and amplification stages between the BT chip and the speaker ensure optimal audio quality.
[0245] 7. Alternative Communication Devices: [0246] The same process can be applied to other communication devices like laptops, iPads, and more. [0247] In such cases, the call signal may travel through the internet instead of relying on the cellular network.
By utilizing the Quietphone® or similar devices, users can make phone calls in a quiet and private manner. The technology enables wireless communication between the user's device and the cell phone or other devices, ensuring that the user's voice is transmitted clearly while also offering signal conditioning and audio enhancements.
[0248] Although the method and apparatus is described above in terms of various exemplary embodiments and implementations, it should be understood that the various features, aspects and functionality described in one or more of the individual embodiments are not limited in their applicability to the particular embodiment with which they are described, but instead might be applied, alone or in various combinations, to one or more of the other embodiments of the disclosed method and apparatus, whether or not such embodiments are described and whether or not such features are presented as being a part of a described embodiment. Thus, the breadth and scope of the claimed invention should not be limited by any of the above-described embodiments.
[0249] Terms and phrases used in this document, and variations thereof, unless otherwise expressly stated, should be construed as open-ended as opposed to limiting. As examples of the foregoing: the term “including” should be read as meaning “including, without limitation” or the like, the term “example” is used to provide exemplary instances of the item in discussion, not an exhaustive or limiting list thereof, the terms “a” or “an” should be read as meaning “at least one,” “one or more,” or the like, and adjectives such as “conventional,” “traditional,” “normal,” “standard,” “known” and terms of similar meaning should not be construed as limiting the item described to a given time period or to an item available as of a given time, but instead should be read to encompass conventional, traditional, normal, or standard technologies that might be available or known now or at any time in the future. Likewise, where this document refers to technologies that would be apparent or known to one of ordinary skill in the art, such technologies encompass those apparent or known to the skilled artisan now or at any time in the future. The presence of broadening words and phrases such as “one or more,” “at least,” “but not limited to” or other like phrases in some instances shall not be read to mean that the narrower case is intended or required in instances where such broadening phrases might be absent. Additionally, the various embodiments set forth herein are described in terms of exemplary block diagrams, flow charts and other illustrations. As will become apparent to one of ordinary skill in the art after reading this document, the illustrated embodiments and their various alternatives might be implemented without confinement to the illustrated examples. For example, block diagrams and their accompanying description should not be construed as mandating a particular architecture or configuration.
[0250] All original claims submitted with this specification are incorporated by reference in their entirety as if fully set forth herein.