APPARATUS AND METHOD OF PROCESSING AUDIO SIGNALS
20210392450 · 2021-12-16
Inventors
Cpc classification
H04R1/1091
ELECTRICITY
H04R2225/67
ELECTRICITY
H04R25/70
ELECTRICITY
G10H2210/066
PHYSICS
H04R2430/01
ELECTRICITY
H04R2430/03
ELECTRICITY
International classification
Abstract
A method for processing audio signals includes extracting a fundamental frequency (F0) component from a first audio signal; processing the first audio signal with Dominant Melody Enhancement (DoME) based on a hearing profile and output a second audio signal; and providing the second audio signal to the user. The DoME enhances the F0 component. The enhancement weight of the DoME is corresponding to the hearing profile.
Claims
1. A method for processing audio signals comprising: extracting a fundamental frequency (F0) component from a first audio signal; processing the first audio signal with Dominant Melody Enhancement (DoME) based on a hearing profile to generate a second audio signal; and providing the second audio signal to a user; wherein the DoME enhances the F0 component, and the enhancement weight of the DoME corresponds to the hearing profile.
2. The method of claim 1, wherein the F0 component is enhanced by adding a frequency-modulated sine consisting of only the F0 component.
3. The method of claim 2, wherein the frequency-modulated sine is added from approximately −21.1 dB to −6.2 dB.
4. The method of claim 2, wherein the frequency-modulated sine is added from approximately −9.6 dB to −4.3 dB below −20 LUFS.
5. The method of claim 1, wherein the F0 component ranges from approximately 212 Hz to 1.4 kHz.
6. The method of claim 1, wherein the first audio signal includes a vocal group and an instrumental group, and the processing includes adjusting the weights of the vocal group and the instrumental group.
7. The method of claim 1, wherein the hearing profile comprises one or more settings for enhancing or reducing existing features of the first audio signal, and settings for synthesizing new features based on characteristics of the first audio signal and user calibration.
8. The method of claim 1, further comprising conducting a user calibration process comprising: obtaining settings for enhancing or reducing existing features of the first audio signal specific to the user's preferences and hearing loss, and electrical characteristic of hardware executing the method for processing audio signals.
9. An audio processing system, including: an audio source; a signal output; and a first processor electrically connected the audio source and the signal output, wherein the audio source generates a first audio signal, and the first processor extracts a F0 component from the first audio signal, and the first processor processes the first audio signal with DoME based on a hearing profile to generate a second audio signal, and the enhancement weight of the F0 component in the DoME corresponds to the hearing profile, and the signal output stimulates a cochlea of a user with the second audio signal.
10. The audio processing system of claim 9, wherein the signal output comprises a cochlear implant.
11. The audio processing system of claim 9, further including a first input device, wherein the first input device is electrically connected to the first processor, and the first input device is configured to generate a first controlling signal to the first processor, and the first processor adjusts the enhancement weight of the F0 component based on the first controlling signal and the hearing profile.
12. The audio processing system of claim 9, further including a second input device and a second processor, wherein the second processor is electrically connected to the first processor, the audio source and the signal output, and the second input device is electrically connected to the second processor, and the second input device is configured to generate a second controlling signal to the second processor, and the second processor adjusts enhancement weights of a vocal group and an instrumental group of the first audio signal based on the second controlling signal and the hearing profile.
13. The audio processing system of claim 9, wherein the signal output includes one or more dominant electrodes, and the first processor enhances stimulations by the dominant electrodes through the second audio signal, and the dominant electrodes are corresponded to signals range from approximately 212 Hz to 1.4 kHz.
14. The audio processing system of claim 9, wherein the hearing profile comprises one or more settings for enhancing or reducing existing features of the first audio signal, and settings for synthesizing new features based on characteristics of the first audio signal and user calibration.
15. An audio processing system, including: an audio source; an acoustic device; and a first processor electrically connected the audio source and the acoustic device, wherein the audio source generates a first audio signal, and the first processor extracts a F0 component from the first audio signal, and the first processor processes the first audio signal with DoME based on a hearing profile to generate a second audio signal, and the enhancement weight of the F0 component in the DoME corresponds to the hearing profile, and the acoustic device outputs the second audio signal to a user.
16. The audio processing system of claim 15, wherein the acoustic device comprises a loudspeaker, headphones, earphones, headsets, or earbuds.
17. The audio processing system of claim 15, wherein the hearing profile comprises one or more settings for enhancing or reducing existing features of the first audio signal, and settings for synthesizing new features based on characteristics of the first audio signal and user calibration.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0040] Embodiments of the invention are described in more details hereinafter with reference to the drawings, in which:
[0041]
[0042]
[0043]
[0044]
[0045]
[0046]
[0047]
DETAILED DESCRIPTION
[0048] The embodiments of the present invention provide a new preprocessing method and apparatus formed by extracting and enhancing the dominant melody (DoME) of typical music recordings, rather than taking the approach of subtracting elements of the audio signal with the goal of reducing harmonic complexity or reducing the music to elements assumed to translate best to CI listener.
[0049] Referring to
[0050] The method for processing audio signal AS1 includes: extracting a fundamental frequency (F0) from an audio signal AS1 (Step S1); processing the audio signal AS1 with DoME based on a hearing profile and output an audio signal AS2 (Step S2); and providing the audio signal AS2 to a user 50 (Step S3). During the processing step S2, the DoME enhances the F0 component, and the enhancement weight of the DoME is corresponding to the hearing profile.
[0051] In one aspect, the system 10 may utilize the method, and the audio source 100 generates the audio signal AS1, and the processor 120 extracts a F0 component from the audio signal AS1. The processor 120 processes the audio signal AS1 with DoME based on a hearing profile and output an audio signal AS2. The enhancement weight of the F0 component in the DoME is corresponding to the hearing profile, and the signal output 110 stimulates a cochlea 51 of a user 50 with the audio signal AS2.
[0052] In this embodiment, the hearing profile comprises settings to either enhance/reduce existing features of the audio and/or to synthesize new features based on characteristics of the source audio and user calibration.
[0053] To be specific, the signal output 110 may comprise a loudspeaker (e.g., one or more speakers, headphones, earphones, headsets, earbuds, etc.), a cochlear implant (electronic device), or a hearing aid (electronic device).
[0054] Referring to
[0055] The extracted F0 is then mixed with the original music recordings. Moreover, a user may adjust the volume of the F0 melody before mixing with the original music recordings until the music sounded most pleasant to the user. The adjusted volume is then saved as one of the parameters of the hearing profile of the audio processing system 10, and the hearing profile of the user of the hearing device (signal output 110) is determined (Step S21). In other words, the hearing profile is made to correspond to the audio source 100, the signal output 110, and the processor 120 of the audio processing system 10, and the method for processing audio signal AS1 (Step S22) may incorporate a user-adjustable calibration process, allowing each user to configure the music signal processing accordingly to allow for enhancement of musical features specific to that person's preferences and the electrical characteristic of their CI hardware, hearing loss, and the resulting artifacts.
[0056] However, the hearing profile is not limited to the volume or volume ratio of the dominant melodies of F0 melodies. In one embodiment, the hearing profile may also include volume or volume ratio of vocal group or instrumental group in the music recording. To be specific, the audio signal AS1 may further includes a vocal group and an instrumental group, and the processing step of the method also adjusts the weights of enhancement of the vocal group and the instrumental group. A user may save a preferred volume or volume ratio of vocal group or instrumental group, and with the enhancement of F0 component, the user may enjoy the music through audio signal AS2 (Step S23).
[0057] The audio processing system 10 and the method for processing audio signals have a user-specific calibration process that allows users to tailor-adjust musical features to achieve a more pleasurable music listening experience. Also, in some cases, that the calibration does not require reprogramming of the cochlear implant hardware, which is primarily pre-configured for human speech and not readily accessible by the end user.
[0058] To be specific, the F0 component of the audio signal AS1 is enhanced by adding a frequency-modulated sine consisting of only the F0 component. Moreover, the F0 component of the audio signal's AS1 dominant melody was enhanced by adding a pitch-tracked frequency-modulated sine wave in parallel to the audio signal AS1
[0059] In one embodiment, the frequency-modulated sine is added from −21.1 dB to −6.2 dB, and the effects of the DoME output an audio signal AS2 which is more pleasant to a user.
[0060] In one embodiment, the frequency-modulated sine is added from −9.6 dB to −4.3 dB below −20 LUFS, and the effects of the DoME output an audio signal AS2 which is more pleasant to a user and not having damaging or harmful loudness.
[0061] On the other hand, the frequency of the F0 component ranges from 212 Hz to 1.4 kHz. The F0 component is within the F0 range of the average male and female spoken voice, and within the average melodic range of most targeted musical excerpts.
[0062]
[0063] This method incorporates a user-adjustable calibration process, allowing each user to configure the music signal processing accordingly to allow for enhancement of musical features specific to that person's preferences and the electrical characteristic of their CI hardware, hearing loss, and the resulting artifacts. To be specific, the processing may include adjusting or enhancement of the F0 component, the vocal group, or the instrumental group.
[0064] This can be accomplished by offline software processing hosted on consumer devices, a hardware device arranged between the audio source and the playback device, or real-time via acoustic sensors such as microphones. The audio signals in these cases are processed based on user calibration settings to either enhance/reduce existing features of the audio or to synthesize new features based on characteristics of the source audio.
[0065] The audio processing system 10 and the method are designed specifically for cochlear implant users, and the signal processing employed is designed to compensate for the technological limitations of those devices as well as individual differences in music perception. On a signal processing level, this could be accomplished, for example by enhancing the main melody of the music, enhancing the percussive elements (drums, etc.), using source separation algorithms to enhancing only the vocal or only the bass, reducing the complexity of the music through filtering (e.g., frequency filtering), removing the source music entirely and leaving only the enhanced elements, etc. Some of these signal processing techniques may be based on those disclosed in Cappotto, D., Xuan, W., Meng, Q., Zhang, C., and Schnupp, J. (2018)., “Dominant Melody Enhancement in Cochlear Implants,” 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)—Proceedings (pp. 398-402), [8659661] (Asia-Pacific Signal and Information Processing Association Annual Summit and Conference—Proceedings), IEEE, 2018.
[0066] Various embodiments of the present invention provide the modification of audio signals by signal processing of the original audio source or by the generation of new audio content based on features extracted from the original. The auditory stimulus can be played back by one or more loudspeakers, such as consumer headphones or earphones, or used to modify settings (e.g., hardware settings) of a cochlear implant. The above can be used to personalize the auditory stimulus produced by such devices in order to adjust for the unique characteristics of a user's perception of musical features and the limitations of their cochlear implant.
[0067] Moreover, the method for processing audio signals further includes digitally adjusting audio signals AS1 using the determined hearing profile; and converting the digitally adjusted signals to analog signals using a digital-to-analog converter. The adjusting step includes adjusting amplitude, phase, and/or frequency of one or more or all components of the music signals.
[0068] Referring to
[0069] Referring to
[0070] The input device 130 is electrically connected to the processor 120, and the input device 130 is configured to generate a controlling signal to the processor 120, and the processor 120 adjusts the enhancement weight of the F0 component based on the controlling signal from the input device 130 and the hearing profile saved in the audio processing system 10.
[0071] During the process of determining the hearing profile or calibrate the processed audio signal AS2, the input device 130 can control the volume or volume ratio of the F0 component.
[0072] The processor 140 is electrically connected to the processor 120, the audio source 100, and the signal output 110. The input device 150 is electrically connected to the processor 140.
[0073] The input device 150 is configured to generated a controlling signal to the processor 140, and the processor 140 adjusts enhancement weights of a vocal group and an instrumental group of the audio signal AS1 based on the controlling signal and the hearing profile.
[0074] During the process of determining the hearing profile or calibrating the processed audio signal AS2, the input device 150 can control the volume or volume ratio of the vocal group and the instrumental group.
[0075] In the embodiment, the input devices 130, 150 may include a keyboard, a mouse, a stylus, an image scanner, a microphone, a tactile input device (e.g., touch sensitive screen), and an image/video input device (e.g., camera).
[0076] The signal output 110 includes dominant electrode 112. The processor 120 enhances stimulations by the dominant electrode 112 through the audio signal AS2, and the dominant electrodes 112 are corresponded to signals range from 212 Hz to 1.4 kHz. In other words, the dominant electrodes 112 correspond to the F0 component, and the F0 component is within the F0 range of the average male and female spoken voice, and within the average melodic range of most targeted musical excerpts.
[0077] Referring to
[0078] Moreover, the acoustic device 160 may be a loudspeaker, headphones, earphones, headsets, or earbuds.
[0079] Referring to
[0080] The audio processing system 200 may have different configurations, and it generally comprises suitable components necessary to receive, store, and execute appropriate computer instructions, commands, or codes.
[0081] The main components of the audio processing system 200 are a processor 202 and a memory unit 204. The processor 202 may be formed by one or more of: CPU, MCU, controllers, logic circuits, Raspberry Pi chip, digital signal processor (DSP), application-specific integrated circuit (ASIC), Field-Programmable Gate Array (FPGA), or any other digital or analog circuitry configured to interpret and/or to execute program instructions and/or to process data. The memory unit 204 may include one or more volatile memory unit (such as RAM, DRAM, SRAM), one or more non-volatile memory unit (such as ROM, PROM, EPROM, EEPROM, FRAM, MRAM, FLASH, SSD, NAND, and NVDIMM), or any of their combinations.
[0082] Preferably, the audio processing system 200 further includes one or more input devices 206 such as a keyboard, a mouse, a stylus, an image scanner, a microphone, a tactile input device (e.g., touch sensitive screen), and an image/video input device (e.g., camera).
[0083] The audio processing system 200 may further include one or more output devices 208 such as one or more displays (e.g., monitor), speakers, disk drives, headphones, earphones, printers, 3D printers, etc. The display may include an LCD display, an LED/OLED display, or any other suitable display that may or may not be touch sensitive.
[0084] The audio processing system 200 may further include one or more disk drives 212, which may encompass solid state drives, hard disk drives, optical drives, flash drives, and/or magnetic tape drives. A suitable operating system may be installed in the audio processing system 200, e.g., on the disk drive 212 or in the memory unit 204. The memory unit 204 and the disk drive 212 may be operated by the processor 202.
[0085] The audio processing system 200 also preferably includes a communication device 210 for establishing one or more communication links (not shown) with one or more other computing devices such as servers, personal computers, terminals, tablets, phones, or other wireless or handheld computing devices. The communication device 210 may be a modem, a Network Interface Card (NIC), an integrated network interface, a radio frequency transceiver, an optical port, an infrared port, a USB connection, or other wired or wireless communication interfaces. The communication links may be wired or wireless for communicating commands, instructions, information and/or data. Preferably, the processor 202, the memory unit 204, and optionally the input devices 206, the output devices 208, the communication device 210 and the disk drives 212 are connected with each other through a bus (e.g., a Peripheral Component Interconnect (PCI) such as PCI Express, a Universal Serial Bus (USB), an optical bus, or other like bus structure). In one embodiment, some of these components may be connected through a network such as the Internet or a cloud computing network. A person skilled in the art would appreciate that the audio processing system 200 shown in
[0086] It should be apparent to those skilled in the art that many modifications besides those already described are possible without departing from the inventive concepts herein. The inventive subject matter, therefore, is not to be restricted except in the spirit of the invention. Moreover, in interpreting the invention, all terms should be interpreted in the broadest possible manner consistent with the context. In particular, the terms “includes”, “including”, “comprises” and “comprising” should be interpreted as referring to elements, components, or steps in a non-exclusive manner, indicating that the referenced elements, components, or steps may be present, or utilized, or combined with other elements, components, or steps that are not expressly referenced.