Methods and systems for enhancing audio signals corrupted by noise
10726856 ยท 2020-07-28
Assignee
Inventors
- Jonathan Le Roux (Arlington, MA)
- Shinji Watanabe (Baltimore, MD)
- John Hershey (Winchester, MA)
- Gordon Wichern (Boston, MA, US)
Cpc classification
G10L21/02
PHYSICS
International classification
Abstract
Systems and methods for audio signal processing including an input interface to receive a noisy audio signal including a mixture of target audio signal and noise. An encoder to map each time-frequency bin of the noisy audio signal to one or more phase-related value from one or more phase quantization codebook of phase-related values indicative of the phase of the target signal. Calculate, for each time-frequency bin of the noisy audio signal, a magnitude ratio value indicative of a ratio of a magnitude of the target audio signal to a magnitude of the noisy audio signal. A filter to cancel the noise from the noisy audio signal based on the phase-related values and the magnitude ratio values to produce an enhanced audio signal. An output interface to output the enhanced audio signal.
Claims
1. An audio signal processing system, comprising: an input interface to receive a noisy audio signal including a mixture of a target audio signal and noise; an encoder to map each time-frequency bin of the noisy audio signal to one or more phase-related values from one or more phase quantization codebooks of phase-related values indicative of the phase of the target signal, and to calculate, for each time-frequency bin of the noisy audio signal, a magnitude ratio value indicative of a ratio of a magnitude of the target audio signal to a magnitude of the noisy audio signal; a filter to cancel the noise from the noisy audio signal based on the one or more phase-related values and the magnitude ratio values to produce an enhanced audio signal; and an output interface to output the enhanced audio signal.
2. The audio signal processing system of claim 1, wherein one of the one or more phase-related values represents an approximate value of the phase of a target signal in each time-frequency bin.
3. The audio signal processing system of claim 1, wherein one of the one or more phase-related values represents an approximate difference between the phase of a target signal in each time-frequency bin and a phase of the noisy audio signal in the corresponding time-frequency bin.
4. The audio signal processing system of claim 1, wherein one of the one or more phase-related values represents an approximate difference between the phase of a target signal in each time-frequency bin and the phase of a target signal in a different time-frequency bin.
5. The audio signal processing system of claim 1, further comprising a phase-related-value weights estimator, wherein the phase-related-value weights estimator estimates phase-related-value weights for each time-frequency bin, and the phase-related-value weights are used to combine the different phase-related values.
6. The audio signal processing system of claim 1, wherein the encoder includes parameters that determine the mappings of the time-frequency bins to the one or more phase-related values in the one or more phase quantization codebook.
7. The audio signal processing system of claim 6, wherein, given a predetermined set of phase values for the one or more phase quantization codebook, the parameters of the encoder are optimized so as to minimize an estimation error between training enhanced audio signal and corresponding training target audio signal on a training dataset of pairs of training noisy audio signal and training target audio signal.
8. The audio signal processing system of claim 6, wherein the phase values of the first quantization codebook are optimized together with the parameters of the encoder in order to minimize an estimation error between training enhanced audio signal and corresponding training target audio signal on a training dataset of pairs of training noisy audio signal and training target audio signal.
9. The audio signal processing system of claim 1, wherein the encoder maps each time-frequency bin of the noisy speech to a magnitude ratio value from a magnitude quantization codebook of magnitude ratio values indicative of quantized ratios of magnitudes of the target audio signal to magnitudes of the noisy audio signal.
10. The audio signal processing system of claim 9, wherein the magnitude quantization codebook includes multiple magnitude ratio values including at least one magnitude ratio value greater than one.
11. The audio signal processing system of claim 9, further comprising: a memory to store the first quantization codebook and the second quantization codebook, and to store a neural network trained to process the noisy audio signal to produce a first index of the phase value in the phase quantization codebook and a second index of the magnitude ratio value in the magnitude quantization codebook, wherein the encoder determines the first index and the second index using the neural network, and retrieves the phase value from the memory using the first index, and retrieves the magnitude ratio value from the memory using the second index.
12. The audio signal processing system of claim 9, wherein the phase values and the magnitude ratio values are optimized together with the parameters of the encoder in order to minimize an estimation error between training enhanced speech and corresponding training target speech.
13. The audio signal processing system of claim 9, wherein the first quantization codebook and the second quantization codebook form a joint quantization codebook with combinations of the phase values and the magnitude ratio values, such that the encoder maps each time-frequency bin of the noisy speech to the phase value and the magnitude ratio value forming a combination in the joint quantization codebook.
14. The audio signal processing system of claim 13, wherein the phase values and the magnitude ratio values are combined such that the joint quantization codebook includes a subset of all possible combinations of phase values and magnitude ratio values.
15. The audio signal processing system of claim 13, wherein the phase values and the magnitude ratio values are combined, such that the joint quantization codebook includes all possible combinations of phase values and magnitude ratio values.
16. A method for audio signal processing that includes a hardware processor coupled with a memory, wherein the memory has stored instructions and other data, the method comprising: accepting by an input interface, a noisy audio signal including a mixture of target audio signal and noise; mapping by the hardware processor, each time-frequency bin of the noisy audio signal to one or more phase-related values from one or more phase quantization codebook of phase-related values indicative of the phase of the target signal; calculating by the hardware processor, for each time-frequency bin of the noisy audio signal, a magnitude ratio value indicative of a ratio of a magnitude of the target audio signal to a magnitude of the noisy audio signal; cancelling using a filter, the noise from the noisy audio signal based on the phase values and the magnitude ratio values to produce an enhanced audio signal; and outputting by an output interface, the enhanced audio signal.
17. The method of claim 16, wherein the cancelling further comprising: updating time-frequency coefficients of the filter using the one or more phase values and the magnitude ratio values determined by the hardware processor for each time-frequency bin and to multiply the time-frequency coefficients of the filter with a time-frequency representation of the noisy audio signal to produce a time-frequency representation of the enhanced audio signal.
18. The method of claim 16, wherein the stored other data includes a first quantization codebook, a second quantization codebook, and a neural network trained to process the noisy audio signal to produce a first index of the phase value in the first quantization codebook and a second index of the magnitude ratio value in the second quantization codebook, wherein the hardware processor determines the first index and the second index using the neural network, and retrieves the phase value from the memory using the first index, and retrieves the magnitude ratio value from the memory using the second index.
19. The method of claim 18, wherein the first quantization codebook and the second quantization codebook form a joint quantization codebook with combinations of the phase values and the magnitude ratio values, such that the hardware processor maps each time-frequency bin of the noisy speech to the phase value and the magnitude ratio value forming a combination in the joint quantization codebook.
20. A non-transitory computer readable storage medium embodied thereon a program executable by a hardware processor for performing a method, the method comprising: accepting a noisy audio signal including a mixture of target audio signal and noise; mapping each time-frequency bin of the noisy audio signal to a phase value from a first quantization codebook of phase values indicative of quantized phase differences between phases of the noisy audio signal and phases of the target audio signal; mapping by the hardware processor, each time-frequency bin of the noisy audio signal to one or more phase-related values from one or more phase quantization codebook of phase-related values indicative of the phase of the target signal; calculating by the hardware processor, for each time-frequency bin of the noisy audio signal, a magnitude ratio value indicative of a ratio of a magnitude of the target audio signal to a magnitude of the noisy audio signal; cancelling using a filter, the noise from the noisy audio signal based on the phase values and the magnitude ratio values to produce an enhanced audio signal; and outputting by an output interface, the enhanced audio signal.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The presently disclosed embodiments will be further explained with reference to the attached drawings. The drawings shown are not necessarily to scale, with emphasis instead generally being placed upon illustrating the principles of the presently disclosed embodiments.
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15) While the above-identified drawings set forth presently disclosed embodiments, other embodiments are also contemplated, as noted in the discussion. This disclosure presents illustrative embodiments by way of representation and not limitation. Numerous other modifications and embodiments can be devised by those skilled in the art which fall within the scope and spirit of the principles of the presently disclosed embodiments.
DETAILED DESCRIPTION
(16) Overview
(17) The present disclosure relates to providing systems and methods for speech processing, including speech enhancement with noise suppression.
(18) Some embodiments of the present disclosure include an audio signal processing system having an input interface to receive a noisy audio signal including a mixture of target audio signal and noise. An encoder to map each time-frequency bin of the noisy audio signal to one or more phase-related value from one or more phase quantization codebook of phase-related values indicative of the phase of the target signal. Calculate, for each time-frequency bin of the noisy audio signal, a magnitude ratio value indicative of a ratio of a magnitude of the target audio signal to a magnitude of the noisy audio signal. A filter to cancel the noise from the noisy audio signal based on the phase-related values and the magnitude ratio values to produce an enhanced audio signal. An output interface to output the enhanced audio signal.
(19) Referring to
(20) Step 115 of
(21)
and the system may select the value 0 for bins whose energy is strongly dominated by the target signal energy: selecting the value 0 for such bins results in using the phase of the noisy signal as is for these bins, as the phase component of the filter at those bins will be equal to e.sup.0*i=1, where i denotes the imaginary unit of complex numbers, which will leave the phase of the noisy signal unchanged.
(22) Step 120 of
(23) Step 125 of
(24) The speech enhancement method 100 is directed to, among other things, obtain enhanced speech which is a processed version of the noisy speech that is closer in a certain sense to the underlying true clean speech or target speech.
(25) Note that target speech, i.e. clean speech, can be assumed to be only available during training, and not available during the real-world use of the system, according to some embodiments. For training, clean speech can be obtained with a close talking microphone, whereas the noisy speech can be obtained with a far-field microphone recorded at the same time, according to some embodiments. Or, given separate clean speech signals and noise signals, one can add the signals together to obtain noisy speech signals, where the clean and noisy pairs can be used together for training.
(26) Step 130 of
(27) Embodiments of the present disclosure provide unique aspects, by non-limiting example, an estimate of the phase of the target signal is obtained by relying on the selection or combination of a limited number of values within one or more phase quantization codebooks. These aspects allow the present disclosure to obtain a better estimate of the phase of the target signal, resulting in a better quality for the enhanced target signal.
(28) Referring to
(29) It is contemplated the hardware processor 140 can include two or more hardware processors depending upon the requirements of the specific application. Certainly, other components may be incorporated with method 100 including input interfaces, output interfaces and transceivers.
(30)
(31)
(32) For example,
(33)
(34)
For each time-frequency bin, the one or more phase codes 272 are used to select or combine phase-related values corresponding to the one or more phase codes within a phase codebook 280 to obtain a filter phase 278 for that time-frequency bin. For example, if the phase codebook 280 contains four values
(35)
the enhancement network 254 may estimate a code c.sub.t,f.sup.(p){0,1,2,3} for a time-frequency bin t,f, in which case the value of the filter phase 278 at time-frequency bin t,f may be set to
(36)
The filter magnitudes 274 and filter phases 278 are combined to obtain a filter 260. For example they can be combined by multiplying their values at each time-frequency bin t,f, in which case the value of the filter 260 at time-frequency bin t,f may be set to
(37)
A speech estimation module 265 then multiplies at each time-frequency bin the time-frequency representation of the noisy speech 105 with the filter 260 to obtain a time-frequency representation of the enhanced speech, and inverts that time-frequency representation of the enhanced speech to obtain the enhanced speech signal 290.
(38)
(39)
the enhancement network 354 may estimate a code c.sub.t,f.sup.(p){0,1,2,3} for a time-frequency bin t,f, in which case the value of the filter phase 378 at time-frequency bin t,f may be set to
(40)
The filter magnitudes 374 and filter phases 378 are combined to obtain a filter 360. For example they can be combined by multiplying their values at each time-frequency bin t,f, in which case the value of the filter 360 at time-frequency bin t,f may be set to
(41)
A speech estimation module 365 then multiplies at each time-frequency bin the time-frequency representation of the noisy speech 105 with the filter 360 to obtain a time-frequency representation of the enhanced speech, and inverts that time-frequency representation of the enhanced speech to obtain the enhanced speech signal 390.
(42)
(43)
(44) Features
(45) According to aspects of the present disclosure, the combinations of the phase values and the magnitude ratio values can minimize an estimation error between training enhanced speech and corresponding training target speech.
(46) Another aspect of the present disclosure can include the phase values and the magnitude ratio values being combined regularly and fully such that each phase value in the joint quantization codebook forms a combination with each magnitude ratio value in the joint quantization codebook. This is illustrated in
(47) Further, the phase values and the magnitude ratio values can be combined irregularly such that the joint quantization codebook includes a first magnitude ratio value forming combinations with a first set of phase values and includes a second magnitude ratio value forming combinations with a second set of phase values, wherein the first set of phase values differs from the second set of phase values. This is illustrated in
(48) Another aspect of the present disclosure can include one of the one or more phase-related values represents an approximate value of the phase of a target signal in each time-frequency bin. Further, another aspect can be that one of the one or more phase-related values represents an approximate difference between the phase of a target signal in each time-frequency bin and a phase of the noisy audio signal in the corresponding time-frequency bin.
(49) It is possible that one of the one or more phase-related values represents an approximate difference between the phase of a target signal in each time-frequency bin and the phase of a target signal in a different time-frequency bin. Wherein the different phase-related values are combined using phase-related-value weights. Such that, the phase-related-value weights are estimated for each time-frequency bin. This estimation can be performed by the network, or it can be performed offline by estimating the best combination according to some performance criterion on some training data.
(50) Another aspect can include the one or more phase-related values in the one or more phase quantization codebook minimize an estimation error between a training enhanced audio signal and a corresponding training target audio signal.
(51) Another aspect can include the encoder includes parameters that determine the mappings of the time-frequency bins to the one or more phase-related values in the one or more phase quantization codebook. Wherein, given a predetermined set of phase values for the one or more phase quantization codebook, the parameters of the encoder are optimized so as to minimize an estimation error between training enhanced audio signal and corresponding training target audio signal. Wherein the phase values of the first quantization codebook are optimized together with the parameters of the encoder in order to minimize an estimation error between training enhanced audio signal and corresponding training target audio signal. Another aspect can include that at least one magnitude ratio value can be greater than one.
(52) Another aspect can include the encoder that maps each time-frequency bin of the noisy speech to a magnitude ratio value from a magnitude quantization codebook of magnitude ratio values indicative of quantized ratios of magnitudes of the target audio signal to magnitudes of the noisy audio signal. Wherein the magnitude quantization codebook includes multiple magnitude ratio values including at least one magnitude ratio value greater than one. It is possible to further comprise a memory to store the first quantization codebook and the second quantization codebook, and to store a neural network trained to process the noisy audio signal to produce a first index of the phase value in the phase quantization codebook and a second index of the magnitude ratio value in the magnitude quantization codebook. Wherein the encoder determines the first index and the second index using the neural network, and retrieves the phase value from the memory using the first index, and retrieves the magnitude ratio value from the memory using the second index. Wherein the combinations of the phase values and the magnitude ratio values are optimized together with the parameters of the encoder in order to minimize an estimation error between training enhanced speech and corresponding training target speech. Wherein the first quantization codebook and the second quantization codebook form a joint quantization codebook with combinations of the phase values and the magnitude ratio values, such that the encoder maps each time-frequency bin of the noisy speech to the phase value and the magnitude ratio value forming a combination in the joint quantization codebook. Wherein the phase values and the magnitude ratio values are combined such that the joint quantization codebook includes a subset of all possible combinations of phase values and magnitude ratio values. Such that the phase values and the magnitude ratio values are combined, such that the joint quantization codebook includes all possible combinations of phase values and magnitude ratio values.
(53) An aspect further includes a processor to update time-frequency coefficients of the filter using the phase values and the magnitude ratio values determined by the encoder for each time-frequency bin and to multiply the time-frequency coefficients of the filter with a time-frequency representation of the noisy audio signal to produce a time-frequency representation of the enhanced audio signal.
(54) Another aspect can include a processor to update time-frequency coefficients of the filter using the phase values and the magnitude ratio values determined by the encoder for each time-frequency bin and to multiply the time-frequency coefficients of the filter with a time-frequency representation of the noisy audio signal to produce a time-frequency representation of the enhanced audio signal.
(55)
(56) The computing device 700A can include a power source 708, a processor 709, a memory 710, a storage device 711, all connected to a bus 750. Further, a high-speed interface 712, a low-speed interface 713, high-speed expansion ports 714 and low speed connection ports 715, can be connected to the bus 750. Also, a low-speed expansion port 716 is in connection with the bus 750.
(57) Contemplated are various component configurations that may be mounted on a common motherboard depending upon the specific application. Further still, an input interface 717 can be connected via bus 750 to an external receiver 706 and an output interface 718. A receiver 719 can be connected to an external transmitter 707 and a transmitter 720 via the bus 750. Also connected to the bus 750 can be an external memory 704, external sensors 703, machine(s) 702 and an environment 701. Further, one or more external input/output devices 705 can be connected to the bus 750. A network interface controller (NIC) 721 can be adapted to connect through the bus 750 to a network 722, wherein data or other data, among other things, can be rendered on a third party display device, third party imaging device, and/or third party printing device outside of the computer device 700A.
(58) Contemplated also is that the memory 710 can store instructions that are executable by the computer device 700A, historical data, and any data that can be utilized by the methods and systems of the present disclosure. The memory 710 can include random access memory (RAM), read only memory (ROM), flash memory, or any other suitable memory systems. The memory 710 can be a volatile memory unit or units, and/or a non-volatile memory unit or units. The memory 710 may also be another form of computer-readable medium, such as a magnetic or optical disk.
(59) Still referring to
(60) The system can be linked through the bus 750 optionally to a display interface or user Interface (HMI) 723 adapted to connect the system to a display device 725 and keyboard 724, wherein the display device 725 can include a computer monitor, camera, television, projector, or mobile device, among others.
(61) Still referring to
(62) The high-speed interface 712 manages bandwidth-intensive operations for the computing device 700A, while the low-speed interface 713 manages lower bandwidth-intensive operations. Such allocation of functions is an example only. In some implementations, the high-speed interface 712 can be coupled to the memory 710, a user interface (HMI) 723, and to a keyboard 724 and display 725 (e.g., through a graphics processor or accelerator), and to the high-speed expansion ports 714, which may accept various expansion cards (not shown) via bus 750. In the implementation, the low-speed interface 713 is coupled to the storage device 711 and the low-speed expansion port 715, via bus 750. The low-speed expansion port 715, which may include various communication ports (e.g., USB, Bluetooth, Ethernet, wireless Ethernet) may be coupled to one or more input/output devices 705, and other devices a keyboard 724, a pointing device (not shown), a scanner (not shown), or a networking device such as a switch or router, e.g., through a network adapter.
(63) Still referring to
(64)
(65) Referring to
(66) The processor 761 may communicate with a user through a control interface 766 and a display interface 767 coupled to the display 768. The display 768 may be, for example, a TFT (Thin-Film-Transistor Liquid Crystal Display) display or an OLED (Organic Light Emitting Diode) display, or other appropriate display technology. The display interface 767 may comprise appropriate circuitry for driving the display 768 to present graphical and other information to a user. The control interface 766 may receive commands from a user and convert them for submission to the processor 761. In addition, an external interface 769 may provide communication with the processor 761, so as to enable near area communication of the mobile computing device 700B with other devices. The external interface 769 may provide, for example, for wired communication in some implementations, or for wireless communication in other implementations, and multiple interfaces may also be used.
(67) Still referring to
(68) The memory 762 may include, for example, flash memory and/or NVRAM memory (non-volatile random access memory), as discussed below. In some implementations, instructions are stored in an information carrier, that the instructions, when executed by one or more processing devices (for example, processor 761), perform one or more methods, such as those described above. The instructions can also be stored by one or more storage devices, such as one or more computer or machine readable mediums (for example, the memory 762, the expansion memory 770, or memory on the processor 762). In some implementations, the instructions can be received in a propagated signal, for example, over the transceiver 771 or the external interface 769.
(69)
(70) The mobile computing device 700B may also communicate audibly using an audio codec 772, which may receive spoken information from a user and convert it to usable digital information. The audio codec 772 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of the mobile computing device 700B. Such sound may include sound from voice telephone calls, may include recorded sound (e.g., voice messages, music files, etc.) and may also include sound generated by applications operating on the mobile computing device 700B.
(71) Still referring to
Embodiments
(72) The following description provides exemplary embodiments only, and is not intended to limit the scope, applicability, or configuration of the disclosure. Rather, the following description of the exemplary embodiments will provide those skilled in the art with an enabling description for implementing one or more exemplary embodiments. Contemplated are various changes that may be made in the function and arrangement of elements without departing from the spirit and scope of the subject matter disclosed as set forth in the appended claims.
(73) Specific details are given in the following description to provide a thorough understanding of the embodiments. However, understood by one of ordinary skill in the art can be that the embodiments may be practiced without these specific details. For example, systems, processes, and other elements in the subject matter disclosed may be shown as components in block diagram form in order not to obscure the embodiments in unnecessary detail. In other instances, well-known processes, structures, and techniques may be shown without unnecessary detail in order to avoid obscuring the embodiments. Further, like reference numbers and designations in the various drawings indicated like elements.
(74) Also, individual embodiments may be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged. A process may be terminated when its operations are completed, but may have additional steps not discussed or included in a figure. Furthermore, not all operations in any particularly described process may occur in all embodiments. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, the function's termination can correspond to a return of the function to the calling function or the main function.
(75) Furthermore, embodiments of the subject matter disclosed may be implemented, at least in part, either manually or automatically. Manual or automatic implementations may be executed, or at least assisted, through the use of machines, hardware, software, firmware, middleware, microcode, hardware description languages, or any combination thereof. When implemented in software, firmware, middleware or microcode, the program code or code segments to perform the necessary tasks may be stored in a machine readable medium. A processor(s) may perform the necessary tasks.
(76) Further, embodiments of the present disclosure and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly-embodied computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Further some embodiments of the present disclosure can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions encoded on a tangible non transitory program carrier for execution by, or to control the operation of, data processing apparatus. Further still, program instructions can be encoded on an artificially generated propagated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus. The computer storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them.
(77) According to embodiments of the present disclosure the term data processing apparatus can encompass all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit). The apparatus can also include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
(78) A computer program (which may also be referred to or described as a program, software, a software application, a module, a software module, a script, or code) can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data, e.g., one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, e.g., files that store one or more modules, sub programs, or portions of code. A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network. Computers suitable for the execution of a computer program include, by way of example, can be based on general or special purpose microprocessors or both, or any other kind of central processing unit. Generally, a central processing unit will receive instructions and data from a read only memory or a random access memory or both. The essential elements of a computer are a central processing unit for performing or executing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Moreover, a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device, e.g., a universal serial bus (USB) flash drive, to name just a few.
(79) To provide for interaction with a user, embodiments of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user's client device in response to requests received from the web browser.
(80) Embodiments of the subject matter described in this specification can be implemented in a computing system that includes a back end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network (LAN) and a wide area network (WAN), e.g., the Internet.
(81) The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
(82) Although the present disclosure has been described with reference to certain preferred embodiments, it is to be understood that various other adaptations and modifications can be made within the spirit and scope of the present disclosure. Therefore, it is the aspect of the append claims to cover all such variations and modifications as come within the true spirit and scope of the present disclosure.