METHOD FOR CREATING A VIRTUAL ACOUSTIC STEREO SYSTEM WITH AN UNDISTORTED ACOUSTIC CENTER
20170230772 · 2017-08-10
Assignee
Inventors
- Martin E. Johnson (Los Gatos, CA)
- Sylvain J. Choisel (San Francisco, CA, US)
- Daniel K. Boothe (San Francisco, CA, US)
- Mitchell R. Lerner (Mountain View, CA, US)
Cpc classification
H04S1/002
ELECTRICITY
H04S2400/09
ELECTRICITY
H04S2420/01
ELECTRICITY
H04S2400/11
ELECTRICITY
H04S7/30
ELECTRICITY
International classification
Abstract
A system and method are described for transforming stereo signals into mid and side components xm and xs to apply processing to only the side-component xs and avoid processing the mid-component. By avoiding alteration to the mid-component XM, the system and method may reduce the effects of ill-conditioning, such as coloration that may be caused by processing a problematic mid component x.sub.M while still performing crosstalk cancellation and/or generating virtual sound sources. Additional processing may be separately applied to the mid and side components x.sub.M and xs and/or particular frequency bands of the original stereo signals to further reduce ill-conditioning.
Claims
1. A method for generating a set of virtual sound sources based on a left audio signal and a right audio signal corresponding to left and right channels for a piece of sound program content, comprising: transforming the left and right audio signals to a mid-component signal and a side-component signal; generating a set of filter values for the mid-component signal and the side-component signal, wherein the filter values 1) provide crosstalk cancellation between two speakers and 2) simulate virtual sound sources for the left and right channels of the piece of sound program content; normalizing the set of filter values such that the filter values corresponding to the mid-component signal avoid altering the mid-component signal; and applying the normalized set of filter values to one or more of the mid-component signal and the side-component signal.
2. The method of claim 1, wherein the mid-component signal is the sum of the right and left audio signals and the side-component signal is the difference between the left and right audio signals.
3. The method of claim 1, further comprising: transforming the resulting signals produced from the application of the set of normalized filter values to the one or more of the mid-component signal and the side-component signal to produce left and right filtered stereo audio signals; and driving the two speakers using the left and right filtered stereo audio signals to generate the virtual sound sources.
4. The method of claim 3, further comprising: band pass filtering the left audio signal using a first cutoff frequency and a second cutoff frequency to produce a band pass left signal, such that the band pass left signal includes frequencies from the left audio signal between the first and second cutoff frequencies; and band pass filtering the right audio signal using the first and second cutoff frequencies to produce a band pass right signal, such that the band pass right signal includes frequencies from the right audio signal between the first and second cutoff frequencies, wherein the band pass left and right signals are transformed to produce the mid-component signal and the side-component signal.
5. The method of claim 4, further comprising: low pass filtering the left audio signal using the first cutoff frequency to produce a low pass left signal; low pass filtering the right audio signal using the first cutoff frequency to produce a low pass right signal; high pass filtering the left audio signal using the second cutoff frequency to produce a high pass left signal; high pass filtering the right audio signal using the second cutoff frequency to produce a high pass right signal; combining the low pass left signal and the high pass left signal with the left filtered stereo audio signal; and combining the low pass right signal and the high pass right signal with the right filtered stereo audio signal, wherein the left filtered stereo audio signal after combination with the low pass left signal and the high pass left signal and the right filtered stereo audio signal after combination with the low pass right signal and the high pass right signal are used to drive the two speakers
6. The method of claim 3, further comprising: compressing the mid-component signal; and compressing the side-component signal, wherein compression of the mid-component signal is performed separately from compression of the side-component signal.
7. The method of claim 1, wherein the normalized set of filter values are applied to the side-component signal, the method further comprising: applying a delay to the mid-component signal while the side-component signal is being filtered using the normalized set of filter values such that the mid-component signal remains in sync with the side-component signal as a result of the delay.
8. The method of claim 1 wherein normalizing the set of filter values comprises dividing each non-zero filter value by the filter values corresponding to the mid-component signal such that the filter values corresponding to the mid-component are equal to one.
9. The method of claim 1 further comprising: equalizing the mid-component signal; and equalizing the side-component signal, wherein equalization of the mid-component signal is performed separately from equalization of the side-component signal.
10. A system for generating a set of virtual sound sources based on a left audio signal and a right audio signal corresponding to left and right channels for a piece of sound program content, comprising: a first set of filters to transform the left and right audio signals to a mid-component signal and a side-component signal; a processor to: generate a set of filter values for the mid-component signal and the side-component signal, wherein the filter values 1) provide crosstalk cancellation between two speakers and 2) simulate virtual sound sources for the left and right channels of the piece of sound program content, and normalize the set of filter values such that the filter values corresponding to the mid-component signal avoid altering the mid-component signal; and a second set of filters to apply the normalized set of filter values to one or more of the mid-component signal and the side-component signal.
11. The system of claim 10, wherein the mid-component signal is the sum of the right and left audio signals and the side-component signal is the difference between the left and right audio signals.
12. The system of claim 10, further comprising: a third set of filters to transform the resulting signals produced from the application of the set of filter values to one or more of the mid-component signal and the side-component signal to produce left and right filtered audio signals; and a set of drivers to drive the two speakers using the left and right filtered audio signals to generate the virtual sound sources.
13. The system of claim 10, wherein normalizing the set of filter values comprises dividing each non-zero filter value by the filter values corresponding to the mid-component signal such that the filter values corresponding to the mid-component are equal to one.
14. The system of claim 12, further comprising: a band pass filter to 1) filter the left audio signal using a first cutoff frequency and a second cutoff frequency to produce a band pass left signal, such that the band pass left signal includes frequencies from the left audio signal between the first and second cutoff frequencies and 2) filter the right audio signal using the first and second cutoff frequencies to produce a band pass right signal, such that the band pass right signal includes frequencies from the right audio signal between the first and second cutoff frequencies, wherein the band pass left and right signals are transformed by the first set of filters to produce the mid-component signal and the side-component signal.
15. The system of claim 14, further comprising: a low pass filter to filter 1) the left audio signal using the first cutoff frequency to produce a low pass left signal and 2) the right audio signal using the first cutoff frequency to produce a low pass right signal; a high pass filter to filter 1) the left audio signal using the second cutoff frequency to produce a high pass left signal and 2) the right audio signal using the second cutoff frequency to produce a high pass right signal; a summation unit to combine 1) the low pass left signal and the high pass left signal to the left filtered audio signal and 2) the low pass right signal and the high pass right signal to the right filtered audio signal, wherein the left filtered audio signal after combination with the low pass left signal and the high pass left signal and the right filtered audio signal after combination with the low pass right signal and the high pass right signal are used to drive the two speakers.
16. The system of claim 12, wherein first set of filters, the second set of filters, and the third set of filters are finite impulse response (FIR) filters.
17. An article of manufacture for generating a set of virtual sound sources based on a left audio signal and a right audio signal corresponding to left and right channels for a piece of sound program content, comprising: a non-transitory machine-readable storage medium that stores instructions which, when executed by a processor in a computing device, transform the left and right audio signals to a mid-component signal and a side-component signal; generate a set of filter values for the mid-component signal and the side-component signal, wherein the filter values 1) provide crosstalk cancellation between two speakers and 2) simulate virtual sound sources for the left and right channels of the piece of sound program content; normalize the set of filter values such that the filter values corresponding to the mid-component signal avoid altering the mid-component signal; and apply the normalized set of filter values to one or more of the mid-component signal and the side-component signal.
18. The article of manufacture of claim 17, wherein the mid-component signal is the sum of the right and left audio signals and the side-component signal is the difference between the left and right audio signals.
19. The article of manufacture of claim 17, wherein the non-transitory machine-readable storage medium stores further instructions which when executed by the processor: transform the resulting signals produced from the application of the set of filter values to one or more of the mid-component signal and the side-component signal to produce left and right filtered audio signals; and drive the two speakers using the left and right filtered audio signals to generate the virtual sound sources.
20. The article of manufacture of claim 17, wherein normalizing the set of filter values comprises dividing each non-zero filter value by the filter values corresponding to the mid-component signal such that the filter values corresponding to the mid-component are equal to one.
21. The article of manufacture of claim 20, wherein the non-transitory machine-readable storage medium stores further instructions which when executed by the processor: equalize the mid-component signal; and equalize the side-component signal, wherein equalization of the mid-component signal is performed separately from equalization of the side-component signal.
22. The article of manufacture of claim 20, wherein the non-transitory machine-readable storage medium stores further instructions which when executed by the processor: compress the mid-component signal; and compress the side-component signal, wherein compression of the mid-component signal is performed separately from compression of the side-component signal.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0011] The embodiments of the invention are illustrated by way of example and not by way of limitation in the figures of the accompanying drawings in which like references indicate similar elements. It should be noted that references to “an” or “one” embodiment of the invention in this disclosure are not necessarily to the same embodiment, and they mean at least one. Also, in the interest of conciseness and reducing the total number of figures, a given figure may be used to illustrate the features of more than one embodiment of the invention, and not all elements in the figure may be required for a given embodiment.
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022]
DETAILED DESCRIPTION
[0023] Several embodiments are described with reference to the appended drawings are now explained. While numerous details are set forth, it is understood that some embodiments of the invention may be practiced without these details. In other instances, well-known circuits, structures, and techniques have not been shown in detail so as not to obscure the understanding of this description.
[0024]
[0025] As shown in
[0026]
[0027] As shown in
[0028] Although the rendering strategy unit 209 is shown and described as a segment of software stored within the memory unit 203, in other embodiments the rendering strategy unit 209 may be implemented in hardware. For example, the rendering strategy unit 209 may be composed of a set of hardware circuitry, including filters (e.g., finite impulse response (FIR) filters) and processing units, that are used to implement the various operations and attributes described herein in relation to the rendering strategy unit 209.
[0029] In one embodiment, the audio source 103 may include one or more audio inputs 205 for receiving audio signals from external and/or remote devices. For example, the audio source 103 may receive audio signals from a streaming media service and/or a remote server. The audio signals may represent one or more channels of a piece of sound program content (e.g., a musical composition or an audio track for a movie). For example, a single signal corresponding to a single channel of a piece of multichannel sound program content may be received by an input 205 of the audio source 103. In another example, a single signal may correspond to multiple channels of a piece of sound program content, which are multiplexed onto the single signal.
[0030] In one embodiment, the audio source 103 may include a digital audio input 205A that receives digital audio signals from an external device and/or a remote device. For example, the audio input 205A may be a TOSLINK connector or a digital wireless interface (e.g., a wireless local area network (WLAN) adapter or a Bluetooth receiver). In one embodiment, the audio source 103 may include an analog audio input 205B that receives analog audio signals from an external device. For example, the audio input 205B may be a binding post, a Fahnestock clip, or a phono plug that is designed to receive and/or utilize a wire or conduit and a corresponding analog signal from an external device.
[0031] Although described as receiving pieces of sound program content from an external or remote source, in some embodiments pieces of sound program content may be stored locally on the audio source 103. For example, one or more pieces of sound program content may be stored within the memory unit 203.
[0032] In one embodiment, the audio source 103 may include an interface 207 for communicating with the loudspeakers 105 and/or other devices (e.g., remote audio/video streaming services). The interface 207 may utilize wired mediums (e.g., conduit or wire) to communicate with the loudspeakers 105. In another embodiment, the interface 207 may communicate with the loudspeakers 105 through a wireless connection as shown in
[0033] As described above, the loudspeakers 105 may be any device that includes at least one transducer 109 to produce sound in response to signals received from the audio source 103. For example, the loudspeakers 105 may each include a single transducer 109 to produce sound in the listening area 101. However, in other embodiments, the loudspeakers 105 may be loudspeaker arrays that include two or more transducers 109.
[0034] The transducers 109 may be any combination of full-range drivers, mid-range drivers, subwoofers, woofers, and tweeters. Each of the transducers 109 may use a lightweight diaphragm, or cone, connected to a rigid basket, or frame, via a flexible suspension that constrains a coil of wire (e.g., a voice coil) to move axially through a cylindrical magnetic gap. When an electrical audio signal is applied to the voice coil, a magnetic field is created by the electric current in the voice coil, making it a variable electromagnet. The coil and the transducers' 109 magnetic system interact, generating a mechanical force that causes the coil (and thus, the attached cone) to move back and forth, thereby reproducing sound under the control of the applied electrical audio signal coming from an audio source, such as the audio source 103. Although electromagnetic dynamic loudspeaker drivers are described for use as the transducers 109, those skilled in the art will recognize that other types of loudspeaker drivers, such as piezoelectric, planar electromagnetic and electrostatic drivers are possible.
[0035] Each transducer 109 may be individually and separately driven to produce sound in response to separate and discrete audio signals received from an audio source 103. By allowing the transducers 109 in the loudspeakers 105 to be individually and separately driven according to different parameters and settings (including delays and energy levels), the loudspeakers 105 may produce numerous separate sounds that represent each channel of a piece of sound program content output by the audio source 103.
[0036] Although shown in
[0037] Although described and shown as being separate from the audio source 103, in some embodiments, one or more components of the audio source 103 may be integrated within the loudspeakers 105. For example, one or more of the loudspeakers 105 may include the hardware processor 201, the memory unit 203, and the one or more audio inputs 205. In this example, a single loudspeaker 105 may be designated as a master loudspeaker 105. This master loudspeaker 105 may distribute sound program content and/or control signals (e.g., data describing beam pattern types) to each of the other loudspeakers 105 in the audio system 100.
[0038] As noted above, the rendering strategy unit 209 may be used to crosstalk cancel a set of audio signals and generate a set of virtual acoustic sound sources 111 based on this crosstalk cancellation. The objective of the virtual acoustic sound sources 111 is to create the illusion that sound is emanating from a direction which there is no real sound source (e.g., a loudspeaker 105). One example application might be stereo widening where two closely spaced loudspeakers 105 are too close together to give a good stereo rendering of sound program content (e.g., music or movies). For example, two loudspeakers 105 may be located within a compact audio source 103 such as a telephone or tablet computing device as shown in
[0039] In one embodiment, crosstalk cancellation may be used for generating the virtual sound sources 111. In this embodiment, a two-by-two matrix H of loudspeakers 105 to ears of the listener 107 describing the transfer functions may be inverted to allow independent control of sound at the right and left ears of the listener 107 as shown in
[0040] To address the issues related to coloration and ill-conditioning, such as coloration, in one embodiment the rendering strategy unit 209 may transform the problem from left-right stereo to mid-side stereo. In particular,
[0041] As described above, the signals x.sub.L and x.sub.R represent left-right stereo channels for a piece of sound program content. In this context, the signal x.sub.L characterizes sound in the left aural field represented by the piece of sound program content and the signal x.sub.R characterizes sound in the right aural field represented by the piece of sound program content. The signals x.sub.L and x.sub.R are synchronized such that playback of these signals through the loudspeaker 105 would create the illusion of directionality and audible perspective.
[0042] In a typical set of left-right stereo signals x.sub.L and x.sub.R, an instrument or vocal can be panned from left to right to generate what may be termed as the sound stage. Many times, but not necessarily always, the main focus of the piece of sound program content being played is panned down the middle (i.e., x.sub.L=x.sub.R). The most important example would be vocals (e.g., main vocals for a musical composition instead of background vocals or reverberation/effects, which are panned left or right). Also, low frequency instruments, such as bass and kick drums are typically panned down the middle. Accordingly, in the bass region, where it is important to maintain output levels (especially for small loudspeaker systems, such as those in consumer products), it may be important to reduce the effects of ill-conditioning, such as coloration. Further, for centrally panned vocals, it is important not to add coloration to the signals used to drive the loudspeakers 105. Coloration may also vary from listener-to-listener. Thus, it may be difficult to equalize out these coloration effects. Given these issues, the rendering strategy unit 209 may keep the centrally panned or mid-components untouched while making adjustments to side-components.
[0043] To allow for this independent handling/adjustment of mid-components and side-components, in one embodiment, the signals x.sub.L and x.sub.R may be transformed from left-right stereo to mid-side stereo using a mid-side transformation matrix T as shown in
x.sub.M=x.sub.L+x.sub.R
[0044] Similar to the value of the mid-component x.sub.M shown above, in one embodiment, the side-component x.sub.S may be generated based on the following equation:
x.sub.S=x.sub.L−x.sub.R
[0045] Accordingly, in contrast to the signals x.sub.L and x.sub.R that represented separate left and right components for a piece of sound program content, the mid-component x.sub.M represents the combined left-right stereo signals x.sub.L and x.sub.R (i.e., a center channel) while the mid-component x.sub.M represents the difference between these left-right stereo signals x.sub.L and x.sub.R. In these embodiments, the transformation matrix T may be calculated to generate the mid-component x.sub.M and the side-component x.sub.S according to the above equations. The transformation matrix T may be composed of real numbers and independent of frequency. Thus, the transformation matrix T may be applied using multiplication instead use of a filter. For example, in one embodiment the transformation matrix T may include the values shown below:
[0046] In other embodiments, different values for the transformation matrix T may be used such that the mid-component x.sub.M and the side-component x.sub.S are generated/isolated according to the above equations. Accordingly, the values for the transformation matrix T are provided by way of example and are not limiting on the possible values of the matrix T.
[0047] Following the conversion of the left-right stereo signals x.sub.L and x.sub.R to the mid-side components x.sub.M and x.sub.S, a set of filters may be applied to the mid-side components x.sub.M and x.sub.S. The set of filters may be represented by the matrix W shown in
[0048] In one embodiment, the matrix W may be represented by the values shown below, wherein i represents the imaginary number in the complex domain:
[0049] In the example matrix W shown above, values in the leftmost column of the matrix W represent filters that would be applied to the mid-component x.sub.M while the values in the rightmost column of the matrix W represent filters that would be applied to the side-component x.sub.S. As noted above, these filter values in the matrix W 1) perform crosstalk cancellation such that sound originating from the left loudspeaker 105 is not heard/picked-up by the right ear of the listener 107 and sound originating from the right loudspeaker 105 is not heard/picked-up by the left ear of the listener 107, 2) generate the virtual sound sources 111 in the listening area 101, and 3) provide transformation back to left-right stereo. Accordingly, the signals y.sub.L and y.sub.R represent left-right stereo signals after the filters represented by the matrix W have been applied to the mid-side stereo signals x.sub.M and x.sub.S.
[0050] As shown in
z.sub.LR=d=Dx.sub.LR=HWTx.sub.LR
[0051] In the above representation of the left-right stereo signals z.sub.L and z.sub.R and the desired signal d, the matrix W may be represented according to the equation below:
W=H.sup.−1DT.sup.−1
[0052] Accordingly, the matrix W 1) accounts for the effects of sound propagating from the loudspeakers 105 to the ears of the listener 107 through the inversion of the loudspeaker-to-ear transfer function H (i.e., H.sup.−1), 2) adjusts the mid-side stereo signals x.sub.M and x.sub.S to represent the virtual sound sources 111 represented by the matrix D, and 3) transforms the mid-side stereo signals x.sub.M and x.sub.S back to left-right stereo domain through the inversion of the transformation matrix T (i.e., T.sup.−1).
[0053] As described above, the mid-component of audio is especially susceptible to ill-conditioning and general poor results when crosstalk cancellation is applied. To avoid or mitigate these effects, in one embodiment, the matrix W may be normalized to avoid alteration of the mid-component signal x.sub.M. For example, the values in the matrix W corresponding to the mid-component signal x.sub.M may be set to a value of one (1.0) such that the mid-component signal x.sub.M is not altered when the matrix W is applied as described and shown above. In one embodiment, the normalized matrix W.sub.norm1 may be generated by dividing each value in the matrix W by the value of the values in the matrix W corresponding to the mid-component signal x.sub.M. As noted above, the values in the leftmost column of the matrix W represent filters that would be applied to the mid-component x.sub.M while the values in the rightmost column of the matrix W represent filters that would be applied to the side-component x.sub.S. In one embodiment, this normalized matrix W.sub.norm1 may be generated according to the equation below:
[0054] In the above equation, represents the top-left value of the matrix W as shown below:
##STR00001##
[0055] Accordingly, the normalized matrix W.sub.norm1 may be computed as shown below:
[0056] Accordingly, by altering the mid-components of the matrix W (i.e., the leftmost column of the matrix W) such that these value are equal to 1.0000, the normalized matrix W.sub.norm1 guarantees that the mid-component signal x.sub.M passes through without being altered by the matrix W.sub.norm1. By allowing the mid-component signal x.sub.M to remain unchanged and unaffected by the effects of crosstalk cancellation and other alterations caused by application of the matrices W and W.sub.norm1, ill-conditioning and other undesirable effects, which would be most noticeable in the mid-component signal x.sub.M as described above, may be reduced.
[0057] In one embodiment, the normalized matrix W.sub.norm1 may be compressed to generate the normalized matrix W.sub.norm2. In particular, in one embodiment, the normalized matrix W.sub.norm1 may be compressed such that the values corresponding to the side-component signal x.sub.S avoid becoming too large and consequently may reduce ill-conditioned effects, such as coloration effects. For example, the normalized matrix W.sub.norm2 may be represented by the values shown below, wherein α is less than one, may be frequency dependent, and represents an attenuation factor used to reduce excessively larger terms:
[0058] By compressing the values in the normalized matrix W.sub.norm1 to form the normalized matrix W.sub.norm2, ill-conditioning issues (e.g., coloration) that result in the loudspeakers 105 being driven hard and/or over-sensitivity related to assumptions regarding the HRTFs corresponding to the listener 107 may be reduced.
[0059] As described above and shown in
[0060] Although described above and shown in
[0061] In particular, as shown in
[0062] Following transformation by the matrix T, the matrix W.sub.MS may process the mid-side stereo signals x.sub.M and x.sub.S. In this embodiment, the desired signal d at the ears of the listener 107 may be defined by the HRTFs H for the desired angles of the virtual sound sources 111 represented by the matrix D. Accordingly, the left-right stereo signals z.sub.L and z.sub.R and the desired signal d detected at the ears of the listener 107 may be represented by the following equation:
z.sub.LR=d=Dx.sub.LR=HT.sup.−1W.sub.MSTx.sub.LR
[0063] In the above representation of the left-right stereo signals z.sub.L and z.sub.R and the desired signal d, the matrix W.sub.MS may be represented by the equation shown below:
W.sub.MS=TH.sup.−1DT.sup.−1
[0064] As noted above, the virtual sound sources 111 may be defined by the values in the matrix D. If D is symmetric (i.e., the virtual sound sources 111 are symmetrically placed and/or widened in relation to the loudspeakers 105) and H is symmetric (i.e., the loudspeakers 105 are symmetrically placed), then the matrix W.sub.MS may be a diagonal matrix (i.e., the values outside a main diagonal line within the matrix W.sub.MS are zero). For example, in one embodiment, the matrix W.sub.MS may be represented by the values shown in the diagonal matrix below:
[0065] In the example matrix W.sub.MS shown above, the top left value may be applied to the mid-component signal x.sub.M while the bottom right value may be applied to the side-component signal x.sub.S. In some embodiments, separate W.sub.MS matrices may be used for separate frequencies or frequency bands of the mid-side signals x.sub.M and x.sub.S. For example, 512 separate W.sub.MS matrices may be used for separate frequencies or frequency bands represented by the mid-side stereo signals x.sub.M and x.sub.S.
[0066] Similar to the signal processing shown and described in relation to
[0067] In the above equation, W.sub.MS.sub._.sub.11 represents the top-left value of the matrix W.sub.MS as shown below:
[0068] As noted above, in one embodiment, the matrix W.sub.MS may be a diagonal matrix (i.e., the values outside a main diagonal line within the matrix W.sub.MS are zero). In this embodiment, since the matrix W.sub.MS is a diagonal matrix, the computation of values for the matrix W.sub.MS.sub._.sub.norm1 may be performed on only the main diagonal of the matrix W.sub.MS (i.e., the non-zero values in the matrix W.sub.MS). Accordingly, the normalized matrix W.sub.MS.sub._.sub.norm1 may be computed as shown in the examples below:
[0069] As noted above in relation to the matrix W.sub.MS, separate W.sub.MS.sub._.sub.norm1 matrices may be used for separate frequencies or frequency bands represented by the mid-side signals x.sub.M and x.sub.S. Accordingly, different values may be applied to frequency components of the side-component signal x.sub.S.
[0070] By normalizing the mid-component signal x.sub.M, the mid-component signal x.sub.M may avoid processing by the matrix W.sub.MS.sub._.sub.norm1. Instead, as shown in
[0071] In one embodiment, compression and equalization may be independently applied to the separate chains of mid and side components. For example, as shown in
[0072] In some embodiments, ill-conditioning may be a factor of frequency with respect to the original left and right audio signals x.sub.L and x.sub.R. In particular, low frequency and high frequency content may suffer from ill-conditioning issues. In these embodiments, low pass, high pass, and band pass filtering may be used to separate each of the signals x.sub.L and x.sub.R by corresponding frequency bands. For example, as shown in
[0073] Following processing and delay, the signals produced by the VS system v.sub.L and v.sub.R may be summed by a summation unit with their delayed/unprocessed counterparts x.sub.LLow, x.sub.RLow, x.sub.LHigh and x.sub.RHigh to produce the signals y.sub.L and y.sub.R. These signals y.sub.L and y.sub.R may be played through the loudspeakers 105 to produce the left-right stereo signals z.sub.L and z.sub.R, which represent sound respectively heard at the left and right ears of the listener 107. As noted above, by sequestering low and high components of the original signals x.sub.L and x.sub.R, the system and method for processing described herein may reduce the effects of ill-conditioning, such as coloration that may be caused by processing problematic frequency bands.
[0074] As noted above, the system and method described herein transforms stereo signals into mid and side components x.sub.M and x.sub.S to apply processing to only the side-component x.sub.S and avoid processing the mid-component x.sub.M. By avoiding alteration to the mid-component x.sub.M, the system and method described herein may eliminate or greatly reduce the effects of ill-conditioning, such as coloration that may be caused by processing the problematic mid-component x.sub.M while still performing crosstalk cancellation and/or generating the virtual sound sources 111.
[0075] As explained above, an embodiment of the invention may be an article of manufacture in which a machine-readable medium (such as microelectronic memory) has stored thereon instructions that program one or more data processing components (generically referred to here as a “processor”) to perform the operations described above. In other embodiments, some of these operations might be performed by specific hardware components that contain hardwired logic (e.g., dedicated digital filter blocks and state machines). Those operations might alternatively be performed by any combination of programmed data processing components and fixed hardwired circuit components.
[0076] While certain embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that the invention is not limited to the specific constructions and arrangements shown and described, since various other modifications may occur to those of ordinary skill in the art. The description is thus to be regarded as illustrative instead of limiting.