A METHOD OF OUTPUTTING SOUND AND A LOUDSPEAKER
20230370777 · 2023-11-16
Inventors
Cpc classification
H04S2420/07
ELECTRICITY
H04S5/005
ELECTRICITY
H04R2420/03
ELECTRICITY
H04S7/30
ELECTRICITY
International classification
Abstract
A method of converting an audio signal into signals for a number of loudspeaker transducers, where the audio signal is divided up into audio sub signals each representing a particular frequency interval, and where the signal for each loudspeaker transducer comprises a portion of each audio sub signal which varies over time.
Claims
1.-14. (canceled)
15. A method of outputting sound based on an audio signal, the method comprising: receiving the audio signal, generating a number of audio sub signals from the audio signal, each audio sub-signal representing the audio signal within a frequency interval within the frequency interval of 100-8000 Hz, where the frequency interval of one sub signal is not fully included in the frequency interval of another sub signal, providing a speaker comprising a plurality of sound output loudspeaker transducers each capable of outputting sound in at least the interval of 100-8000 Hz, the loudspeaker transducers being positioned within a room or venue, generating an electrical sub signal for each loudspeaker transducer, each electrical sub signal comprising a predetermined portion of each audio sub signal, and feeding the electrical sub signals to the loudspeaker transducers, wherein the generation of the electrical sub signals comprises: altering, over time, the predetermined portions of the audio sub signals in each electrical sub signal and providing the electrical sub signals with the same or at least substantially the same sound energy, loudness or intensity.
16. The method according to claim 15, wherein the step of receiving the audio signal comprises receiving a stereo signal, and wherein the step of generating the audio sub signals comprises generating, for each channel in the stereo audio signal, a plurality of audio sub signals.
17. The method according to claim 15, wherein the step of receiving the audio signal comprises receiving a mono signal and generating from the audio signal a second signal being at least substantially phase inverted to the mono signal, and wherein the step of generating the audio sub signals comprises generating a plurality of audio sub signals for each of the mono audio signal and the second signal.
18. The method according to claim 15, further comprising the step of deriving, from the audio signal, a low frequency portion thereof having frequencies below a first threshold frequency and including the low frequency portion at least substantially evenly in all electrical sub signals.
19. The method according to claim 15, further comprising the step of deriving, from the audio signal, a high frequency portion thereof having frequencies above a second threshold frequency and including the high frequency portion at least substantially evenly in all electrical sub signals.
20. The method according to claim 15, wherein the step of generating the audio sub signals comprises selecting the frequency interval for one or more of the audio sub signals so that an energy/loudness in each audio sub signal is within 10% of a predetermined energy/loudness value.
21. The method according to claim 15, wherein the step of generating the electrical sub signals comprises, for one or more electrical sub signal(s), generating the electrical sub signal so that a portion of an audio sub band represented in the electrical sub band increases or decreases by at least 5% per second.
22. A system for outputting sound based on an audio signal, the system comprising: an input for receiving the audio signal, a speaker comprising a plurality of sound output loudspeaker transducers each capable of outputting sound in at least the interval of 100-8000 Hz, the loudspeaker transducers being positioned within a room or venue, a controller configured to: generate a number of audio sub signals from the audio signal, each audio sub-signal representing the audio signal within a frequency interval within the frequency interval of 100-8000 Hz, where the frequency interval of one sub signal is not fully included in the frequency interval of another sub signal, generate an electrical sub signal for each loudspeaker transducer, each electrical sub signal comprising a predetermined portion of each audio sub signal, and means for feeding the electrical sub signals to the loudspeaker transducers, wherein the controller is configured to generate each of the electrical sub signal so that: the predetermined portions of the audio sub signals in each electrical sub signal altering over time and a sound energy, loudness or intensity of the electrical sub signals is the same or at least substantially the same.
23. The system according to claim 22, wherein the input is configured to receive a stereo signal, and wherein the controller is configured to generate a plurality of audio sub signals for each channel in the stereo audio signal.
24. The system according to claim 22, wherein the input is configured to receive a mono signal and wherein the controller is configured to generate, from the audio signal, a second signal being at least substantially phase inverted to the mono signal, and to generate a plurality of audio sub signals for each of the mono audio signal and the second signal.
25. The system according to claim 22, wherein the controller is further configured to derive, from the audio signal, a low frequency portion thereof having frequencies below a first threshold frequency and include the low frequency portion at least substantially evenly in all electrical sub signals.
26. The system according to claim 22, wherein the controller is further configured to derive, from the audio signal, a high frequency portion thereof having frequencies above a second threshold frequency and include the high frequency portion at least substantially evenly in all electrical sub signals.
27. The system according to claim 22, wherein the controller is further configured to select the frequency interval for one or more of the audio sub signals so that an energy/loudness in each audio sub signal is within 10% of a predetermined energy/loudness value.
28. The system according to claim 22, wherein the controller is further configured to, for one or more electrical sub signal(s), generate the electrical sub signal so that a portion of an audio sub band represented in the electrical sub band increases or decreases by at least 5% per second.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0099] Unless specified otherwise, the accompanying drawings illustrate aspects of the innovations described herein. Referring to the drawings, wherein like numerals refer to like parts throughout the several views and this specification, several embodiments of presently disclosed principles are illustrated by way of example, and not by way of limitation.
[0100]
[0101]
[0102]
[0103]
[0104]
[0105]
[0106]
[0107]
[0108]
[0109]
[0110]
DETAILED DESCRIPTION
[0111] The following describes various innovative principles related to systems for providing sound spheres having smoothly changing, or constant, three-dimensional in-air transitions. For example, certain aspects of disclosed principles pertain to an audio device configured to project a desired sound sphere, or an approximation thereof, throughout a listening environment.
[0112] Embodiments of such systems described in the context of method acts are but particular examples of contemplated systems, chosen as being convenient illustrative examples of disclosed principles. One or more of the disclosed principles can be incorporated in various other audio systems to achieve any of a variety of corresponding system characteristics.
[0113] Thus, systems having attributes that are different from the specific examples discussed herein can embody one or more presently disclosed innovative principles, and can be used in applications not described herein in detail. Accordingly, such alternative embodiments also fall within the scope of this disclosure.
[0114] In some implementations, the innovation disclosed herein generally concern systems and associated techniques for providing three-dimensional sound spheres with multiple beams, that combine to provide smoothly changing sound localization information. For example, some disclosed audio systems can project subsections in frequency bands of the sound in subtly changing, or constant, phase relationships, and independent amplitude to the loudspeaker transducers. Thereby, the audio system can render added, or procured, spatial information to any input audio throughout a listening environment.
[0115] As but one example, an audio device can have an array of loudspeaker transducers constituting each an independent full-range transducer. The audio device includes a processor and a memory containing instructions that, when executed by the processor, cause the audio device to render a three-dimensional waveform as a 360 degree spherical shape, in weighted combination of individual virtual shape components, as coordinated pairs of shape component or otherwise, that are slowly moved along the loudspeaker transducers by a panning process of the audio signals. For each loudspeaker transducer, the audio device can filter a received audio signal according to a designated procedure. When executing the dynamic sound sphere, the audio device retains the original sound across the combined sphere components, when they are summed in the acoustic space. Therefore, for the listener the resulting sound retains the original sound's frequency envelope, but with the addition, or procurement, of a dynamic, or constant, three-dimensional audio spatialization.
[0116] The disclosure can combine its three-dimensional audio rendering with a summed signal above and below two designated thresholds, where the audio signal outside the thresholds holds no information about a sound's localization, discernible to the cognitive listening apparatus. These two ranges are summed separately into two monophonic audio signals and can be sent to all loudspeaker transducers simultaneously. The audio device can thereby provide the full three-dimensional spatialization that the cognitive listening apparatus can recognize, together with an independent control for all loudspeaker transducers of the low and high frequency ranges.
[0117] The disclosure can manage one mono signal input on one audio device in a number of independent sphere components that is equal to the number of the device's loudspeaker transducers, or a number of virtual sphere components that is different from the number of the device's loudspeaker transducers. Each sphere component can be a subset of a frequency range, and all components can be evenly distributed along the range as a balanced sum total of the components. These components can then be panned independently on all loudspeaker transducers on the geometric solid's planes, or as polar inverted pairs at opposite points on the geometric solid, or otherwise modified, and they can be positioned at any point between adjacent planes. Used in a paired stereo configuration with two devices, such a system will provide separate three-dimensional spatialization on each of the monophonic audio channels, and, rendered the left channel and the right channel separately to the two audio devices, resulting in a three-dimensional stereophonic audio rendering system. The stereo pairs can also be panned individually, and not observe any correlation in opposite points.
[0118] The disclosure can manage one stereo signal on one audio system in a number of independent iterations that is equal to half the number of the unit's loudspeaker transducers. Each pair is a subset of the frequency range of the stereo signal and can be positioned at opposite points on the geometric solid, or at any point between the solid's adjacent planes. The stereo pairs are panned equally, so that a single audio device will give a satisfactory rendering of the input stereo signal, hereby eschewing the need for two devices for rendering the full information of the original stereophonic signal, while still procuring the described three-dimensional audio cues. The result is a point source, three-dimensional stereophonic audio rendering system.
[0119] The instructions stored in processor memory can produce an adaptable division of the frequency bands that can, if so desired, observe equal loudness between the bands. This will avoid sudden directional changes due to changes in energy/loudness at very localized frequency ranges.
[0120] I Overview
[0121] Referring now to
[0122]
[0123] As will be explained more fully below, a three-dimensional sound sphere can be constructed by a combination of sphere components. A three-dimensional sound sphere is dependent on change of amplitude, phase and time along different audio frequencies, or frequency bands. A methodology can be devised to manage such dependencies, and disclosed audio devices can apply these methods to an acoustic signal, or a digital signal, containing an audio content to render as a three-dimensional sound sphere.
[0124] Section II describes principles related to such an audio device by way of reference to the device depicted in
[0125] II. Audio Devices
[0126]
[0127] In general, a loudspeaker array can have any number of individual loudspeaker transducers, despite that the illustrated array has six loudspeaker transducers. The number of loudspeaker transducers depicted in
[0128] In
[0129] Each of the loudspeaker transducers S1, S2, . . . , S6 in the illustrated loudspeaker array are distributed evenly on the cube's planes at a constant, or a substantially constant, position relative to, and at a uniform radial distance, polar, and azimuth angle from, the axis ‘center. In
[0130] Other arrangements for the loudspeaker transducers are possible. For instance, the loudspeaker transducers in the array may be distributed evenly within the loudspeaker cabinet 10, or unevenly. As well, the loudspeaker transducers S1, S2, . . . , S6 can be positioned at various selected spherical positions measured from the axis center, rather than at constant distance position as shown in
[0131] Each transducer S1, S2, . . . , S6 may be an electrodynamic or other type of loudspeaker transducer that may be specially designed for sound output at particular frequency bands, such as woofer, tweeter, midrange, or full-range, for example. The audio device 10 can be combined with a seventh loudspeaker transducer SO, to supplement output from the array. For example, the supplemental loudspeaker transducer SO can be so configured to radiate selected frequencies, e.g., low-end frequencies as a subwoofer. The supplemental loudspeaker transducer SO can be built into the audio device 10, or it can be housed in a separate cabinet. In addition or alternatively, the SO loudspeaker transducer may be used for high frequency output.
[0132] Although the loudspeaker cabinet 10 is shown as being cubic, other embodiments of a loudspeaker cabinet 10 have another shape. For example, some loudspeaker cabinets can be arranged as, e.g., a general prismatic structure, a tetrahedral structure, a spherical structure, an ellipsoidal structure a toroidal structure, or as any other desired three-dimensional shape.
[0133] III. The Three-Dimensional Sound Sphere
[0134] Referring again to
[0135] By projecting acoustic energy in a three dimensional sphere, a user's listening experience can be enhanced in comparison to two-dimensional audio system, since, and in contrast to prior art in one and two-dimensional sound fields, the three-dimensional listening cues provided by the disclosure are spatial and hence immersive, similarly to sound cues in the physical world.
[0136] Furthermore, the disclosure's listening space provides infinite listening positions around the device 10, as the added spatial audio cues do not operate on the basis of an ideal listening position, as long as the entire listening field, or sphere, contains an even balance, or an almost even balance, of the salient features of the original sound input.
[0137]
[0138] In some embodiments of audio devices, a three-dimensional sound field can be modified when the audio device's 10 proximity to a wall 22 is extreme, or very pronounced. For example, by representing the three-dimensional sound sphere 30 using polar coordinates with the z-axis of the audio device 10 positioned at the origin, a user can modify the sound sphere 30 from a sphere to an asymmetrical tri-axial ellipsoidal shape, by means of “drawing”, as on a touch screen, a directional scaling of the loudspeaker transducers ‘amplitude, relative to the z-axis of the audio device 10.
[0139] In still other embodiments, a user can select from a plurality of three-dimensional asymmetrical tri-axial ellipsoids stored by the audio device 10 or remotely. If stored remotely, the audio device 10 can load the selected tri-axial asymmetrical ellipsoid over a communication connection. And in still further embodiments, a user can “draw” a desired tri-axial asymmetrical ellipsoid contour or existing room boundary, as above, on a smartphone or a tablet, and the audio device 10 can receive a representation of the desired asymmetrical tri-axial ellipsoid, or room boundary, directly or indirectly from the user's device over a communication connection. Other forms of user input beside touch screens can be used, as described more fully below in connection with computer environments.
[0140] IV. Modal Decomposition and Reassembly of a Three-Dimensional Sound Sphere
[0141]
[0142] As means of but one example, and not of all possible embodiments, in
[0143]
[0144]
[0145] V. Directivity Considerations
[0146]
[0147] To achieve a desired sound sphere or smoothly varying sphere components (or pattern) over all frequencies, the sphere components described above can undergo equalization so each sphere component provides a corresponding sound field with a desired frequency response throughout. Stated differently, a filter can be designed to provide the desired frequency response throughout the sphere component. And, the equalized sphere components can then be combined to render a sound sphere having a smooth transition of sphere components across the range of audible frequencies and/or selected frequency bands, within the range of audible frequencies.
[0148] VI. Audio Processors
[0149]
[0150] The audio rendering processor 50 may be a special purpose processor such as an application specific integrated circuit (ASIC), a general purpose microprocessor, a field programmable gate array (FPGA), a digital signal controller, or a set of hardware logic structures (e.g., filters, arithmetic logic units, and dedicated state machines). In some instances, the audio rendering processor can be implemented using a combination of machine-executable instructions, that, when executed by a processor, cause the audio device to process one or more input channels as described. The rendering processor 50 is to receive the input channel of a piece of sound program content from an input audio source 51.
[0151] The input audio source 51 may provide a digital input or an analog input. The input audio source or input 51 may include a programmed processor that is running a media player application program and may include a decoder that produces the digital audio input to the rendering processor. To do so, the decoder may be capable of decoding an encoded audio signal, which has been encoded using any suitable audio codec, e.g., Advanced Audio Codec (AAC), MPEG Audio Layer II, MPEG Audio LAYER III, and Free Lossless Audio Codec (FLAC). Alternatively, the input audio source may include a codec that is converting an analog or optical audio signal, from a line input, for example, into digital form for the audio rendering processor 50. Alternatively, there may be more than one input audio channel, such as a two-channel input, namely left and right channels of a stereophonic recording of a musical work, or there may be more than two input audio channels, such as for example the entire audio soundtrack in 5.1-surround format of a motion picture film or movie. Other audio format examples are 7.1 and 9.1-surround formats.
[0152] The array of loudspeaker transducers 58 can render a desired sound sphere (or approximation thereof) based on a combination of sphere component segmentations 52a . . . 52N applied to the audio content by the audio rendering processor 50. Rendering processors 50 according to
[0153] In the loudspeaker transducer domain, a Sphere Domain Matrix can be applied to the various sphere domain signals to provide a signal to be reproduced by each respective loudspeaker transducer in the array 58. Generally speaking, the matrix is an M×N sized matrix, with N is the number of loudspeaker transducers, and M=(2×N)+(2×O) where O represents the number of virtual sphere components. An equalizer 56a . . . 56N can provide equalization to each respective sphere component 57a . . . 57N to adjust for variation in Directivity Factor arising from the particular audio device 10, and from any sphere adjustment towards a desired ellipsoid sphere contour, mentioned above.
[0154] It should be understood the audio rendering processor 50 is capable of performing other signal processing operations in order to render the input audio signal for playback by the transducer array 58 in a desired manner. In another embodiment, in order to determine how to modify the loudspeaker transducer signal, the audio rendering processor may use an adaptive filter process to determine constant, or varying, boundary frequencies.
[0155] VII. Computing Environments
[0156]
[0157] The computing environment 100 includes at least one central processing unit 110 and memory 120. In
[0158] A computing environment may have additional features. For example, the computing environment 100 includes storage 140, one or more input devices 150, one or more output devices 160, and one or more communication connections 170. An interconnection mechanism (not shown) such as a bus, a controller, or a network, interconnects the components of the computing environment 100. Typically, operating system software (not shown) provides an operating environment for other software executing in the computing environment 100, and coordinates activities of the components of the computing environment 100.
[0159] The storage 140 may be removable or non-removable, and can include selected forms of machine-readable media that includes magnetic disks, magnetic tapes or cassettes, non-volatile solid-state memory, CD-ROMs, CD-RWs, DVDs, magnetic tape, optical data storage devices, and carrier waves, or any other machine-readable medium which can be used to store information and which can be accessed within the computing environment 100. The storage 140 stores instructions for the software 180b, which can implement technologies described herein.
[0160] The storage 140 can also be distributed over a network so that software instructions are stored and executed in a distributed fashion. In other embodiments, some of these operations might be performed by specific hardware components that contain hardwired logic. Those operations might alternatively be performed by any combination of programmed data processing components and fixed hardwired circuit components.
[0161] The input device(s) 150 may be a touch input device, such as a keyboard, keypad, mouse, pen, touchscreen, touch pad, or trackball, a voice input device, a scanning device, or another device, that provides input to the computing environment 100. For audio, the input device(s) 150 may include a microphone or other transducer (e.g., a sound card or similar device that accepts audio input in analog or digital form), or a computer-readable media reader that provides audio samples to the computing environment 100.
[0162] The output device(s) 160 may be a display, printer, speaker transducer, DVD writer, or another device that provides output from the computing environment 100.
[0163] The communication connection(s) 170 enable communication over communication medium (e.g., a connecting network) to another computing entity. The communication medium conveys information such as computer-executable instructions, compressed graphics information, processed signal information (including processed audio signals), or other data in a modulated signal.
[0164] Thus, disclosed computing environments are suitable for performing disclosed orientation estimation and audio rendering processes as disclosed herein.
[0165] Machine-readable media are any available media that can be accessed within a computing environment 100. By way of example, and not limitation, with the computing environment 100, machine-readable media include memory 120, storage 140, communication media (not shown), and combinations of any of the above. Tangible machine-readable (or computer-readable) media exclude transitory signals.
[0166] As explained above, some disclosed principles can be embodied in a tangible, non-transitory machine-readable medium (such as a micro-electronic memory) having stored thereon instructions, which program one or more data processing components (generically referred to here as a “processor”) to perform the digital signal processing operations described above including estimating, adapting, computing, calculating, measuring, adjusting (by the audio processor 50), sensing, measuring, filtering, addition, subtraction, inversion, comparisons, and decision-making. In other embodiments, some of these operations (of a machine process) might be performed by specific electronic hardware components that contain hardwired logic (e.g., dedicated digital filter blocks). Those operations might alternatively be performed by any combination of programmed data processing components and fixed hardwired circuit components.
[0167] The audio device 10 can include a loudspeaker cabinet 12 configured to produce sound. The audio device 10 can also include a processor, and a non-transitory machine readable medium (memory) in which instructions are stored which, when executed by the processor, automatically perform the three-dimensional sphere construct processes, and supporting processes, as described herein.
[0168] The examples described above generally concern apparatus, methods, and related systems for rendering audio, and more particularly, to providing desired three-dimensional sphere patterns. Nonetheless, embodiments other than those described above in detail are contemplated based on the principles disclosed herein, together with any attendant changes in configurations of the respective apparatus described herein.
[0169] Directions and other relative references (e.g., up, down, top, bottom, left, right, rearward, forward, etc.) may be used to facilitate discussion of the drawings and principles herein, but are not intended to be limiting. For example, certain terms may be used such as “up”, “down”, “upper”, “lower”, “horizontal”, “vertical”, “left”, “right”, and the like. Such terms are used, where applicable, to provide some clarity of description when dealing with relative relationships, particularly with respect to the illustrated embodiments. Such terms are not, however, intended to imply absolute relationships, positions, and/or orientations. For example, with respect to an object, an “upper” surface can become a “lower” surface simply by turning the object over. Nevertheless, it is still the same surface and the object remains the same. As used herein, “and/or” means “and” or “or”, as well as “and” and “or”. Moreover, all patent and non-patent literature cited herein is hereby incorporated by reference in its entirety for all purposes.
[0170] The principles described above in connection with any particular example can be combined with the principles described in connection with another example described herein. Accordingly, this detailed description shall not be construed in a limiting sense, and following a review of this disclosure, those of ordinary skill in the art will appreciate the wide variety of signal processing and audio rendering techniques that can be devised using the various concepts described herein.
[0171] Moreover, those of ordinary skill in the art will appreciate that the exemplary embodiments disclosed herein can be adapted to various configurations and/or uses without departing from the disclosed principles. Applying the principles disclosed herein, it is possible to provide a wide variety of systems adapted to providing a desired three-dimensional spherical sound field. For example, modules identified as constituting a portion of a given computational engine in the above description or in the drawings can be partitioned differently than described herein, distributed among one or more modules, or omitted altogether. As well, such modules can be implemented as a portion of a different computational engine without departing from some disclosed principles.
[0172] The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the disclosed innovations. Various modifications to those embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of this disclosure. Thus, the claimed inventions are not intended to be limited to the embodiments shown herein, but are to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular, such as by use of the article “a” or “an” is not intended to mean “one and only one” unless specifically so stated, but rather “one or more”. All structural and functional equivalents to the features and methods acts of the various embodiments described throughout the disclosure that are known or later come to be known to those of ordinary skill in the art are intended to be encompassed by the features described and claimed herein. Moreover, nothing disclosed herein is intended to be dedicated to the public regardless of whether such disclosure is explicitly recited in the claims. No claim recitation is to be construed, unless the recitation is expressly recited using the phrase “means for” or “step for”.
[0173] Thus, in view of the many possible embodiments to which the disclosed principles can be applied, we reserve the right to claim any and all combinations of features and technologies described herein as understood by a person ordinary skilled in the art, including, for example, all that comes within the scope of the technology.