Patent classifications
H04S2420/07
SOUND REPRODUCTION SYSTEM
A sound reproduction system includes a sound processing device connected to a stationary first output device including a plurality of sound output units and a portable second output device including a plurality of sound output units. In the sound reproduction system, the sound processing device generates a first sound output signal to be output to the first output device and a second sound output signal, which is different from the first sound output signal, to be output to the second output device. At least the second sound output signal out of the first and second sound output signals includes a signal that is obtained by performing 3-D sound processing.
AUDIO SIGNAL PROCESSING APPARATUS AND METHOD FOR BINAURAL RENDERING
The present invention relates to an audio signal processing apparatus and an audio signal processing method to perform binaural rendering. To this end, provided are an audio signal processing apparatus to perform binaural filtering an input audio signal, including: a first filtering unit configured to filter the input audio signal by a first lateral transfer function to generate a first lateral output signal; and a second filtering unit configured to filter the input audio signal by a second lateral transfer function to generate a second lateral output signal, wherein the first lateral transfer function and the second lateral transfer function are generated by modifying an interaural transfer function (ITF) with respect to the input audio signal and an audio signal processing method using the same.
Method for Processing Sound Signal and Terminal Device
A method includes: receiving, by using channels located in different positions of a terminal device, at least three signals emit by a same sound source; determining, according to three signals in the at least three signals, a signal delay difference between every two of the three signals; determining, according to the signal delay difference, the position of the sound source relative to the terminal device; and when the sound source is located in front of the terminal device, performing orientation enhancement processing on a target signal in the at least three signals, and obtaining a first output signal and a second output signal of the terminal device according to a result of the orientation enhancement processing, where the orientation enhancement processing is used to increase a degree of discrimination between a front characteristic frequency band and a rear characteristic frequency band of the target signal.
Signal Processing Device and Signal Processing Method
A signal processing device includes: a calculating unit performing calculation using signal levels of first and second acoustic signals; a determining unit, based on a result of a comparison between: the signal level of at least one of the first and second acoustic signals before the calculation; and a result of the calculation, determining whether a component of a third acoustic signal to be output from a position between a position from which the first acoustic signal is output and a position from which the second acoustic signal is output is included in the first and second acoustic signals; and a signal generating unit generating the third acoustic signal from the first and second acoustic signals when the determining unit determines that the component of the third acoustic signal is included in the first and second acoustic signals.
Subband spatial processing and crosstalk cancellation system for conferencing
Embodiments relate to providing a conference for client devices with spatialized audio. Input audio streams are received from the client devices. For each client device, placement data defining spatial locations of other client devices within a sound field is determined. A mixed stream including a left mixed channel and a right mixed channel for the client device is generated by mixing and panning input audio streams of the other client devices according to the placement data. A spatially enhanced stream including a left enhanced channel for a left speaker and a right enhanced channel for a right speaker is generated by applying subband spatial processing and crosstalk processing on the left mixed channel and the right mixed channel of the mixed stream.
Method for and apparatus for decoding an ambisonics audio soundfield representation for audio playback using 2D setups
Sound scenes in 3D can be synthesized or captured as a natural sound field. For decoding, a decode matrix is required that is specific for a given loudspeaker setup and is generated using the known loudspeaker positions. However, some source directions are attenuated for 2D loudspeaker setups like e.g. 5.1 surround. An improved method for decoding an encoded audio signal in soundfield format for L loudspeakers at known positions comprises steps of adding (10) a position of at least one virtual loudspeaker to the positions of the L loudspeakers, generating (11) a 3D decode matrix (D′), wherein the positions (Formula I) of the L loudspeakers and the at least one virtual position (Formula II) are used, downmixing (12) the 3D decode matrix (D′), and decoding (14) the encoded audio signal (i14) using the downscaled 3D decode matrix (Formula III). As a result, a plurality of decoded loudspeaker signals (q14) is obtained.
Multiband ducker
A multiband ducker is configured to duck a specific range of frequencies within a music signal in proportion to a corresponding range of frequencies within a speech signal, and then combine the ducked music signal with the speech signal for output to a user. In doing so, the multiband ducker separates the music signal into different frequency ranges, which may include low, middle, and high-range frequencies. The multiband ducker then reduces the amplitude of the specific range of frequencies found in the speech signal, typically the mid-range frequencies. When the ducked music signal and the speech signal are combined, the resultant signal includes important frequencies of the original music signal, including low-range and high-range, thereby allowing perception of the music signal to continue in relatively uninterrupted fashion. Additionally, the combined signal also includes the speech signal, allowing for the perception of intelligible speech concordant with the perception of music.
TRANSITION FUNCTIONS OF DECOMPOSED SIGNALS
A device including: first and second input units providing first and second input signals of first and second audio tracks, a decomposition unit to decompose the first input audio signal to obtain decomposed signals, a playback unit to start playback of a first output signal obtained from recombining at least first and second decomposed signals at first and second volume levels, respectively, and a transition unit for performing a transition between playback of the first output signal and playback of a second output signal obtained from the second input signal. The transition unit is adapted for reducing the first/second volume levels according to first/second transition functions. The device includes an analyzing unit to analyze an audio signal to determine a song part junction between two song parts. The transition time interval of at least one of the transition functions is set such as to include the song part junction.
Method, Apparatus or Systems for Processing Audio Objects
Diffuse or spatially large audio objects may be identified for special processing. A decorrelation process may be performed on audio signals corresponding to the large audio objects to produce decorrelated large audio object audio signals. These decorrelated large audio object audio signals may be associated with object locations, which may be stationary or time-varying locations. For example, the decorrelated large audio object audio signals may be rendered to virtual or actual speaker locations. The output of such a rendering process may be input to a scene simplification process. The decorrelation, associating and/or scene simplification processes may be performed prior to a process of encoding the audio data.
METHOD FOR CONVERTING A FIRST SET OF SIGNALS REPRESENTATIVE OF A SOUND FIELD INTO A SECOND SET OF SIGNALS AND ASSOCIATED ELECTRONIC DEVICE
A conversion method comprises the following steps: for each of the signals (b.sub.E(t)) of a first set, determining values (α(f)) respectively associated with frequency bands; for each frequency band, converting the values (α(f)) associated with the relevant frequency band into at least one value representative of a virtual sound source oriented along the spatial direction associated with a data item stored for the relevant temporal frequency band; for each temporal frequency band, determining, on the basis of an above-mentioned representative value, a plurality of values (γ(f)) associated with the different signals (b.sub.S(t)) of the second set, respectively; constructing each signal (b.sub.S(t)) of the second set on the basis of the values (γ(f)) associated with this signal of the second set.