H04S5/02

ACOUSTIC SIGNAL PROCESSING APPARATUS, ACOUSTIC SIGNAL PROCESSING METHOD, AND PROGRAM
20180007485 · 2018-01-04 ·

The present technology relates to an acoustic signal processing apparatus, an acoustic signal processing method, and a program for expanding a range of listening positions in which an effect of a transaural reproduction system can be obtained. First and second output signals for localizing a sound image in front of or behind and on the left of a first position located on the left of a listening position are output from first and second speakers, respectively. Third and fourth output signals for localizing a sound image in front of or behind and on the right of a second position located on the right of the listening position are output from third and fourth speakers, respectively. The first speaker is disposed in a first direction in front of or behind the listening position and on the left of the listening position. The second speaker is disposed in the first direction and on the right of the listening position. The third speaker is disposed in the first direction and on the left of the listening position and on the right of the first speaker. The fourth speaker is disposed in the first direction of the listening position and on the right of the second speaker. The present technology can be applied, for example, to an acoustic processing system.

ACOUSTIC SIGNAL PROCESSING APPARATUS, ACOUSTIC SIGNAL PROCESSING METHOD, AND PROGRAM
20180007485 · 2018-01-04 ·

The present technology relates to an acoustic signal processing apparatus, an acoustic signal processing method, and a program for expanding a range of listening positions in which an effect of a transaural reproduction system can be obtained. First and second output signals for localizing a sound image in front of or behind and on the left of a first position located on the left of a listening position are output from first and second speakers, respectively. Third and fourth output signals for localizing a sound image in front of or behind and on the right of a second position located on the right of the listening position are output from third and fourth speakers, respectively. The first speaker is disposed in a first direction in front of or behind the listening position and on the left of the listening position. The second speaker is disposed in the first direction and on the right of the listening position. The third speaker is disposed in the first direction and on the left of the listening position and on the right of the first speaker. The fourth speaker is disposed in the first direction of the listening position and on the right of the second speaker. The present technology can be applied, for example, to an acoustic processing system.

APPARATUS AND METHOD FOR PROCESSING STEREO SIGNALS FOR REPRODUCTION IN CARS TO ACHIEVE INDIVIDUAL THREE-DIMENSIONAL SOUND BY FRONTAL LOUDSPEAKERS

Embodiments provide a digital processor including an ambient portion extractor and a spatial effect processing stage. The ambient portion extractor is configured to extract an ambient portion from a multi-channel signal. The spatial effect processing stage is configured to generate a spatial effect signal based on the ambient portion of the multi-channel signal. The digital processor is configured to combine the multi-channel signal or a processed version thereof with the spatial effect signal.

APPARATUS AND METHOD FOR PROCESSING STEREO SIGNALS FOR REPRODUCTION IN CARS TO ACHIEVE INDIVIDUAL THREE-DIMENSIONAL SOUND BY FRONTAL LOUDSPEAKERS

Embodiments provide a digital processor including an ambient portion extractor and a spatial effect processing stage. The ambient portion extractor is configured to extract an ambient portion from a multi-channel signal. The spatial effect processing stage is configured to generate a spatial effect signal based on the ambient portion of the multi-channel signal. The digital processor is configured to combine the multi-channel signal or a processed version thereof with the spatial effect signal.

METHOD FOR PROCESSING SOUND ON BASIS OF IMAGE INFORMATION, AND CORRESPONDING DEVICE

A method of processing an audio signal including at least one audio object based on image information includes: obtaining the audio signal and a current image that corresponds to the audio signal; dividing the current image into at least one block; obtaining motion information of the at least one block; generating index information including information for giving a three-dimensional (3D) effect in at least one direction to the at least one audio object, based on the motion information of the at least one block; and processing the audio object, in order to give the 3D effect in the at least one direction to the audio object, based on the index information.

TRANSMISSION DEVICE, TRANSMISSION METHOD, RECEPTION DEVICE, AND RECEPTION METHOD
20230230601 · 2023-07-20 · ·

A processing load at a receiving side is reduced in a case where a plurality of classes of audio data is transmitted. A predetermined number of audio streams including coded data of a plurality of groups is generated and a container of a predetermined format having this predetermined number of audio streams is transmitted. Command information for creating a command specifying a group to be decoded from among the plurality of groups is inserted into the container and/or the audio stream. For example, a command insertion area for the receiving side to insert a command for specifying a group to be decoded is provided in at least one audio stream among the predetermined number of audio streams.

Inter-channel bandwidth extension spectral mapping and adjustment

A method includes generating a synthesized non-reference high-band channel based on a non-reference high-band excitation corresponding to a non-reference target channel. The method further includes estimating one or more spectral mapping parameters based on the synthesized non-reference high-band channel and a high-band portion of the non-reference target channel. The method also includes applying the one or more spectral mapping parameters to the synthesized non-reference high-band channel to generate a spectrally shaped synthesized non-reference high-band channel. The method further includes generating an encoded bitstream based on the one or more spectral mapping parameters and the spectrally shaped synthesized non-reference high-band channel.

Inter-channel bandwidth extension spectral mapping and adjustment

A method includes generating a synthesized non-reference high-band channel based on a non-reference high-band excitation corresponding to a non-reference target channel. The method further includes estimating one or more spectral mapping parameters based on the synthesized non-reference high-band channel and a high-band portion of the non-reference target channel. The method also includes applying the one or more spectral mapping parameters to the synthesized non-reference high-band channel to generate a spectrally shaped synthesized non-reference high-band channel. The method further includes generating an encoded bitstream based on the one or more spectral mapping parameters and the spectrally shaped synthesized non-reference high-band channel.

Headtracking for pre-rendered binaural audio

A system and method of modifying a binaural signal using headtracking information. The system calculates a delay, a first filter response, and a second filter response, and applies these to the left and right components of the binaural signal according to the headtracking information. The system may also apply headtracking to parametric binaural signals. In this manner, headtracking may be applied to pre-rendered binaural audio.

System and method for an audio reproduction device

System and method for enhancing audio reproduced by an audio reproduction device with a first channel and second channel is described. X samples of audio signals are received and stored in a portion of an input buffer with 2x positions and rest of the x positions are padded with zero for both the channels. Contents of the input buffer are transformed to frequency domain (FD) components. FD components are multiplied with a first filter coefficient to generate FD components with short echo effect and with a second filter coefficient to generate FD components with long echo effect. Then, they are converted to time domain (TD) components with short echo effect and TD components with long echo effect. Selective TD components with short echo effect and long echo effect are combined to generate a convolved first channel output and a convolved second channel output.