Patent classifications
H04S2420/13
SYSTEM AND METHOD FOR ADAPTIVE AUDIO SIGNAL GENERATION, CODING AND RENDERING
Embodiments are described for an adaptive audio system that processes audio data comprising a number of independent monophonic audio streams. One or more of the streams has associated with it metadata that specifies whether the stream is a channel-based or object-based stream. Channel-based streams have rendering information encoded by means of channel name; and the object-based streams have location information encoded through location expressions encoded in the associated metadata. A codec packages the independent audio streams into a single serial bitstream that contains all of the audio data. This configuration allows for the sound to be rendered according to an allocentric frame of reference, in which the rendering location of a sound is based on the characteristics of the playback environment (e.g., room size, shape, etc.) to correspond to the mixer's intent. The object position metadata contains the appropriate allocentric frame of reference information required to play the sound correctly using the available speaker positions in a room that is set up to play the adaptive audio content.
Apparatus, system, method and computer program for distributing announcement messages
An apparatus comprising circuitry configured to generate one or more focused sound sources as virtual loudspeakers of an announcement system.
Rendering audio objects having apparent size
Methods, systems, and computer program products for rending an audio object having an apparent size are disclosed. An audio processing system receives audio panning data including a first grid mapping first virtual sound sources in a space and speaker positions to speaker gains. The first grid specifies first speaker gains of the first virtual sound sources in the space. The audio processing system determines a second grid of second virtual sound sources in the space, including mapping the first virtual sound sources into the second virtual sound sources of the second virtual sources. The audio processing system selects at least one of the first grid or second grid for rendering an audio object based on an apparent size of the audio object. The audio processing system renders the audio object based on the selected grid or grids.
SIGNAL PROCESSING APPARATUS AND METHOD, AND PROGRAM
The present technology relates to a signal processing apparatus and method, and a program that can easily make a leaking sound difficult to hear.
A signal processing apparatus includes a masking sound generation unit that, in a case where a first content is reproduced in a first region and a second content is reproduced in a second region by wave field synthesis using a speaker array, generates a masking sound for masking a sound of the first content and a sound of the second content heard in a region between the first region and the second region. The present technology can be applied to content reproduction systems.
INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING SYSTEM, INFORMATION PROCESSING METHOD, AND PROGRAM
An information processing apparatus, an information processing system, an information processing method, and a program according to the present disclosure each include a sound-source position determining section (131) and an output control section (133). The sound-source position determining section (131) determines a position of a virtual sound source that is generated by a speaker array (20) and produces a spherical wave. The output control section (133) controls an output from the speaker array (20) to cause the virtual sound source to be located at the position. The virtual sound source outputs voice information corresponding to visual information that is information shown on a display (30).
METHOD FOR PROVIDING A SPATIALIZED SOUNDFIELD
A signal processing system and method for delivering spatialized sound by optimizing sound waveforms from a sparse array of speakers to the ears of a user. The system can provide listening areas within a room or space, to provide spatialization sounds to create a 3D audio effect. In a binaural mode, a binary speaker array provides targeted beams aimed towards a user's ears.
Audio Processing Apparatus and Method Therefor
An audio processing apparatus comprises a receiver (705) which receives audio data including audio components and render configuration data including audio transducer position data for a set of audio transducers (703). A renderer (707) generating audio transducer signals for the set of audio transducers from the audio data. The renderer (7010) is capable of rendering audio components in accordance with a plurality of rendering modes. A render controller (709) selects the rendering modes for the renderer (707) from the plurality of rendering modes based on the audio transducer position data. The renderer (707) can employ different rendering modes for different subsets of the set of audio transducers the render controller (709) can independently select rendering modes for each of the different subsets of the set of audio transducers (703). The render controller (709) can select the rendering mode for a first audio transducer of the set of audio transducers (703) in response to a position of the first audio transducer relative to a predetermined position for the audio transducer. The approach may provide improved adaptation, e.g. to scenarios where most speakers are at desired positions whereas a subset deviate from the desired position(s).
Methods and apparatus for compressing and decompressing a higher order ambisonics representation
Higher Order Ambisonics represents three-dimensional sound independent of a specific loudspeaker set-up. However, transmission of an HOA representation results in a very high bit rate. Therefore, compression with a fixed number of channels is used, in which directional and ambient signal components are processed differently. The ambient HOA component is represented by a minimum number of HOA coefficient sequences. The remaining channels contain either directional signals or additional coefficient sequences of the ambient HOA component, depending on what will result in optimum perceptual quality. This processing can change on a frame-by-frame basis.
Audio Processing Apparatus and Method Therefor
An audio processing apparatus comprises a receiver (705) which receives audio data including audio components and render configuration data including audio transducer position data for a set of audio transducers (703). A renderer (707) generating audio transducer signals for the set of audio transducers from the audio data. The renderer (7010) is capable of rendering audio components in accordance with a plurality of rendering modes. A render controller (709) selects the rendering modes for the renderer (707) from the plurality of rendering modes based on the audio transducer position data. The renderer (707) can employ different rendering modes for different subsets of the set of audio transducers the render controller (709) can independently select rendering modes for each of the different subsets of the set of audio transducers (703). The render controller (709) can select the rendering mode for a first audio transducer of the set of audio transducers (703) in response to a position of the first audio transducer relative to a predetermined position for the audio transducer. The approach may provide improved adaptation, e.g. to scenarios where most speakers are at desired positions whereas a subset deviate from the desired position(s).
SOUND IMAGE REPRODUCTION DEVICE, SOUND IMAGE REPRODUCTION METHOD, AND SOUND IMAGE REPRODUCTION PROGRAM
Provided is a sound image reproduction device, sound image reproduction method, and sound image reproduction program that can support monaural sound sources and is capable of imparting directivity to virtual sound sources in a space. An acoustic signal processing device (sound image reproduction device) 1 that generates virtual sound sources in a space using multiple loudspeakers arranged in a straight line, includes: a focal-point position determination unit 12 that determines the position of each virtual sound source to generate multiple virtual sound sources in a circular arrangement; a filter-coefficient determination unit 13 that calculates an impulse response vector for each loudspeaker by performing an inverse Fourier transform on a driving function for each loudspeaker that is used to generate a virtual sound source at the position of each virtual sound source and in which different weights are given to some of the virtual sound sources; and a convolution calculation unit 14 that calculates the convolution of one inputted acoustic signal with the impulse response vector for each loudspeaker and outputs each acoustic signal to the corresponding the multiple loudspeakers.