H04S2400/15

PROCESSING DEVICE AND PROCESSING METHOD
20230045207 · 2023-02-09 ·

A processing device according to this embodiment includes: a frequency characteristics acquisition unit configured to acquire frequency characteristics of an input signal; an extreme value extraction unit configured to extract an extreme value of spectral data; a kurtosis calculation unit configured to: calculate an evaluation value from spectral data; and calculate a kurtosis of a peak or a dip based on a plurality of evaluation values calculated by changing a calculation width, the evaluation value being used for evaluating the peak or the dip corresponding to the extreme value; a determination unit configured to determine whether to suppress the peak or the dip according to a comparison result between the kurtosis and a threshold value; and a suppression unit configured to suppress the peak or the dip with the extreme value that is determined to be suppressed.

PROCESSING DEVICE AND PROCESSING METHOD
20230040821 · 2023-02-09 ·

A processing device according to an embodiment includes: a frequency characteristics acquisition unit configured to acquire frequency characteristics of at least one sound pickup signal; a smoothing processing unit configured to perform smoothing processing so as to generate second spectral data smoother than first spectral data based on the frequency characteristics; a first compression unit configured to calculate a first difference value corresponding to a difference between the second spectral data and the first spectral data in a first band, and to compress the second spectral data based on the first difference value; and a filter generation unit configured to generate a filter, based on the second spectral data.

Discrete binaural spatialization of sound sources on two audio channels

Embodiments relate to binaural spatialization of more than two sound sources on two audio channels of an audio system. Sound signals each emitted from a corresponding sound source are collected, and a respective virtual position within an angular range of a sound scene is assigned to each sound source. Multi-source audio signals are generated by panning each sound signal according to the respective virtual position. A first multi-source audio signal is spatialized to a first direction to generate a first left signal and a first right signal. A second multi-source audio signal is spatialized to a second direction to generate a second left signal and a second right signal. A binaural signal is generated using the first left signal, the second left signal, the first right signal, and the second right signal. The binaural signal is such that each sound source appears to originate from its respective virtual position.

Wireless Sound Tranmission System and Method
20180006752 · 2018-01-04 ·

A method for providing sound to at least one user, in which audio signals are captured and transformed into audio data that is transmitted to at least one receiver unit; audio signals are generated from the received audio data and the hearing of the user(s) stimulated thereby; wherein the audio data is transmitted as audio data packets in separate slots of a TDMA frame structure, wherein the transmission unit and the receiver unit(s) are synchronized to form a wireless network, wherein each TDMA frame structure has at least one listening slot during which the synchronized network members do not transmit data and at least one network members listens, and wherein control data is transmitted from an external control device according to a sequence pattern selected according to the duration and periodicity of the listening slot(s) to be received by the at least one synchronized network member during said listening slot(s).

Spatial audio processing

An apparatus comprising at least one processor and at least one memory, the memory comprising machine-readable instructions, that when executed cause the apparatus to: store in a non-volatile memory multiple sets of predetermined spatial audio processing parameters for differently moving sound sources; provide in a man machine interface an option for a user to select one of the stored multiple sets of predetermined spatial audio processing parameters for differently moving sound sources; and in response to the user selecting one of the stored multiple sets of predetermined spatial audio processing parameters for differently moving sound sources, the apparatus is further caused to use the selected one of the stored multiple sets of predetermined spatial audio processing parameters to spatially process audio from one or more sound sources.

Controlling Audio Signal Parameters

A method and corresponding system for correcting for deviations in a performance that includes a plurality of audio sources, the method comprising detecting a parameter relating to an audio source, determining if the parameter deviates from a predetermined characteristic and in response to it being determined that the parameter deviates from the predetermined characteristic, causing display of a user interface configured to control the parameter, to allow a user to correct the deviation.

Audio Volume Handling

Apparatus is configured to associate each of one or more spatially-distributed audio sources in a virtual space, each audio source providing one or more audio signals representing audio for playback through a user device, with a respective fade-in profile which defines how audio volume for the audio source is gradually increased from a minimum level to a target volume level as a function of time. It is configured also to identify, based on user position, a current field-of-view within a virtual space and, in response to detecting that one or more new audio sources have a predetermined relationship with respect to the current field-of-view, fading-in the audio from the or each new audio source according to the fade-in profile for the respective audio source so as to increase their volume gradually towards the target volume level defined by the fade-in profile.

METHODS AND APPARATUSES RELATING TO AN ESTIMATED POSITION OF AN AUDIO CAPTURE DEVICE
20180007486 · 2018-01-04 ·

This specification describes a method comprising determining whether an estimated position of an audio capture device which captures audio data is within boundaries of a predetermined area, and in response to a determination that the estimated position is not within the boundaries of the predetermined area, associating the captured audio data with an adjusted position.

ADAPTIVE AUDIO CONSTRUCTION

Described herein is a method for creating an object-based audio signal from an audio input, the audio input including one or more audio channels that are recorded to collectively define an audio scene. The one or more audio channels are captured from a respective one or more spatially separated microphones disposed in a stable spatial configuration. The method includes the steps of: a) receiving the audio input; b) performing spatial analysis on the one or more audio channels to identify one or more audio objects within the audio scene; c) determining contextual information relating to the one or more audio objects; d) defining respective audio streams including audio data relating to at least one of the identified one or more audio objects; and e) outputting an object-based audio signal including the audio streams and the contextual information.

METHODS AND SYSTEMS FOR GENERATING AND RENDERING OBJECT BASED AUDIO WITH CONDITIONAL RENDERING METADATA

Methods and audio processing units for generating an object based audio program including conditional rendering metadata corresponding to at least one object channel of the program, where the conditional rendering metadata is indicative of at least one rendering constraint, based on playback speaker array configuration, which applies to each corresponding object channel, and methods for rendering audio content determined by such a program, including by rendering content of at least one audio channel of the program in a manner compliant with each applicable rendering constraint in response to at least some of the conditional rendering metadata. Rendering of a selected mix of content of the program may provide an immersive experience.