Patent classifications
G10H2210/301
METHOD AND SYSTEM FOR INSTRUMENT SEPARATING AND REPRODUCING FOR MIXTURE AUDIO SOURCE
A method and a system for instrument separating and reproducing for a mixture audio source is provided. The method and/or the system includes inputting selected music into an instrument separation model for extracting features therefrom, determining audio source signals of multiple channels for the separation of all instruments, each channel containing sound of one instrument, and transmitting the signals of the different channels to multiple speakers placed at designated positions for playing, which can reproduce or recreate an immersive sound field listening experience for users.
SYSTEM AND METHOD FOR 3D SOUND PLACEMENT
A phone app is disclosed that enables a user to place 3D sound in a room. The user of this app is able to locate precisely where sound is perceived to originate by aiming their phone. This app may be used by audio professionals in place of the controls on a traditional sound mixer.
Systems and Methods for Acoustic Simulation
Systems and methods for acoustic simulation in accordance with embodiments of the invention are illustrated. One embodiment includes a method for simulating acoustic responses, including obtaining a digital model of an object, calculating a plurality of vibrational modes of the object, conflating the plurality of vibrational modes into a plurality of chords, where each chord includes a subset of the plurality of vibrational modes, calculating, for each chord, a chord sound field in the time domain, where the chord sound field describes acoustic pressure surrounding the object when the object oscillates in accordance with the subset of the plurality of vibrational modes, deconflating each chord sound field into a plurality of modal sound fields, where each modal sound field describes acoustic pressure surrounding the object when the object oscillates in accordance with a single vibrational mode, and storing each modal sound field in a far-field acoustic transfer (FFAT) map.
Apparatuses and methods for audio classifying and processing
Apparatus and methods for audio classifying and processing are disclosed. In one embodiment, an audio processing apparatus includes an audio classifier for classifying an audio signal into at least one audio type in real time; an audio improving device for improving experience of audience; and an adjusting unit for adjusting at least one parameter of the audio improving device in a continuous manner based on the confidence value of the at least one audio type.
Facilitating inferential sound recognition based on patterns of sound primitives
The disclosed embodiments provide a system that performs a sound-recognition operation. During operation, the system recognizes a sequence of sound primitives in an audio stream, wherein a sound primitive is associated with a semantic label comprising one or more words that describe a sound characterized by the sound primitive. Next, the system feeds the sequence of sound primitives into a finite-state automaton that recognizes events associated with sequences of sound primitives. Finally, the system feeds the recognized events into an output system that generates an output associated with the recognized events to be displayed to a user.
Enhanced 3D audio/video processing apparatus and method
The enhanced 3D audio/video processing apparatus according to one embodiment of the present invention may comprise: a three-dimensional (3D) content generating unit for generating 3D content including video content and analog content; a depth information generating unit for generating depth information for the video frames constituting the video content; and a signal generating unit for generating a 3D enhanced signal including the generated 3D content and the depth information. Further, the enhanced 3D audio/video processing apparatus according to another embodiment of the present invention may comprise: a signal processing unit for processing the 3D enhanced signal including the 3D content including the video content and the audio content: a depth information extraction unit for acquiring the depth information of the video frames constituting the video content from the processed 3D enhanced signal; a 3D audio effect generating unit for generating 3D audio effect based on the acquired depth information; and a 3D audio content generating unit for generating 3D audio content by applying the generated 3D audio effect.
Systems and methods for identifying a sound event
Systems and methods for identifying a perceived sound event are provided. In one exemplary embodiment, the system includes an audio signal receiver, a processor, and an analyzer. The system deconstructs a received audio signal into a plurality of audio chunks, for which one or more sound identification characteristics are determined. One ore more distances of a distance vector are then calculated based on one or more of the sound identification characteristics. The distance vector can be a sound gene that serves as an identifier for the sound event. The distance vector for a received audio signal is compared to distance vectors of predefined sound events to identify the source of the received audio signal. A variety of other systems and methods related to sound identification are also provided.
Sound Signal Generation Method, Sound Signal Generation Device, Non-transitory Computer Readable Medium Storing Sound Signal Generation Program and Electronic Musical Apparatus
A sound signal generation method includes receiving a pitch and an intensity, and generating a sound signal corresponding to the pitch such that a size of a sound image of the sound signal is adjusted in accordance with the intensity.
Sound effect synthesis
Disclosed herein is a sound synthesis system for generating a user defined synthesised sound effect, the system comprising: a receiver of user defined inputs for defining a sound effect; a generator of control parameters in dependence on the received user defined inputs; a plurality of sound effect objects, wherein each sound effect object is arranged to generate a different class of sound and each sound effect object comprises a sound synthesis model arranged to generate a sound in dependence on one or more of the control parameters; a plurality of audio effect objects, wherein each audio effect object is arranged to receive a sound from one or more sound effect objects and/or one or more other audio effect objects, process the received sound in dependence on one or more of the control parameters and output the processed sound; a scene creation function arranged to receive sound output from one or more sound effect objects and/or audio effect objects and to generate a synthesised sound effect in dependence on the received sound; and an audio routing function arranged to determine the arrangement of audio effect objects, sound effect objects and scene creation function such that one or more sounds received by the scene creation function are dependent on the audio routing function; wherein the determined arrangement of audio effect objects, sound effect objects and the scene creation function by the audio routing function is dependent on the user defined inputs.
Automatic multi-channel music mix from multiple audio stems
There are disclosed automatic mixers and methods for creating a surround audio mix. A set of rules may be stored in a rule base. A rule engine may select a subset of the set of rules based, at least in part, on metadata associated with a plurality of stems. A mixing matrix may mix the plurality of stems in accordance with the selected subset of rules to provide three or more output channels.