Patent classifications
H04S7/40
Spatial audio processing
An apparatus comprising at least one processor and at least one memory, the memory comprising machine-readable instructions, that when executed cause the apparatus to: store in a non-volatile memory multiple sets of predetermined spatial audio processing parameters for differently moving sound sources; provide in a man machine interface an option for a user to select one of the stored multiple sets of predetermined spatial audio processing parameters for differently moving sound sources; and in response to the user selecting one of the stored multiple sets of predetermined spatial audio processing parameters for differently moving sound sources, the apparatus is further caused to use the selected one of the stored multiple sets of predetermined spatial audio processing parameters to spatially process audio from one or more sound sources.
Controlling Audio Signal Parameters
A method and corresponding system for correcting for deviations in a performance that includes a plurality of audio sources, the method comprising detecting a parameter relating to an audio source, determining if the parameter deviates from a predetermined characteristic and in response to it being determined that the parameter deviates from the predetermined characteristic, causing display of a user interface configured to control the parameter, to allow a user to correct the deviation.
METHOD FOR PROCESSING SOUND ON BASIS OF IMAGE INFORMATION, AND CORRESPONDING DEVICE
A method of processing an audio signal including at least one audio object based on image information includes: obtaining the audio signal and a current image that corresponds to the audio signal; dividing the current image into at least one block; obtaining motion information of the at least one block; generating index information including information for giving a three-dimensional (3D) effect in at least one direction to the at least one audio object, based on the motion information of the at least one block; and processing the audio object, in order to give the 3D effect in the at least one direction to the audio object, based on the index information.
SERVICE FOR TARGETED CROWD SOURCED AUDIO FOR VIRTUAL INTERACTION
An audio generation system is provided to enable coordinated control of multiple IoT devices for audio collection and distribution of one or more audio sources according to location and user preference. The audio generation system enables a location sensitive acoustic control of sound, both as a shaped envelope for a particular source, and as an individualized experience. The audio generation system also facilitates an interactive visual system for visualization and manipulation of the audio environment including via the use of augmented reality and/or virtual reality to depict soundscapes. The audio generation system can also facilitate a system for improving and achieving an audio environment (sound influence zone) and an intuitive way to understand where sounds will be heard.
DYNAMIC VISUAL INTENSITY RENDERING
The present technology can provide a mechanism for adjusting a visual effect that is associated with an audio artifact at a given frequency bandwidth that is attenuated by speaker characteristics. The intensity of the visual effects that is adjusted can also be attributed to a change in volume settings of a processing device as well as an intensity of a multimedia skin in which the visual effect is encoded. The multimedia skin includes filters, transitions/animations, and/or image universal processing, that can be applied to any set of photos, videos, and/or songs, in order to create, in real-time, many variations of the same digital multimedia file, wherein each multimedia skin leads to a specific video rendering.
System and a processing method for customizing audio experience
The present disclosure relates to a system and a processing method in association with the system for customizing audio experience. Customization of audio experience can be based on derivation of at least one customized audio response characteristic which can be applied to an audio device used by a person. The customized audio response characteristic(s) can be unique to the person.
LOUDSPEAKER SYSTEM AND CONTROL
A system for providing an audio interface at a mobile device is provided. The mobile device includes an interface programmed detect a loudspeaker system. The mobile device presents, via a user interface, a display screen to receive user input of sweet-spot commands. The mobile devices send sweet-spot parameters to the loudspeaker system in response to the sweet-spot commands.
Systems and methods for videoconferencing with spatial audio
A system may provide for the generation of spatial audio for audiovisual conferences, video conferences, etc. (referred to herein simply as “conferences”). Spatial audio may include audio encoding and/or decoding techniques in which a sound source may be specified at a location, such as on a two-dimensional plane and/or within a three-dimensional field, and/or in which a direction or target for a given sound source may be specified. A conference participant's position within a conference user interface (“UI”) may be set as the source of sound associated with the conference participant, such that different conference participants may be associated with different sound source positions within the conference UI.
Acoustic neural network scene detection
An acoustic environment identification system is disclosed that can use neural networks to accurately identify environments. The acoustic environment identification system can use one or more convolutional neural networks to generate audio feature data. A recursive neural network can process the audio feature data to generate characterization data. The characterization data can be modified using a weighting system that weights signature data items. Classification neural networks can be used to generate a classification of an environment.
Information processing method, information processing device, and non-transitory storage medium
An information processing method includes: receiving first space information including a first coordinate system of one of a logical space or a physical space, and second space information including a second coordinate system of the other of the logical space or the physical space; receiving first sound localization information indicating a position where a sound image is to be localized in the first coordinate system; and transforming the first sound localization information into second sound localization information indicating a position where the sound image is to be localized in the second coordinate system.