Patent classifications
H04S1/007
Spatial audio navigation
Methods and apparatus for spatial audio navigation that may, for example, be implemented by mobile multipurpose devices. A spatial audio navigation system provides navigational information in audio form to direct users to target locations. The system uses directionality of audio played through a binaural audio device to provide navigational cues to the user. A current location, target location, and map information may be input to pathfinding algorithms to determine a real world path between the user's current location and the target location. The system may then use directional audio played through a headset to guide the user on the path from the current location to the target location. The system may implement one or more of several different spatial audio navigation methods to direct a user when following a path using spatial audio-based cues.
Binaural Sound in Visual Entertainment Media
A method provides binaural sound to a listener while the listener watches a movie so sounds from the movie localize to a location of a character in the movie. Sound is convolved with head related transfer functions (HRTFs) of the listener, and the convolved sound is provided to the listener who wears a wearable electronic device.
NOISE SUPPRESSION USING TANDEM NETWORKS
A device includes a memory configured to store instructions and one or more processors configured to execute the instructions. The one or more processors are configured to execute the instructions to receive audio data including a first audio frame corresponding to a first output of a first microphone and a second audio frame corresponding to a second output of a second microphone. The one or more processors are also configured to execute the instructions to provide the audio data to a first noise-suppression network and a second noise-suppression network. The first noise-suppression network is configured to generate a first noise-suppressed audio frame and the second noise-suppression network is configured to generate a second noise-suppressed audio frame. The one or more processors are further configured to execute the instructions to provide the noise-suppressed audio frames to an attention-pooling network. The attention-pooling network is configured to generate an output noise-suppressed audio frame.
COLORLESS GENERATION OF ELEVATION PERCEPTUAL CUES USING ALL-PASS FILTER NETWORKS
A system includes one or more computing devices that encode spatial perceptual cues into a monaural channel to generate a plurality of output channels. A computing device determines a target amplitude response for the mid and side channels of the plurality of output channels, defining a spatial perceptual associated with one or more frequency-dependent phase shifts. The computing device determines a transfer function of a single-input, multi-output allpass filter based on the target amplitude response and determines coefficients of the allpass filter based on the transfer function, and processes the monaural channel with the coefficients of the allpass filter to generate the plurality of channels having the encoded spatial perceptual cues. The allpass filter is configured to be colorless with respect to the individual output channels, allowing for the placement of spatial cues into the audio stream to be decoupled from the overall coloration of the audio.
ARRANGEMENT FOR DISTRIBUTING HEAD RELATED TRANSFER FUNCTION FILTERS
Arrangement for distributing head related transfer function filters. In the arrangement a user device sends a request for a head related transfer function filter to the service being used. The service verifies if the user of the device has a subscription for a head related transfer function filters in the service being used and retrieves a filter as a response to a positive verification result. The service may filter audio channels and transmit filtered audio further. In an alternative embodiment the service transmits the filter to the user device for filtering the audio.
AUTOMATED MIXING OF AUDIO DESCRIPTION
A computer-implemented method of audio processing, the method comprising: receiving audio object data and audio description data, wherein the audio object data includes a first plurality of audio objects; calculating a long-term loudness of the audio object data and a long- term loudness of the audio description data; calculating a plurality of short-term loudnesses of the audio object data and a plurality of short-term loudnesses of the audio description data; reading a first plurality of mixing parameters that correspond to the audio object data; generating a second plurality of mixing parameters based on the first plurality of mixing parameters, the long-term loudness of the audio object data, the long-term loudness of the audio description data, the plurality of short-term loudnesses of the audio object data, and the plurality of short-term loudnesses of the audio description data; generating a gain adjustment visualization corresponding to the second plurality of mixing parameters, the audio object data and the audio description data; and generating mixed audio object data by mixing the audio object data and the audio description data according to the second plurality of mixing parameters, wherein the mixed audio object data includes a second plurality of audio objects, wherein the second plurality of audio objects correspond to the first plurality of audio objects mixed with the audio description data according to the second plurality of mixing parameters.
PERCEPTUAL OPTIMIZATION OF MAGNITUDE AND PHASE FOR TIME-FREQUENCY AND SOFTMASK SOURCE SEPARATION SYSTEMS
A method comprises: obtaining softmask values for frequency bins of time-frequency tiles representing an audio signal; reducing, or expanding and limiting, the softmask values; and applying the reduced, or expanded and limited, softmask values to the frequency bins to create a time-frequency representation of an estimated target source. An alternative method comprises, for each time-frequency tile: obtaining softmask values; applying the softmask values to the frequency bins to create a time-frequency domain representation of an estimated target source; obtaining a panning parameter and a source concentration estimates for the target source; determining, using the panning parameter estimate and the softmask values, a magnitude for the time-frequency representation of the estimated target source; determining, using the panning parameter estimate and the source phase concentration estimate, a phase for the time-frequency representation of the estimated target source; and combining the magnitude and the phase.
A SOUND OUTPUT UNIT AND A METHOD OF OPERATING IT
A system and a method for outputting sound where one or more first sound outputting units are identified and other, second, sound outputting units define a sound delay in accordance with the relative positions between each sound outputting unit and the first sound outputting unit(s). Microphones may be added to e.g. determine the amount and positions of persons.
The system may also use Intelligent cameras to determine number of people their position the room face direction and age distribution in order optimize audio level and equalisation of the frequency response—like in a church with many elder people or in a young audience at a live concert.
TEST INSTRUMENT FOR 5G CPE ANTENNA INSTALLATION
Disclosed are a test instrument and testing methods for audibly providing signal metrics (such as signal strength and/or signal quality) of fifth-generation network (5G) beams to assist installation of 5G Customer Premises Equipment (CPE) antenna at a premises. A test instrument may obtain signal metrics and provide audio output based on the signal metrics at various locations of the premises. The audio output may be transmitted to a headphone device worn by a user. In this manner, the user may select an appropriate location on the premises at which to install the 5G CPE antenna via audible queues that are based on the measured signal metric at a given location. The test instrument may provide fine-tuning capabilities by also audibly providing directional information that indicates where the 5G CPE antenna should be pointed or moved to align the 5G CPE antenna to a 5G beam.
Inter-channel bandwidth extension spectral mapping and adjustment
A method includes generating a synthesized non-reference high-band channel based on a non-reference high-band excitation corresponding to a non-reference target channel. The method further includes estimating one or more spectral mapping parameters based on the synthesized non-reference high-band channel and a high-band portion of the non-reference target channel. The method also includes applying the one or more spectral mapping parameters to the synthesized non-reference high-band channel to generate a spectrally shaped synthesized non-reference high-band channel. The method further includes generating an encoded bitstream based on the one or more spectral mapping parameters and the spectrally shaped synthesized non-reference high-band channel.