Patent classifications
G10L19/0212
APPLICATION-SPECIFIC INTEGRATED CIRCUIT FOR ACCELERATING ENCODING AND DECODING, AND METHOD THEREFOR
An application-specific integrated circuit for accelerated encoding and decoding and a method, which are related to the technical field of Bluetooth mobile communication. The application-specific integrated circuit for accelerated encoding and decoding includes: a hardware accelerator, wherein the hardware accelerator includes a pre-processing and pronation processing module, which performs a pre-processing and pronation processing of data, a discrete Fourier transform module is used for performing a multi-level discrete Fourier transform, in an accelerated low-delay modified discrete cosine transform operation LD-MDCT and/or an accelerated the low-delay inverse modified discrete cosine transform operation LD-IMDCT. The application-specific integrated circuit for accelerated encoding and decoding and a method of the present invention adopts an ASIC application-specific integrated circuit, and adopts multi-level discrete Fourier transforms, so that the complex operations are completed by the ASIC application-specific integrated circuit.
PHASE RECONSTRUCTION IN A SPEECH DECODER
Innovations in phase quantization during speech encoding and phase reconstruction during speech decoding are described. For example, to encode a set of phase values, a speech encoder omits higher-frequency phase values and/or represents at least some of the phase values as a weighted sum of basis functions. Or, as another example, to decode a set of phase values, a speech decoder reconstructs at least some of the phase values using a weighted sum of basis functions and/or reconstructs lower-frequency phase values then uses at least some of the lower-frequency phase values to synthesize higher-frequency phase values. In many cases, the innovations improve the performance of a speech codec in low bitrate scenarios, even when encoded data is delivered over a network that suffers from insufficient bandwidth or transmission quality problems.
Apparatus and method realizing a fading of an MDCT spectrum to white noise prior to FDNS application
An apparatus for decoding an encoded audio signal to obtain a reconstructed audio signal includes a receiving interface for receiving one or more frames comprising information on a plurality of audio signal samples of an audio signal spectrum of the encoded audio signal, and a processor for generating the reconstructed audio signal. The processor is configured to generate the reconstructed audio signal by fading a modified spectrum to a target spectrum, if a current frame is not received by the receiving interface or if the current frame is received by the receiving interface but is corrupted, wherein the modified spectrum includes a plurality of modified signal samples, wherein, for each of the modified signal samples of the modified spectrum, an absolute value of the modified signal sample is equal to an absolute value of one of the audio signal samples of the audio signal spectrum.
MDCT-based complex prediction stereo coding
The invention provides methods and devices for stereo encoding and decoding using complex prediction in the frequency domain. In one embodiment, a decoding method, for obtaining an output stereo signal from an input stereo signal encoded by complex prediction coding and comprising first frequency-domain representations of two input channels, comprises the upmixing steps of: (i) computing a second frequency-domain representation of a first input channel; and (ii) computing an output channel on the basis of the first and second frequency-domain representations of the first input channel, the first frequency-domain representation of the second input channel and a complex prediction coefficient. The upmixing can be suspended responsive to control data.
Transform encoding/decoding of harmonic audio signals
An encoder for encoding frequency transform coefficients of a harmonic audio signal include the following elements: A peak locator configured to locate spectral peaks having magnitudes exceeding a predetermined frequency dependent threshold. A peak region encoder configured to encode peak regions including and surrounding the located peaks. A low-frequency set encoder configured to encode at least one low-frequency set of coefficients outside the peak regions and below a crossover frequency that depends on the number of bits used to encode the peak regions. A noise-floor gain encoder configured to encode a noise-floor gain of at least one high-frequency set of not yet encoded coefficients outside the peak regions.
AUDIO DECODER FOR INTERLEAVING SIGNALS
A method for decoding an encoded audio bitstream in an audio processing system is disclosed. The method includes extracting from the encoded audio bitstream a first waveform-coded signal comprising spectral coefficients corresponding to frequencies up to a first cross-over frequency for a time frame and performing parametric decoding at a second cross-over frequency for the time frame to generate a reconstructed signal. The second cross-over frequency is above the first cross-over frequency and the parametric decoding uses reconstruction parameters derived from the encoded audio bitstream to generate the reconstructed signal. The method also includes extracting from the encoded audio bitstream a second waveform-coded signal comprising spectral coefficients corresponding to a subset of frequencies above the first cross-over frequency for the time frame and interleaving the second waveform-coded signal with the reconstructed signal to produce an interleaved signal for the time frame.
Audio bandwidth extension by insertion of temporal pre-shaped noise in frequency domain
An audio decoder device for decoding a bitstream includes a bitstream receiver configured to receive the bitstream and to derive an encoded audio signal from the bitstream; a core decoder module configured for deriving a decoded audio signal in a time domain from the encoded audio signal; a temporal envelope generator configured to determine a temporal envelope of the decoded audio signal; a bandwidth extension module configured to produce a frequency domain bandwidth extension signal; a time-to-frequency converter configured to transform the decoded audio signal into a frequency domain decoded audio signal; a combiner configured to combine the frequency domain decoded audio signal and the frequency domain bandwidth extension signal in order to produce a bandwidth extended frequency domain audio signal; and a frequency-to-time converter configured to transform the bandwidth extended frequency domain audio signal into a bandwidth-extended time domain audio signal.
APPARATUS AND METHOD FOR SELECTING ONE OF A FIRST ENCODING ALGORITHM AND A SECOND ENCODING ALGORITHM USING HARMONICS REDUCTION
An apparatus for selecting one of a first encoding algorithm and a second encoding algorithm includes a filter configured to receive the audio signal, to reduce the amplitude of harmonics in the audio signal and to output a filtered version of the audio signal. First and second estimators are provided for estimating first and second quality measures in the form of SNRs of segmented SNRs associated with the first and second encoding algorithms without actually encoding and decoding the portion of the audio signal using the first and second encoding algorithms. A controller is provided for selecting the first encoding algorithm or the second encoding algorithm based on a comparison between the first quality measure and the second quality measure.
INTEGRATED SENSOR-ARRAY PROCESSOR
An integrated sensor-array processor and method includes sensor array time-domain input ports to receive sensor signals from time-domain sensors. A sensor transform engine (STE) creates sensor transform data from the sensor signals and applies sensor calibration adjustments. Transducer time-domain input ports receive time-domain transducer signals, and a transducer output transform engine (TTE) generates transducer output transform data from the transducer signals. A spatial filter engine (SFE) applies suppression coefficients to the sensor transform data, to suppress target signals received from noise locations and/or amplification locations. A blocking filter engine (BFE) applies subtraction coefficients to the sensor transform data, to subtract the target signals from the sensor transform data. A noise reduction filter engine (NRE) subtracts noise signals from the BFE output. An inverse transform engine (ITE) generates time-domain data from the NRE output.
Downscaled decoding
A downscaled version of an audio decoding procedure may more effectively and/or at improved compliance maintenance be achieved if the synthesis window used for downscaled audio decoding is a downsampled version of a reference synthesis window involved in the non-downscaled audio decoding procedure by downsampling by the downsampling factor by which the downsampled sampling rate and the original sampling rate deviate, and downsampled using a segmental interpolation in segments of ¼ of the frame length.