Abstract:
A method for processing speech signals prior to encoding a digital signal comprising audio data includes selecting frequency domain coding or time domain coding based on a coding bit rate to be used for coding the digital signal and a short pitch lag detection of the digital signal.
Abstract:
A method of decoding an encoded audio bitstream and generating frequency bandwidth extension is disclosed. The method includes decoding the audio bitstream to produce a decoded low band audio signal and generate a low band excitation spectrum corresponding to a low frequency band. A sub-band area is identified within the low frequency band using a parameter which indicates energy information of a low band spectral envelope. A high band excitation spectrum is generated for a high frequency band by copying a sub-band excitation spectrum from the identified sub-band area to a high sub-band area corresponding to the high frequency band. Using the generated high band excitation spectrum, an extended high band audio signal is generated by applying a high band spectral envelope. The extended high band audio signal is added to the decoded low band audio signal to generate an audio output signal having an extended frequency bandwidth.
Abstract:
The quality of encoded signals can be improved by reclassifying AUDIO signals carrying non-speech data as VOICE signals when periodicity parameters of the signal satisfy one or more criteria. In some embodiments, only low or medium bit rate signals are considered for re-classification. The periodicity parameters can include any characteristic or set of characteristics indicative of periodicity. For example, the periodicity parameter may include pitch differences between subframes in the audio signal, a normalized pitch correlation for one or more subframes, an average normalized pitch correlation for the audio signal, or combinations thereof. Audio signals which are re-classified as VOICED signals may be encoded in the time-domain, while audio signals that remain classified as AUDIO signals may be encoded in the frequency-domain.
Abstract:
In accordance with an embodiment of the present invention, a method for speech processing includes determining an unvoicing/voicing parameter reflecting a characteristic of unvoiced/voicing speech in a current frame of a speech signal comprising a plurality of frames. A smoothed unvoicing/voicing parameter is determined to include information of the unvoicing/voicing parameter in a frame prior to the current frame of the speech signal. A difference between the unvoicing/voicing parameter and the smoothed unvoicing/voicing parameter is computed. The method further includes generating an unvoiced/voiced decision point for determining whether the current frame comprises unvoiced speech or voiced speech using the computed difference as a decision parameter.
Abstract:
MDCT or FFT-based audio coding algorithms often have the problem named here spectral pre-echoes when coding an energy attack signal. This invention presents several possibilities to avoid the spectral pre-echoes existing in decoded signal segment before the energy attack point. The spectral envelope before the attack point can be improved by performing spectrum smoothing, replacing the segment of having spectral pre-echoes or filtering the segment with a combined filter obtained by doing LPC analysis.
Abstract:
MDCT or FFT-based audio coding algorithms often have the problem named here spectral pre-echoes when coding an energy attack signal. This invention presents several possibilities to avoid the spectral pre-echoes existing in decoded signal segment before the energy attack point. The spectral envelope before the attack point can be improved by performing spectrum smoothing, replacing the segment of having spectral pre-echoes or filtering the segment with a combined filter obtained by doing LPC analysis.
Abstract:
The present disclosure relates to a method, apparatus, and system for encoding and decoding signals. The encoding method includes: converting a first-domain signal into a second-domain signal; performing Linear Prediction (LP) processing and Long-Term Prediction (LTP) processing for the second-domain signal; obtaining a long-term flag value according to a decision criterion; obtaining a second-domain predictive signal according to the LP processing result and the LTP processing result when the long-term flag value is a first value; obtaining a second-domain predictive signal according to the LP processing result when the long-term flag value is a second value; converting the second-domain predictive signal into a first-domain predictive signal, and calculating a first-domain predictive residual signal; and outputting a bit stream that includes the first-domain predictive residual signal.
Abstract:
In accordance with an embodiment, a method of decoding an audio/speech signal includes decoding an excitation signal based on an incoming audio/speech information, determining a stability of a high frequency portion of the excitation signal, smoothing an energy of the high frequency portion of the excitation signal based on the stability of the high frequency portion of the excitation signal, and producing an audio signal based on smoothing the high frequency portion of the excitation signal.
Abstract:
In accordance with an embodiment, a method of encoding an audio/speech signal includes determining a mixed codebook vector based on an incoming audio/speech signal, where the mixed codebook vector includes a sum of a first codebook entry from a first codebook and a second codebook entry from a second codebook. The method further includes generating an encoded audio signal based on the determined mixed codebook vector, and transmitting a coded excitation index of the determined mixed codebook vector.
Abstract:
A file generation method includes obtaining first information based on a resource index file, where the resource index file comprises indexes, file names, and first path information of a plurality of resource files, where the resource files include a first resource file and a second resource file, where the first information includes second path information and a file name of a first file; loading the first file based on the first information; obtaining the first resource file and the second resource file from the first file based on second information; and displaying the first page based on the first resource file and the second resource file.