290 likes | 305 Views
Survey of Robust Techniques. 2005/5/19 Presented by Chen-Wei Liu. Conferences. Log-Energy Dynamic Range Normalization for Robust Speech Recognition Weizhong Zhu and Douglas O’Shaughnessy INRS-EMT, University of Quebec, Canada ICASSP 2005
E N D
Survey of Robust Techniques 2005/5/19 Presented by Chen-Wei Liu
Conferences • Log-Energy Dynamic Range Normalization for Robust Speech Recognition • Weizhong Zhu and Douglas O’Shaughnessy • INRS-EMT, University of Quebec, Canada • ICASSP 2005 • Static and Dynamic Spectral Features : Their Noise Robustness and Optimal Weights for ASR • Chen Yang, Tan Lee, The Chinese University of Hong Kong • Frank K. Soong, ATR, Kyoto, Japan • ICASSP 2005
Introduction • Methods of robust speech recognition can be classified into two approaches • Front-end processing for speech feature extraction • Back-end processing for HMM decoding • Compensation for noise in • The front-end processing method • is to suppress the noise and get more robust parameters • The back-end processing • is to compensate for noise and adapt the parameters inside the HMM system • This paper focus on the first approach
Introduction • Comparing with cepstral coefficients, the log-energy feature has quite different characteristics • Log of summation of energy of all samples in one frame • This paper tries to find a more effective way, named log-energy dynamic range normalization (ERN), to remove the effects of additive noise • By minimizing mismatch between training and testing data
Energy Dynamic Range Normalization • Observations • Elevated minimum value • Valleys are buried by additive noise energy, while peaks are not affected as much
Energy Dynamic Range Normalization • The larger difference on valleys leads to a mismatch between the clean and noisy speech • To minimize the mismatch, this paper suggests an algorithm to scale the log-energy feature sequence of clean speech • In which it lifts valleys while it keeps peaks unchanged • Log-energy dynamic range is defined as follows
Energy Dynamic Range Normalization • In the presence of noise, is affected by additive noise, while is not affected as much • Let and target energy dynamic range as X ;then the above equation becomes • Is this way, it can use to set the target minimum value based on a given target dynamic range
Energy Dynamic Range Normalization • The following are the steps of the proposed log-energy feature dynamic range normalization algorithm • 1st : find Max = and Min = • 2nd : calculate target • 3rd : if then 4th • 4th : for i=1…n,
Energy Dynamic Range Normalization • The scaling effect is decreased as its own value goes up and the maximum of the sequence is unchanged
Experimental ResultsLinear Scaling • The proposed method was evaluated on the Aurora 2.0
Experimental ResultsNonLinear Scaling • Using non-linear scaling of equation as follows
Experimental ResultsComparison of Linear Scaling and Non-linear Scaling • Performance comparisons at different SNR levels are shown as follows
Conclusions • When systems were trained on a clean speech training set, the proposed technique can have overall about a 30.83% relative performance • Like CMS, the proposed method does not require any prior knowledge of noise and level • Reducing mismatch in log-energy leads to a large recognition improvement
Conferences • Log-Energy Dynamic Range Normalization for Robust Speech Recognition • Weizhong Zhu and Douglas O’Shaughnessy • INRS-EMT, University of Quebec, Canada • ICASSP 2005 • Static and Dynamic Spectral Features : Their Noise Robustness and Optimal Weights for ASR • Chen Yang, Tan Lee, The Chinese University of Hong Kong • Frank K. Soong, ATR, Kyoto, Japan • ICASSP 2005
Introduction • Dynamic cepstral features can help static features to characterize the speech trajectory on its time varying rate • It has been shown that such a representation yields higher speech and speaker recognition performance than static cepstra only • This paper tries to quantify the robustness of static and dynamic features under different types of noise and variable SNRs
Noise Robustness AnalysisRecognition with only Static or Dynamic Features
Noise Robustness AnalysisStatic and Dynamic Cepstral Distances between Clean and Noisy Speech • For a given sequence of noisy speech observation, the output likelihood is presented as follows by single Gaussians for simplicity : • The mismatch between clean and noisy conditions lies mainly on the exponent term which can be re-written as : Sequence of noisy speech observation A B A B Expected value is zero
Noise Robustness AnalysisStatic and Dynamic Cepstral Distances between Clean and Noisy Speech • Since the expected value of the second term is zero, the difference of likelihood between noisy and clean speech is just the first term, measured by defining a cepstral distance as follows : • Where is used to approximate the diagonal covariance, in the clean speech model • denotes the time average over the whole utterance
Noise Robustness AnalysisStatic and Dynamic Cepstral Distances between Clean and Noisy Speech • The weighted distances between clean and noisy speech for both the static and dynamic features, respectively: • Where the superscripts d and s denote the dynamic and the static features
Noise Robustness AnalysisStatic and Dynamic Cepstral Distances between Clean and Noisy Speech • The following depicts the scatter diagrams of dynamic distance (between clean and noisy dynamic cepstra) vs. its static counterpart :
Noise Robustness AnalysisStatic and Dynamic Cepstral Distances between Clean and Noisy Speech • Two observations can be made on the figure • Both distances are larger for increasingly mismatched conditions at lower SNRs • Majority points fall below the diagonal line. In other words, the dynamic cepstral distance between noisy and clean features is smaller than its static counterpart
Exponential Weighting in DecodingExponential Weightings • Based on the findings in the previous figure • It would make sense to weight the log likelihoods of the static and dynamic features differently in decoding to exploit their uneven noise robustness • The output likelihood of an observation can be split into two separate corresponding terms, d and s, as : • The acoustic likelihood components can be computed with different exponential weightings as :
Exponential Weighting in DecodingRecognition with Bracketed Weightings • Testing by bracketing the two weights at a step of 0.1 with the constraint of unity sum
Exponential Weighting in DecodingDiscriminative Weight Training (Weight Optimization) • The log likelihood difference (lld) between the recognized and the correct states is chosen as the objective function for optimization • For the u-th speech utterance of T observations • The lld is as follows : • The cost averaged over the whole training set of U utterances is :
Exponential Weighting in DecodingDiscriminative Weight Training (Weight Optimization) • This cost is minimized by adjusting iteratively the dynamic weight and the static weight, via the steepest descent as :
Experimental ResultsEvaluation on Aurora2.0 Database • Overall, a 36.6% relative WER reduction is obtained :
Experimental ResultsEvaluation on CUDIGIT Database • The relative WER improvement is 41.9%, averaged over all noise conditions
Conclusions • The dynamic features were found to be more resilient to additive noise interference than their static counterpart • Optimal exponential weights for exploiting the unequal robustness of the two cepstral features were used, and better performance were obtained