검색 전체 메뉴
PDF
맨 위로
OA 학술지
Speech Processing System Using a Noise Reduction Neural Network Based on FFT Spectrums
  • 비영리 CC BY-NC
  • 비영리 CC BY-NC
ABSTRACT
Speech Processing System Using a Noise Reduction Neural Network Based on FFT Spectrums
KEYWORD
Speech processing , Neural network , Amplitude and phase spectrums , Background noise
  • I. INTRODUCTION

    In the field of noise processing, background noise is one of the important research problems. In particular, it has been noted that system reliability, and especially the speech recognition ratio, can be significantly decreased by background noise [1, 2]. Background noise includes various non-stationary noises existing in the real environment such as street noise and babble noise. Thus, typical background noises cannot be simply eliminated with a Wiener filter or spectral subtraction, but require more skillful techniques. Therefore, to solve the above-mentioned problems, this paper proposes a speech processing system using an auditory system and noise reduction neural network that is effective in various noisy environments.

    Several noise suppression and speech processing methods have been proposed, such as spectral subtraction [3-5], neural networks [6, 7], minimum mean-square error estimators [1, 8, 9], and a time-delay neural network [10]. The spectral subtraction method is used in signal processing to process speech adaptively according to the noise intensity, thereby enhancing the performance. For instance, in a study by Lim et al. [4], speech intelligibility was improved when choosing four different lengths of filter for the pitch period according to four different input signal-to-noise ratios (SNRs). In this paper, the parameters of lateral inhibition function are adjusted frame by frame. For details, the parameters are used to adjust the width and amplitude of the lateral inhibition filter at each frame, respectively [10, 11].

    Recently, developments of auditory processing systems have been reported in various studies imitating functions of signal processing in auditory processing systems [10-12], and one of these auditory mechanisms, called lateral inhibition, is used as the speech processing system in this paper. In the area of speech signal processing, a neural network (NN) is applied to speech recognition area, while in the area of speech enhancement and noise reduction, the major application of NNs is the extraction of speech sections from a noisy speech signal. Moreover, amplitude spectrums contain more information than phase spectrums when speech signals are generated by fast Fourier transform (FFT).

    In this paper, the proposed system reduces the noise of a speech signal by using the proposed NN based on FFT amplitude spectrums and phase spectrums, and implements auditory processing frame by frame according to the detected results after detected voiced sections and transitional sections for each frame. Thereafter, the proposed system restores the FFT amplitude and phase spectrums using an inverse fast Fourier transform (IFFT).

    To evaluate the proposed speech processing system, the SNR is measured as an indication of the speech intelligibility, and the results confirm the effectiveness of our solution in the case of white, car, restaurant, subway, and street noise.

    II. MODEL OF SPEECH AND NOISE

      >  A. Speech and Noise Database

    The speech data used in the experiment are included in the Aurora-2 database (DB) that consists of English-connected digits, recorded in clean environments with a sampling frequency of 8 kHz [13]. The Aurora-2 DB offers two different training modes, i.e., clean training and multi-conditional training modes. The clean training mode includes 8,440 utterances, which contains the voices of 55 male and 55 female adult recordings. The same 8,440 utterances are also used in the multi-conditional training mode. For each training mode, three test sets (sets A, B, and C) are provided. These clean utterances are artificially contaminated by adding different types of noise (subway, babble, car, etc.) to the clean utterances at different SNR levels. In test set A, four types of noise (subway, car, exhibition hall, and babble) are added to the clean utterances at SNR levels of -5, 0, 5, 10, 15, and 20 dB, while in test set B, another set of four different types of noise (restaurant, street, airport, and train station) are added to the clean utterances at the same SNR levels. In test set C, two types of noise (subway and street) used in sets A and B are added. The three test sets are composed of 4,004 utterances from 52 male and 52 female speakers [10, 11].

      >  B. Proposed Methods

    In this paper, a weighted spectral average filter is adopted to reduce unexpected peaks at each frame, as shown in Eq. (1) [10, 11].

    image

    In Eq. (1), the weighted values for the parameters of Wj are used as follows: W?2 = 0.5 , W?1 = 1.2 , W0 = 1.6 , W1 = 1.2, and W2 = 0.5 , where,

    image

    is the FFT spectral average of the ith frame, SiA(ω) is the original speech spectrums. Moreover, this study uses the function of spectral lateral inhibition [11, 12], thereby emphasizing the spectral peaks of speech by an excitation area, while simultaneously suppressing the noise spectral valleys by two inhibition areas, as shown in Fig. 1.

    The parameters for Fi= l,e,r show the amplitude of the impulse response as follows: Fe = 1.0 , Fl = ?0.7 and Fr = ?0.3 , where Fe represents the amplitude for an excitation area, F1 and Fr represent the amplitudes for two inhibition areas, α is used to adjust the whole amplitude for an excitation area and two inhibition areas, and the parameters for Bi =l,e,r show the width of the lateral inhibition function. For noise reduction, the parameters are restricted to satisfy the following Eq. (2). Therefore, the noise signal is reduced because the average sum values for the weighted noise is zero [10, 11].

    image

    In this paper, the parameters of Wj , Fi=l,e,r and Bi=l,e,r are experimentally determined to be the optimum values.

      >  C. Proposed Noise Reduction NN

    This paper proposes an algorithm using a noise reduction neural network (NRNN) that is constructed for the FFT amplitude and phase spectrum areas, allowing more effective correlation of the added information. In this experiment, a perceptron type NN is used and trained by a back-propagation (BP) algorithm, as shown in Fig. 2. The input and output for each unit approximates a sigmoid function, whose output range is from ?1.0 to 1.0, as given by Eq. (3) [10, 11].

    image

    In this equation, f(xj) is the output of unit j in the upper subnet, xi is the output of unit i in the lower subnet, wji is the connection weight between the input unit i and the output unit j, and θ is a threshold in each unit. The proposed NN is composed of three layers and the composition of the NN is 32-64-32. In this experiment, the number of units in the hidden layer was adjusted to 64, as this produced the best convergence of the NN.

    In order to essentially reduce the background noise, this paper proposes a noise processing algorithm using the BP training method of the NRNN system, which is composed of an NN based on amplitude and phase spectrums, as shown in Fig. 3.

    The proposed NN was trained using twenty types of noisy speech data selected randomly from test set A, and artificially added at five different SNRs (20 dB, 15 dB, 10 dB, 5 dB, and 0 dB) for four different noises (i.e., white, car, restaurant, and subway noise). Therefore, the proposed NN was trained using five kinds of networks: 1) input signal-to-noise ratio ( SNRIN ) = 20 dB; 2) SNRIN = 15 dB; 3) SNRIN = 10 dB; 4) SNRIN = 5 dB; and 5) SNRIN = 0 dB. Thus, a total of ten simulations were used for the training,for one network. In this experiment, the input signals of the NN are the zero to 31st samples (0 to 3.9 kHz) of the FFT amplitude spectrums. The target signals are samples of the FFT amplitude spectrums with a frame corresponding to a training signal for a clean speech signal. Therefore, the target signals for the NN are the zero to 31st samples of the FFT amplitude spectrums.

    In order to highlight the noise elimination performance by the NRNN system, car noise was added to the original clean signal "MAT_32O2A.08", which was spoken by a male speaker, for the tenth frame. Fig. 4 shows examples of the FFT amplitude spectrums for the target (clean speech), input (noisy speech), and output signal, while Fig. 5 shows the FFT phase spectrums for the target, input, and output signal, in the case of SNRIN = 5 dB. Therefore, these figures show that the background noise was significantly reduced when the proposed NRNN system was used.

    The training coefficient was set at 0.1 and the inertia coefficient set at 0.03 because these training coefficients produced the best convergence of the NN. The maximum number of training iterations was discontinued after 10,000 times because there was almost no decrease in the training error curves at the minimum error points. The error in Fig. 6 is an average value among ten trials for one network. Fig. 6 shows the training error curves for the NN based on the amplitude spectrums in the case of SNRIN = 5 dB.

      >  III. DESIGN OF PROPOSED SPEECH PROCESSING SYSTEM

    In order to consider the practical applications for noise reduction, we adopted the speech processing system as shown in Fig. 7. The proposed system mainly consists of the weighted spectral average filter for reducing unexpected peaks at frames, the NN for training two kinds of different patterns with the FFT amplitude and phase spectrums, and the lateral inhibition function for obtaining a significant auditory spectral representation. In Fig. 7, first, the noisy speech signal x(t), sampled at 8 kHz, is multiplied by a 64 sample Hamming window. After passing through the FFT, the Fourier-transformed signal is separated into its amplitude and phase spectrums. Next, the amplitude spectrums are delayed by 3 frames and averaged using the weighted spectral average filter for each frame. The output signals are added to the input signals for the NRNN system. In the case of the amplitude spectrums, the proposed system detects voiced sections in every frame where Rf ≥ Th , and transitional sections in every frame where Rf < Th . Here, Rf is the effective value obtained for each frame and Th is the threshold value which is determined experimentally. In this experiment, the maximum value for the threshold calculated at the start of seven sections was used [10, 11].

    At the output of the decision of the voiced and transitional sections block, there is a switch to send the FFT or log power spectrums in Fig. 7. When the detected results for each frame are voiced sections, we use FFT power spectrums to convolve with the lateral inhibition function, called algorithm I. When the detected results for each frame are transitional sections, we use log power spectrums and a spectrum smoothing method by using the average filter, that is, algorithm II. Some negative values after convolution do not contain any useful information in the present situation, and therefore are set at zero. In this experiment, the coefficients for the lateral inhibition function were "B = 6" and "? = 2.0" as shown in Fig. 1, where "B" shows the width of the lateral inhibition function and "? " shows the whole amplitude of the lateral inhibition function. After obtaining the amplitude and phase spectrums, the output of the speech signal is regenerated by an inverse fast Fourier transform (IFFT) [10, 11].

    IV. EXPERIMENTAL RESULTS AND CONSIDERATIONS

    Using the basic composition conditions described above, experiments confirmed that the proposed system was effective for speech degraded by additive car and street noise based on measuring the SNR. To evaluate the performance, noisy speech data for twenty different test utterances were randomly selected from test sets A, B, and C [10, 11].

    The proposed system was compared with a conventional spectral subtraction (SS) method [3] and the minimum mean-square error log-spectral amplitude (MMSE-LSA) [8] estimator method for car and street noise. The SS method is the classic algorithm that is used in speech enhancement and noise suppression. This method estimates the noise power spectrums from the magnitude spectrums of the noisy speech measured during non-speech activity and then subtracts the noise power spectrums from the speech power spectrums for each frame. The enhanced speech is reconstructed, through an inverse Fourier transform, from both the enhanced magnitude spectrums and the original phase spectrums [3]. When implementing the SS method in this experiment, Hamming-windowed frames that overlapped by 50% were used to reduce edge effects. Moreover, the MMSE-LSA estimator method is based on a minimum mean-square error short-time spectral amplitude (MMSE-STSA) estimator [9], which can be derived by modeling speech and noise spectrums as statistically independent Gaussian random variables. This MMSE-LSA amplitude estimator method is obtained by means of minimizing the mean-squared error of the log-power spectrums, and the gain function derived in Ephraim and Malah [8] is constructed to minimize the mean-square error estimates of log spectrums. When implementing these conventional methods, the frame length was 64 samples (8 ms) and the overlap was 32 samples (4 ms). At each frame, a Hamming window was used [10, 11].

    Figs. 8 and 9 show the averages of the approximate SNROUT values over the test utterances for the proposed system, as compared with the SS method and the MMSE-LSA estimator method at different noise levels ( IN SNR = 20 to 0 dB) for each noise. In the case of stationary noise, as shown in Fig. 8, the maximal improvement in the SNROUT values, with the proposed method, was approximately 3.0 dB, 4 dB, and 11.5 dB better for car noise, when compared with the MMSE-LSA, without lateral inhibition processing and NRNN (W-LIPN), and SS methods, respectively, where the W-LIPN shows the maximal improvement in the SNROUT values obtained by not using the lateral inhibition function or the NRNN system, in the proposed speech processing system of Fig. 7. Moreover, a similar tendency was found for non-stationary noise as shown in Fig. 9, i.e., the maximal improvement in the SNROUT values, with the proposed method, was approximately 2.5 dB, 3.5 dB, and 11.0 dB better for street noise, when compared with the MMSE-LSA, W-LIPN, and SS methods. The maximal improvement in the SNROUT values with the proposed method was approximately 10.0 dB better for car noise and 8.0 dB better for street noise, when compared with ? = 0.0. Thus, the 10.0 dB improvement in the SNROUT was quite significant, and was evident when listening to the output.

    V. CONCLUSIONS

    A speech processing system was proposed that uses a lateral inhibition mechanism model and noise reduction neural network based on amplitude and phase spectrums to reduce background noise. In summary, the experimental results were as follows: 1) The noise reduction with the function of lateral inhibition was different for car and street noise, and was especially remarkable for car noise; and 2)The noise reduction was significant under serious SNRIN conditions of up to about 0 dB for speech data.

    The proposed system using the function of lateral inhibition was experimentally demonstrated as an effective noise suppression system for white, car, restaurant, subway, and street noise. Therefore, it is believed that the present research results will be useful for noise suppression and speech enhancement under noisy conditions.

참고문헌
  • 1. Yu D., Deng L., Droppo J., Wu J., Gong Y., Acero A. 2008 "A minimum-mean-square-error noise reduction algorithm on Mel-frequency cepstra for robust speech recognition" [Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing] P.4041-4044 google
  • 2. Jeong, S., Yang, H., Hahn, M. 2008 "Two-channel noise reduction for robust speech recognition in car environments" [Electronics Letters] Vol.44 P.1042-1043 google
  • 3. Boll S. 1979 "Suppression of acoustic noise in speech using spectral subtraction" [IEEE Transactions on Acoustics, Speech and Signal Processing] Vol.27 P.113-120 google
  • 4. Lim J., Oppenheim A., Braida L. 1978 "Evaluation of an adaptive comb filtering method for enhancing speech degraded by white noise addition" [IEEE Transactions on Acoustics, Speech and Signal Processing] Vol.26 P.354-358 google
  • 5. Sui L. Y., Zhang X. W., Huang J. J., Zhou B. 2011 "An improved spectral subtraction speech enhancement algorithm under non-stationary noise" [Proceedings of International Conference on Wireless Communications and Signal Processing] P.1-5 google
  • 6. Daqrouq K., Abu-Isbeih I. N., Alfauri M. 2009 "Speech signal enhancement using neural network and wavelet transform" [Proceedings of the 6th International Multi-Conference on Systems, Signals and Devices] P.1-6 google
  • 7. Knecht W. G., Schenkel M. E., Moschytz G. S. 1995 "Neural network filters for speech enhancement" [IEEE Transactions on Speech and Audio Processing] Vol.3 P.433-438 google
  • 8. Ephraim Y., Malah D. 1985 "Speech enhancement using a minimum mean-square error log-spectral amplitude estimator" [IEEE Transactions on Acoustics, Speech and Signal Processing] Vol.33 P.443-445 google
  • 9. Okamoto R., Takahashi Y., Saruwatari H., Shikano K. 2010 "MMSE STSA estimator with nonstationary noise estimation based on ICA for high-quality speech enhancement" [Proceedings of IEEE International Conference on Acoustics Speech and Signal Processing] P.4778-4781 google
  • 10. Choi J. S., Park S. J. "Speech enhancement system based on auditory system and time-delay neural network" [Proceedings of the 8th International Conference on Adaptive and Natural Computing Algorithms] P.153-160 google
  • 11. Choi J. S. 2007 "An adaptive speech enhancement system based on noise level estimation and lateral inhibition" [Acta Acustica united with Acustica] Vol.93 P.632-644 google
  • 12. Glackin C., Maguire L., McDaid L. 2010 "Feature extraction from spectro-temporal signals using dynamic synapses, recurrency, and lateral inhibition" [Proceedings of International Joint Conference on Neural Networks] P.1-6 google
  • 13. Pearce D., Hirsch H. G. 2000 "The aurora experimental framework for the performance evaluation of speech recognition systems under noisy conditions" [Proceedings of the 6th International Conference on Spoken Language Processing] P.29-32 google
이미지 / 테이블
  • [ Fig. 1. ]  Modeling of lateral inhibition function.
    Modeling of lateral inhibition function.
  • [ Fig. 2. ]  The construction of a three-layer neural network. FFT: fastFourier transform.
    The construction of a three-layer neural network. FFT: fastFourier transform.
  • [ Fig. 3. ]  Proposed noise reduction neural network system. FFT: fast Fourier transform, NN: neural network.
    Proposed noise reduction neural network system. FFT: fast Fourier transform, NN: neural network.
  • [ Fig. 4. ]  Comparison of fast Fourier transform (FFT) amplitude spectrums for target, input, and output signal.
    Comparison of fast Fourier transform (FFT) amplitude spectrums for target, input, and output signal.
  • [ Fig. 5. ]  Comparison of fast Fourier transform (FFT) phase spectrums for target, input, and output signal.
    Comparison of fast Fourier transform (FFT) phase spectrums for target, input, and output signal.
  • [ Fig. 6. ]  Error curves of training for neural network with amplitude spectrums.
    Error curves of training for neural network with amplitude spectrums.
  • [ Fig. 7. ]  Proposed speech processing system.
    Proposed speech processing system.
  • [ Fig. 8. ]  A comparison of the proposed method and conventional noise suppression methods with the addition of car noise. SNR: signal-to-noise ratio, SS: spectral subtraction, NRNN: noise reduction neural network, W-LIPN: without lateral inhibition processing and NRNN, MMSE-LSA: minimum mean-square error log-spectral amplitude.
    A comparison of the proposed method and conventional noise suppression methods with the addition of car noise. SNR: signal-to-noise ratio, SS: spectral subtraction, NRNN: noise reduction neural network, W-LIPN: without lateral inhibition processing and NRNN, MMSE-LSA: minimum mean-square error log-spectral amplitude.
  • [ Fig. 9. ]  A comparison of the proposed method and conventional noise suppression methods with the addition of street noise. SNR: signal-to-noise ratio, SS: spectral subtraction, NRNN: noise reduction neural network, W-LIPN: without lateral inhibition processing and NRNN, MMSE-LSA: minimum mean-square error log-spectral amplitude.
    A comparison of the proposed method and conventional noise suppression methods with the addition of street noise. SNR: signal-to-noise ratio, SS: spectral subtraction, NRNN: noise reduction neural network, W-LIPN: without lateral inhibition processing and NRNN, MMSE-LSA: minimum mean-square error log-spectral amplitude.
(우)06579 서울시 서초구 반포대로 201(반포동)
Tel. 02-537-6389 | Fax. 02-590-0571 | 문의 : oak2014@korea.kr
Copyright(c) National Library of Korea. All rights reserved.