I.
INTRODUCTION
B
Figure 2: Structure of the noise cancellation system
The corrupted speech signal is then used as the target
signal to the ADALINE. Based on the LMS rule, the
ADALINE adapts to cancel the out noise from the
noisy signal to produce the clean speech signal.
The testing speech signal recorded via sound card in
nearly ideal conditions. The recording signal with
8000Hz sampling rate is used as clean signal (free of
noise) and added to random noise to accomplish noisy
signal. The noisy speech signal is compared to the free
of noise until we obtain the speech signal that is clear
for the listener. The following algorithm steps present
the enhancement system used.
Step 1: set the learning rate at 0.05.
Step 2: set the weight at
1.2*(rand(1, p) - 0.5)
(1)
Where p is the dimensionality of the input space
Step 3: set the target signal as the corrupted speech
signal.
Step 4: set the input signal as the free of noise signal.
Step 5: for each Epoch calculate the output and error.
Step 6: adjust the weights of the network.
C
Figure 3: Adaline results. A. Free of noise signal B.
Noisy signal. C. Filtered signal.
IV. FEED FORWARD NEURAL NET FFNN
ENHANCEMENT METHOD
There are two basic types of networks, networks with
feedback and those without. In networks with feedback,
the output value can be traced back to the input values.
However there are networks wherein for every input
vector laid on the network, an output vector is
calculated and this can be read from the output neurons.
There is no feedback. Hence only, a forward flow of
information is presented. Network having this structure
high. The bias is provided for the hidden and the output
layer, to act upon the net input to be calculated. The
training algorithm of back propagation involves four
stages:
1. Intialization of weights
2. Feed forward
3. Back propagation of errors
4. Updation of the weight and biases.
During first stage which is the initialization of
weight, some small random values are assigned
during feed forward stage each input unit (Xi) receives
an input signal and transmits this signal to each of the
hidden units Z1Zp. Each hidden
unit
then
calculates the activation function and sends its
signal Zj to each output unit. The output unit
calculates the activation function to form the response
of the net for the given input pattern. During back
propagation of error, each output unit computes
activation Yk with its target value tk to determine the
associated error for that pattern with the unit. Based on
the error, the factor k (k=1,.,m) is computed
and is used to distribute the error at output unit Yk
back to all units in the previous layer. Similarly,
the factor j (j=1,.,p) is computed for each hidden
unit Zj. During final stage, the weight and biases are
update using the factor and the activation.
In this section we use a speech signal that was
recorded in noisy environment. The recorded signal
is used as input signal, which is compared with the
target signal that is free of noise. The Fig. 5 presents the
results of FFNN enhancement method. The SNR is
enhanced for about 3 dB.
The Training here is ON-line training which is
different than the training we used with the
ADALINE neural net, which off-line.
V. NEURAL NETWORK AND DISCRETE
WAVELET TRANSFORM NOISE CANCELATION
C
Figure 5: FFNN results. A. Free of noise signal B.
Noisy signal. C. Filtered signal.
Wavelets
are
mathematical
functions
that
decomposes
data
into
different
frequency
components, and then study each component with a
resolution matched to its scale. They have
advantages over traditional Fourier methods in
analyzing physical situations where the signal
contains discontinuities and sharp spikes. Wavelets
were developed independently in the fields of
mathematics, quantum physics, electrical engineering,
and seismic geology. Interchanges between these fields
during the last ten years have led to many new wavelet
applications such as image compression, turbulence,
human vision, radar, and earthquake prediction. Fig.7
illustrates random signal that has a length of 1000
samples, divided into number of frequency components
(subsignals).
In this suction we use the ADALINE as neural net
to filters each subsignal that produced by the Discrete
Wavelet Transform DWT (Fig.6) [16-18].
D
Figure 9: PSD of filtered signal by A. Adaline B.
FFNN. C. Neural Net Wavelet Transform
Figure 7: Random signal of 1000 samples length
divided into four levels of Discrete Wavelet
Transform subsignals
VI. CONCLUSIO
In this paper, three methods are tested: The Adaline,
Feed Forward Neural Network Enhancement Method
FFNN and Wavelet Transform and Adaline
Enhancement Method. The results obtained showed
high quality due to fast processing and high signal to
noise ratio. The tested signal is enhanced 10 dB by
Adaline, 3 dB by FFNN and 8 dB by Wavelet
Transform and Adaline Enhancement Method. By
comparing the PSD of free of noise with PSD
calculated for each of our three methods filtered signal,
the Adaline filtered signal has the best quality for many
cases. Wavelet Transform and Adaline Enhancement
Method can be used in speaker identification where the
enhancement and feature extracting are both required.
VII. REFERENCES
[1]Alcantara, J.I., et al., Preliminary evaluation of a formant
enhancement algorithm on the perception of speech in noise for
normally hearing listeners. Audiology. 33(1): p. 15-27. 1994.
[2]Kates, J.M., Speech enhancement based on a sinusoidal model.
Journal of Speech and Hearing Research. 37(2):449-64. 1994.
[3]Warren, R.M., et al., Spectral restoration of speech: intelligibility
is increased by inserting noise in spectral gaps. Perception and
Psychophysics. 59(2): 275-83. 1997.
[4]Gautam Moharir, Spectral subtraction method for speech
enhancement, M.Tech . thesis, Department of electrical engineering,
I.I.T. Bombay, Mumbai, India, Jan 2002.
[5] Boll S.F., Suppression of acoustic noise in speech using spectral
subtraction, IEEE Trans. On Acoustics, Speech and Signal
Processing, vol. ASSP-27, pp.113-120, Apr 1979.
[6] Gautam Moharir, Pushkar Patwardhan and Preeti Rao, Spectral
enhancement preprocessing for the HNM coding of noisy speech,
Proc. of International Conference on Spoken Language Processing,
Sep. 2002.
[7] M. Berouti, R. Schwartz, and J. Makhoul, Enhancement of speech
corrupted by acoustic noise, Proc. IEEE Int. Conf. Acoust., Speech
Signal Processing, pp. 208-211 (1979).
[8] Y. Ephriam and H. L. Van Trees, A signal subspace approach for
speech enhancement, in Proc. International Conference on Acoustic,
Speech and Signal Processing, vol. II, Detroit, MI, U.S.A., pp. 355358, May (1993).
[9] Simon Haykin, Adaptive Filter Theory, Prentice-Hall, ISBN 0-13322760-X, (1996).
[10]Grewal, Mohinder S., and Angus P. Andrews, Kalman Filtering
Theory and Practice. Upper Saddle River, NJ USA, Prentice
Hall..1993. [11]Jacobs Jacobs, O. L. R.. Introduction to Control
Theory, 2nd Edition. Oxford University Press. 1993.
[12]Widrow, B. and Hoff (1960), Adaptive switching circuits. In
1960 IRE WESCON Convention Record, pages 96 - 104 RE.1960.