In this paper the problem of noise removal from speech
signals using Variable Step Size based adaptive filtering is
presented. For this, the same formats for representing the data
as well as the filter coefficients as used for the LMS algorithm
were chosen. As a result, the steps related to the filtering
remains unchanged. The proposed treatment, however exploits
the modifications in the weight update formula for all
categories to its advantage and thus pushes up the speed over
the respective LMS-based realizations. Our simulations,
however, confirm that the ability of MRVSSLMS and
RVSSLMS algorithms is better than conventional LMS and
Kowngs VSSLMS algorithms in terms of SNR improvement
and convergence rate. Hence these algorithm is acceptable for
all practical purposes.
5 trang |
Chia sẻ: huongthu9 | Lượt xem: 385 | Lượt tải: 0
Bạn đang xem nội dung tài liệu Non Stationary Noise Removal from Speech Signals using Variable Step Size Strategy, để tải tài liệu về máy bạn click vào nút DOWNLOAD ở trên
Non Stationary Noise Removal from Speech Signals using
Variable Step Size Strategy
K. Prameela, M. Ajay Kumar, Mohammad Zia-Ur-Rahman and Dr B V Rama Mohana Rao
Dept. of E.C.E., Narasaraopeta Engg. College, Narasaraopeta-522 601, India
E-mail: mdzr_5@ieee.org
Abstract— The aim of this paper is to implement various adaptive
noise cancellers (ANC) for speech enhancement based on gradient
descent approach, namely the least-mean square (LMS) algorithm
and then enhanced to variable step size strategy. In practical
application of the LMS algorithm, a key parameter is the step size.
As is well known, if the step size is large, the convergence rate of
the LMS algorithm will be rapid, but the steady-state mean square
error (MSE) will increase. On the other hand, if the step size is
small, the steady state MSE will be small, but the convergence rate
will be slow. Thus, the step size provides a trade-off between the
convergence rate and the steady-state MSE of the LMS algorithm.
An intuitive way to improve the performance of the LMS algorithm
is to make the step size variable rather than fixed, that is, choose
large step size values during the initial convergence of the LMS
algorithm, and use small step size values when the system is close
to its steady state, which results in Variable Step Size LMS
(VSSLMS) algorithms. By utilizing such an approach, both a fast
convergence rate and a small steady-state MSE can be obtained.
By using this approach various forms of VSSLMS algorithms are
implemented. These are robust variable step-size LMS (RVSSLMS)
algorithm providing fast convergence at early stages of adaptation
and modified robust variable step-size LMS (MRVSSLMS)
algorithm. The performance of these algorithms is compared with
conventional LMS and Kowngs VSSLMS algorithm. Finally we
applied these algorithms on speech enhancement application.
Simulation results confirms that the implemented RVSSLMS and
MRVSSLMS are superior than conventional algorithms in terms of
convergence rate and signal to noise ratio improvement (SNRI).
Keywords— Adaptive filtering, LMS algorithm, Noise
Cancellation, Speech Processing, Variable Step Size.
I. INTRODUCTION
In real time environment speech signals are corrupted by
several forms of noise such as such as competing speakers,
background noise, car noise, and also they are subject to
distortion caused by communication channels; examples are
room reverberation, low-quality microphones, etc. In all such
situations extraction of high resolution signals is a key task. In
this aspect filtering come in to the picture. Basically filtering
techniques are broadly classified as non-adaptive and adaptive
filtering techniques. In practical cases the statistical nature of
all speech signals is non-stationary; as a result non-adaptive
filtering may not be suitable. Speech enhancement improves
the signal quality by suppression of noise and reduction of
distortion. Speech enhancement has many applications; for
example, mobile communications, robust speech recognition,
low-quality audio devices, and hearing aids.
Many approaches have been reported in the literature
to address speech enhancement. In recent years, adaptive
filtering has become one of the effective and popular
approaches for the speech enhancement. Adaptive filters
permit to detect time varying potentials and to track the
dynamic variations of the signals. Besides, they modify their
behavior according to the input signal. Therefore, they can
detect shape variations in the ensemble and thus they can
obtain a better signal estimation. The first adaptive noise
cancelling system at Stanford University was designed and
built in 1965 by two students. Their work was undertaken as
part of a term paper project for a course in adaptive systems
given by the Electrical Engineering Department. Since 1965,
adaptive noise cancelling has been successfully applied to a
number of applications. Several methods have been reported
so far in the literature to enhance the performance of speech
processing systems; some of the most important ones are:
Wiener filtering, LMS filtering [1], spectral subtraction [2]-[3],
thresholding [4]-[5]. On the other side, LMS-based adaptive
filters have been widely used for speech enhancement [6]–[8].
In a recent study, however, a steady state convergence
analysis for the LMS algorithm with deterministic reference
inputs showed that the steady-state weight vector is biased,
and thus, the adaptive estimate does not approach the Wiener
solution. To handle this drawback another strategy was
considered for estimating the coefficients of the linear
expansion, namely, the block LMS (BLMS) algorithm [9], in
which the coefficient vector is updated only once every
occurrence based on a block gradient estimation. A major
advantage of the block, or the transform domain LMS
algorithm is that the input signals are approximately
uncorrelated. Recently Jamal Ghasemi et.al [10] proposed a
new approach for speech enhancement based on eigenvalue
spectral subtraction, in [11] authors describes usefulness of
speech coding in voice banking, a new method for voicing
detection and pitch estimation. This method is based on the
spectral analysis of the speech multi-scale product [12].
In practice, LMS is replaced with its Normalized version,
NLMS. In practical applications of LMS filtering, a key
parameter is the step size. If the step size is large, the
convergence rate of the LMS algorithm will be rapid, but the
steady-state mean square error (MSE) will increase. On the
other hand, if the step size is small, the steady state MSE will
be small, but the convergence rate will be slow. Thus, the step
size provides a tradeoff between the convergence rate and the
steady-state MSE of the LMS algorithm. The performance of
the LMS algorithm may be improved by making the step size
variable rather than fixed. The resultant approach with
variable step size is known as variable step size LMS
(VSSLMS) algorithm [13]. By utilizing such an approach,
both a fast convergence rate and a small steady-state MSE can
be obtained. Many VSSLMS algorithms are proposed during
Mohd Zia-Ur-Rahman et al, International Journal of Computer Science & Communication Networks,Vol 1(1),September-October 2011
Available online @ www.ijcscn.com 91
ISSN:2249-5789
recent years [14]-[17]. In this paper, we considered the
problem of noise cancellation in speech signals by effectively
modifying and extending the framework of [1], using
VSSLMS algorithms mentioned in [14]-[17]. For that, we
carried out simulations on various real time speech signals
contaminated with real noise. The simulation results show that
the performances of the VSSLMS based algorithms are
comparable with LMS counterpart to eliminate the noise from
speech signals. Recently in [18] Karthik et.al demonstrated
speech enhancement using variable step size LMS (VSSLMS)
algorithms, in [19], [20] Rahman et.al presented speech
filtering using variable step size least mean fourth based
treatment and unbiased and normalized adaptive filtering
techniques.
II. ADAPTIVE ALGORITHMS
A. Basic Adaptive Filter Structure
Figure 1 shows an adaptive filter with a primary input that is
noisy speech signal s1 with additive noise n1. While the
reference input is noise n2, which is correlated in some way
with n1. If the filter output is y and the filter error e= (s1+n1)-y,
then
𝑒𝑒2 = (s1 + n1)2 – 2y (s1 + n1) + y2
= (n1 – y)2 + s12 + 2 s1 n1 – 2y s1. (1)
Since the signal and noise are uncorrelated, the mean-squared
error (MSE) is
E[e2]=E[(n1 – y)2]+E[s12] (2)
Minimizing the MSE results in a filter error output that is
the best least-squares estimate of the signal s1. The adaptive
filter extracts the signal, or eliminates the noise, by iteratively
minimizing the MSE between the primary and the reference
inputs. Minimizing the MSE results in a filter error output y
that is the best least-squares estimate of the signal s1.
Figure 1: Adaptive Filter Structure.
B. Conventional LMS Algorithms
The LMS algorithm is a method to estimate gradient vector
with instantaneous value. It changes the filter tap weights so
that e(n) is minimized in the mean-square sense. The
conventional LMS algorithm is a stochastic implementation of
the steepest descent algorithm. It simply replaces the cost
function ξ(n) = E[e2(n)] by its instantaneous coarse estimate.
The error estimation e(n) is
e(n) = d(n) – w(n) Φ(n) (3)
Where Φ(n) is input data sequence.
Coefficient updating equation is
w(n+1) = w(n) + µ Φ(n) e(n), (4)
Where µ is an appropriate step size to be chosen as 0 < µ < 2
𝑡𝑡𝑡𝑡 𝑅𝑅 for the convergence of the algorithm.
C. Kwong’s VSSLMS algorithm
The LMS type adaptive algorithm is a gradient search
algorithm which computes a set of weights wk that seeks to
minimize E(dk -XTkWk )The algorithm is of the form
Wk+1 = Wk + μkXkϵk
Where ϵk = dk + XTkW*k
and μk is the step size. In the standard LMS algorithm μk is a
constant. In this μk is time varying with its value determined
by the number of sign changes of an error surface gradient
estimate. Here the new variable step size or VSS algorithm,
for adjusting the step size μk yields :
μ′k+1 = αμk + γϵ2k 0 < α < 1,
γ > 0
and μmax if μ′k+1> μmax
μk+1 = μmin if μ′k+1< μmin
μ′k+1 otherwise
(5)
where 0 < μmin < μmax. The initial step size μ0 is usually
taken to be μmax, although the algorithm is not sensitive to the
choice. The step size μk , is always positive and is controlled
by the size of the prediction error and the parameters α and γ.
Intuitively speaking, a large prediction error increases the step
size to provide faster tracking. If the prediction error decreases,
the step size will be decreased to reduce the misadjustment.
The constant μmax is chosen to ensure that the mean-square
error (MSE) of the algorithm remains bounded. A sufficient
condition for μmax
μmax 2/(3 tr (R)) (6)
μmin is chosen to provide a minimum level of tracking ability.
Usually, μmin will be near the value of μ that would be chosen
for the fix ed step size (FSS) alg orith m. α must be chosen in
the range (0, 1) to provide exponential forgetting.
D. Robust Variable Step-Size LMS (RVSSLMS) algorithm
Mohd Zia-Ur-Rahman et al, International Journal of Computer Science & Communication Networks,Vol 1(1),September-October 2011
Available online @ www.ijcscn.com 92
ISSN:2249-5789
A number of time-varying step-size algorithms have
been proposed to enhance the performance of the conventional
LMS algorithm. Simulation results comparing the proposed
algorithm to current variable step-size algorithms clearly
indicate its superior performance for cases of stationary
environments. For non-stationary environments, our algorithm
performs as well as other variable step-size algorithms in
providing performance equivalent to that of the regular LMS
algorithm [17].
The adaptation step size is adjusted using the energy of the
instantaneous error. The weight update recursion is given by
w (n+1)= w(n)+μ(n)e(n)X(n)
And updated step-size equation is
μ(n+1)=αμ(n)+γe2(n) (7)
where 00 , and μ(n+1) is set to or when it falls below
or above these lower and upper bounds, respectively. The
constant μmax is normally selected near the point of instability
of the conventional LMS to provide the maximum possible
convergence speed. The value of μmax is chosen as a
compromise between the desired level of steady state
misadjustment and the required tracking capabilities of the
algorithm. The parameter γ controls the convergence time as
well as the level of misadjustment of the algorithm. At early
stages of adaptation, the error is large, causing the step size to
increase, thus providing faster convergence speed. When the
error decreases, the step size decreases, thus yielding smaller
misadjustment near the optimum. However, using the
instantaneous error energy as a measure to sense the state of
the adaptation process does not perform as well as expected in
the presence of measurement noise. The output error of the
identification system is
e(n)=d(n)-XT(n)W(n) (8)
where d(n) is the desired signal is given by
d(n)=XT(n)W*(n)+ξ(n) (9)
ξ(n) is a zero-mean independent disturbance, and W*(n) is
the time-varying optimal weight vector. Substituting (8) and
(9) in the step-size recursion, we get
μ(n+1)=αμ(n)+γVT(n)X(n)XT(n)V(n)+γξ2(n)-
2γξ(n)VT(n)X(n) (10)
Where V(n)=W(n)-W*(n) is the weight error vector. The input
signal autocorrelation matrix, which is defined as
R=E{X(n)XT(n)}, can be expressed as R=QᴧQT, where ᴧ is
the matrix of eigen values, and Q is the model matrix of R.
using Ṽ(n)=QTV(n) and Xʹ(n) = Q TX(n), then the statistical
behavior of μ(n+1) is determined.
E{μ(n+1)}=αE{μ(n)}+γ(E{ξ2(n)}+E{ ṼT(n)ᴧ Ṽ(n)})
where we have made use of the common independence
assumption of Ṽ(n) and Xʹ(n). Clearly, the term
E{ ṼT(n)ᴧXʹ(n) } in fluences the p rox imity of the adaptive
system to the optimal solution, and μ(n+1) is adjusted
accordingly. However, due to the presence of E{ξ2(n)}, the
step-size update is not an accurate reflection of the state of
adaptation before or after convergence. This reduces the
efficiency of the algorithm significantly. More specifically,
close to the optimum, μ(n) will still be large due to the
presence of the noise term E{ξ2(n)} .
The step size can be rewritten as
μ(n+1)=αμ(n)+γ[E{VT(n)X(n)XT(n-1)V(n-1)}]2 (11)
It is also clear from above discussion that the update of
μ(n) is dependent on how far we are from the optimum and is
not affected by independent disturbance noise. Finally, the
considered algorithm involves two additional update equations
compared with the standard LMS algorithm. Therefore, the
added complexity is six multiplications per iteration. These
multiplications can be reduced to shifts if the parameters α,β,γ,
are chosen as powers of 2.
E. Modified Robust Variable Step-Size LMS (MRVSSLMS)
algorithm
From the frame work of step size parameter of LMS
algorithm, Kwongs and RVSSLMS algorithms the step size of
MRVSS is given:
𝜇𝜇(𝑛𝑛 + 1) = �𝜇𝜇𝑚𝑚𝑚𝑚𝑚𝑚 ; 𝑖𝑖𝑖𝑖 𝜇𝜇(𝑛𝑛 + 1) > 𝜇𝜇𝑚𝑚𝑚𝑚𝑚𝑚 𝜇𝜇𝑚𝑚𝑖𝑖𝑛𝑛 ; 𝑖𝑖𝑖𝑖 𝜇𝜇(𝑛𝑛 + 1) < 𝜇𝜇𝑚𝑚𝑖𝑖𝑛𝑛
𝛼𝛼𝜇𝜇(𝑛𝑛) + 𝛾𝛾𝑝𝑝2(𝑛𝑛) (12)
p (n +1) = (1−β (n)) p(n) +β (n)e(n)e(n −1) (13)
𝛽𝛽(𝑛𝑛 + 1) = �𝛽𝛽𝑚𝑚𝑚𝑚𝑚𝑚 ; 𝑖𝑖𝑖𝑖 𝛽𝛽(𝑛𝑛 + 1) > 𝛽𝛽𝑚𝑚𝑚𝑚𝑚𝑚 𝛽𝛽𝑚𝑚𝑖𝑖𝑛𝑛 ; 𝑖𝑖𝑖𝑖 𝛽𝛽(𝑛𝑛 + 1) < 𝛽𝛽𝑚𝑚𝑖𝑖𝑛𝑛
𝜂𝜂𝛽𝛽(𝑛𝑛) + 𝜆𝜆𝑒𝑒2(𝑛𝑛) (14)
where the parameters 0 0 . The p (n) is the
time average of the error signal correlation at iteration time n
and n+1, and the β (n) is the time average of the square error
signal, which is used to control the sensitivity of p (n) to the
instantaneous error correlation. min max 0 < μmin < μmax ; 0 <
βmin < βmax <1 . The upper bound of step size μmax satisfied the
mean square stability condition. The lower bound of the step
size μmin is used to guarantee the excess MSE under the
tolerant level. The parameter β should be less than 1 and
larger than zero.
That is to say, when the algorithm is convergent, the
instantaneous error power is very small and the error signal
correlation is not sensitive to instantaneous error, and the
accuracy of error signal correlation is enhanced. If the system
is suddenly changed, the instantaneous error signal power is
increased, which result to the enlargement of the correlation
function of the error signal and the instantaneous error signal
correlation, therefore the algorithm has a good tracking ability.
Mohd Zia-Ur-Rahman et al, International Journal of Computer Science & Communication Networks,Vol 1(1),September-October 2011
Available online @ www.ijcscn.com 93
ISSN:2249-5789
In one word, the MRVSS have good tracking ability and good
anti-noise ability, which are the advantages of algorithm
proposed in reference [15][17]. Using these strategies
different adaptive noise cancellers are implemented to remove
diverse form of noises from speech signals.
III. SIMULATION RESULTS
To show that RVSSLMS and MRVSSLMS algorithms are
appropriate for speech enhancement we have used real speech
signals with noise. In the figure number of samples is taken on
x-axis and amplitude is taken on y-axis. In order to test the
convergence performance we have simulated a sudden noise
spike at 4000th sample. From the figure it is clear that the
performance of the implemented RVSSLMS and
MRVSSLMS algorithms is better than the conventional LMS
and Kwongs VSSLMS algorithm. To prove the concept of
filtering we have considered five speech samples
contaminated with various real noises. These noises are high
voltage murmuring, crane noise. For comparison purpose we
also considered random noise removal. Generally the noise
added to the speech signal when it is transmitted through free
space is random in nature. The noisy speech signal is given as
in put to the adaptive filter structure shown in Figure 1, signal
somewhat correlated with noise is given as reference signal.
As the number of iterations increases error decreases and
clean signal can be extracted from the output of the filter.
These simulation results are shown in Figures 3, 4. To
evaluate the performance of the algorithms SNRI is measured
and tabulated in Tables I, II, III.
IV CONCLUSION
In this paper the problem of noise removal from speech
signals using Variable Step Size based adaptive filtering is
presented. For this, the same formats for representing the data
as well as the filter coefficients as used for the LMS algorithm
were chosen. As a result, the steps related to the filtering
remains unchanged. The proposed treatment, however exploits
the modifications in the weight update formula for all
categories to its advantage and thus pushes up the speed over
the respective LMS-based realizations. Our simulations,
however, confirm that the ability of MRVSSLMS and
RVSSLMS algorithms is better than conventional LMS and
Kowngs VSSLMS algorithms in terms of SNR improvement
and convergence rate. Hence these algorithm is acceptable for
all practical purposes.
Figure 3: Typical filtering results of high voltage murmuring removal (a)
Speech Signal with real noise, (b) recovered signal using LMS algorithm, (c)
recovered signal using Kowngs VSSLMS algorithm, (d) recovered signal
using RVSSLMS algorithm, (e) recovered signal using MRVSSLMS
algorithm.
Figure 4: Typical filtering results of crane noise removal (a) Speech
Signal with real noise, (b) recovered signal using LMS algorithm, (c)
recovered signal using Kowngs VSSLMS algorithm, (d) recovered signal
using RVSSLMS algorithm, (e) recovered signal using MRVSSLMS
algorithm.
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
x 10
4
-20
2
(a)
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
x 10
4
-2
0
2
(b)
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
x 10
4
-2
0
2
(c)
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
x 10
4
-2
0
2
(d)
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
x 10
4
-2
0
2
(e)
0 1 2 3 4 5 6
x 10
4
-2
0
2
(a)
0 1 2 3 4 5 6
x 10
4
-2
0
2
(b)
0 1 2 3 4 5 6
x 10
4
-2
0
2
(c)
0 1 2 3 4 5 6
x 10
4
-2
0
2
(d)
0 1 2 3 4 5 6
x 10
4
-2
0
2
(e)
Mohd Zia-Ur-Rahman et al, International Journal of Computer Science & Communication Networks,Vol 1(1),September-October 2011
Available online @ www.ijcscn.com 94
ISSN:2249-5789
Table I: SNR Contrast for Random noise removal.
Sl. No Sample
No
Before
Filtering
LMS Kowngs VSSLMS RVSSLMS MRVSSLMS
After Imp After Imp After Imp After Imp
1 I 0.7523 5.9077 5.1553 6.5145 5.7621 9.0738 8.3214 10.1066 9.3542
2 II -2.1468 4.1468 6.6975 5.7103 8.2610 6.6617 9.2154 7.9232 10.4730
3 III -4.1554 1.4826 5.6380 1.539 5.6944 3.1546 7.3100 4.7609 8.9163
4 IV -3.6941 1.9213 5.6154 2.0417 5.7358 3.5682 7.2623 5.1431 8.8372
5 V -5.6992 0.5443 6.2435 2.3337 8.0329 2.6920 8.3912 3.8539 9.5531
Average Improvement 5.8699 6.6972 8.1000 9.4269
Table II: SNR Contrast for High voltage murmuring removal.
S.No Sample
No
Before
Filtering
LMS Kowngs VSSLMS RVSSLMS MRVSSLMS
After Imp After Imp After Imp After Imp
1 I -1.5937 2.0034 3.5971 3.0735 4.6672 4.2078 5.8015 4.6311 6.2248
2 II 0.0705 1.7646 1.6940 1.9657 1.8951 5.9283 5.8577 6.5044 6.4338
3 III 2.6032 4.3508 1.7476 5.5225 2.9193 7.4302 4.8270 7.9161 5.3129
4 IV 3.0644 4.9673 1.9029 6.6277 3.5633 7.4096 4.3452 8.5129 5.4485
5 V 0.9671 2.8560 1.8888 3.0758 2.1086 7.1156 6.1484 7.9817 7.0145
Average Improvement 2.1660 3.0307 5.3959 6.0869
Table III: SNR Contrast for Crane noise removal.
S.No Sample
No
Before
Filtering
LMS Kowngs VSSLMS RVSSLMS MRVSSLMS
After Imp After Imp After Imp After Imp
1 I 0.5244 3.2108 2.6863 4.1024 3.5770 4.2822 3.7577 4.6914 4.1669
2 II -1.8459 3.2714 5.1173 5.7327 7.5786 6.0373 7.8832 6.7004 8.5463
3 III -2.1790 3.3691 5.5481 4.2556 6.4346 4.3284 6.5074 4.9409 7.1199
4 IV -1.6394 2.3560 3.9954 4.2422 5.8816 4.4689 6.1083 5.1134 6.7528
5 V -3.6823 0.7695 4.4518 4.9700 8.6523 5.8311 9.5134 6.7282 10.4109
Average Improvement 4.3597 6.4250 6.7540 7.3993
REFERENCES
[1] B. Widrow, J. Glover, J. M. McCool, J. Kaunitz, C. S. Williams, R.
H.Hearn, J. R. Zeidler, E. Dong, and R. Goodlin,“Adaptive noise cancelling:
Principles and applications ”, Proc. IEEE, vol. 63, pp.1692-1716, Dec. 1975.
[2] B. L. Sim, Y. C. Tong, J. S. Chang and C. T. Tan, “A parametric
formulation of the generalized spectral subtraction method,” IEEE Trans. On
Speech and Audio Processing, vol. 6, pp. 328-337, 1998.
[3] I. Y. Soon, S. N. Koh, and C. K. Yeo, “Noisy speech enhancement using
discrete cosine transform,” Speech Communication, vol. 24, pp. 249-257,
1998.
[4] H. Sheikhzadeh, and H. R. Abutalebi, “An improved wavelet-based
speech enhancement system,” Proc. of the Eurospeech, 2001.
[5] S. Salahuddin, S. Z. Al Islam, M. K. Hasan, M. R. Khan, “Soft
thresholding for DCT speech enhancement,” Electron. Letters, vol. 38,
no.24, pp. 1605-1607, 2002.
[6] J. Homer, “Quantifying the convergence speed of LMS adaptive filter
with autoregressive inputs,” Electronics Letters, vol. 36, no. 6, pp. 585– 586,
March 2000.
[7] H. C. Y. Gu, K. Tang and W. Du, “Modifier formula on mean square
convergence of LMS algorithm,” Electronics Letters, vol. 38, no. 19, pp. 1147
–1148, Sep 2002.
[8] M. Chakraborty and H. Sakai, “Convergence analysis of a complex LMS
algorithm with tonal reference signals,” IEEE Trans. on Speech and Audio
Processing, vol. 13, no. 2, pp. 286 – 292, March 2005.
[9] S. Olmos , L. Sornmo and P. Laguna, ``Block adaptive filter with
deterministic reference inputs for event-related signals:BLMS and BRLS,"
IEEE Trans. Signal Processing, vol. 50, pp. 1102-1112, May.2002.
[10] Jamal Ghasemi and Mohammad Reza Karami Mollaei, “A New
Approach for Speech Enhancement Based On Eigenvalue Spectral
Subtraction”, Signal Processing: An International Journal, vol. 3, Issue. 4, pp.
34-41.
[11] Mohamed Anouar Ben Messaoud, Aïcha Bouzid and Noureddine
Ellouze,” A New Method for Pitch Tracking and Voicing Decision Based on
Spectral Multi-scale Analysis”, Signal Processing: An International Journal,
vol. 3, Issue. 5, pp. 144-152.
[12] M.Satya Sai Ram, P. Siddaiah and M. Madhavi Latha,” USEFULNESS
OF SPEECH CODING IN VOICE BANKING”, Signal Processing: An
International Journal, vol. 3, Issue. 4, pp. 42-52.
[13] Yonggang Zhang, Ning Li, Jonathon A. Chambers, and Yanling Hao,
“New Gradient-Based Variable Step Size LMS Algorithms,” EURASIP
Journal on Advances in Signal Processing vol. 2008, Article ID 529480, 9
pages, doi:10.1155/2008/529480.
[14] S. Karni and G. Zeng, “A new convergence factor for adaptive filters,”
IEEE Transactions on Circuits and Systems, vol. 36, no. 7, pp. 1011–1012,
1989.
[15] R. H. Kwong and E.W. Johnson, “A variable step-size LMS algorithm,”
IEEE Transactions on Signal Processing, vol. 40, no. 7, pp. 1633–1642, 1992.
[16] V. J. Mathews and Z. Xie, “A stochastic gradient adaptive filter with
gradient adaptive step-size,” IEEE Transactions on Signal Processing, vol. 41,
no. 6, pp. 2075–2087, 1993.
[17] T. Aboulnasr and K.Mayyas, “A robust variable step-size LMStype
algorithm: analysis and simulations,” IEEE Transactions on Signal
Processing, vol. 45, no. 3, pp. 631–639, 1997.
[18] G.V.S. Karthik, M.Ajay Kumar, Md.Zia Ur Rahman, “Speech
Enhancement Using Gradient Based Variable Step Size Adaptive Filtering
Techniques”, International Journal of Computer Science & Emerging
Technologies, UK, (E ISSN: 2044-6004), Volume 2, Issue 1, February 2011,
pp. 168-177.
[19] Md. Zia Ur Rahman, K.Murali Krishna, G.V.S. Karthik, M. John Joseph
and M.Ajay Kumar, “ Non Stationary Noise Cancellation in Speech Signals
using an Efficient Variable step size higher order filter”, International Journal
of Research and Reviews in Computer Science, UK, Vol. 2, No. 1, 2011.
[20] Md Zia Ur Rahman et al., “Filtering Non Stationary noise in Speech
signals using Computationally efficient unbiased and Normalized Algorithm”,
International Journal on Computer Science and Engineering, ISSN : 0975-
3397 Vol. 3 No. 3 Mar 2011, pp. 1106- 1113.
Mohd Zia-Ur-Rahman et al, International Journal of Computer Science & Communication Networks,Vol 1(1),September-October 2011
Available online @ www.ijcscn.com 95
ISSN:2249-5789
Các file đính kèm theo tài liệu này:
- non_stationary_noise_removal_from_speech_signals_using_varia.pdf