• 검색 결과가 없습니다.

An Efficient Recursive Total Least Squares Algorithm for Training Multilayer Feedforward Neural Networks

N/A
N/A
Protected

Academic year: 2021

Share "An Efficient Recursive Total Least Squares Algorithm for Training Multilayer Feedforward Neural Networks"

Copied!
4
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

2004년도 한국음향학회 학술발표대회 논문집 제23권 제2(s)호

An

Efficient

Recursive

Total

Least

Squares

Algorithm for Training M

니비

ayer

Feedforward

Neural

Networks

Nakjin Choi, *Jun-Seok

m, Koeng-Mo Sung

School

of

Electrical

Engineering

and

Computer Science, Seoul National University *Department

of

Electronics

Engineering,

Sejong University

nakjin@acoustics.snu.ac.kr

,

islim96@cholHan.net, kmsung@acoustics.snu.ac.kr

Abstract

We present a recursive total least squares (RTLS) algorithm for multilayer feedforward neural networks. So far, recursive least squares (RLS) has been successfully applied to training multilayer feedforward neural networks. But, when input data contain additive noise, the results from RLS could be biased. Such biased results can be avoided by using the recursive total least squares (RTLS) algorithm. The RTLS algorithm described in this paper gives better performance than RLS algorithm over a wide range of SNRs and involves approximately the same computational complexity of

1. Introduction

The property that is of primary significance for an artificial neural network is the ability of the network to learn from its environment, and to improve its performance through learning. The multilayer feedforward neural networks (MFNNs) have attracted a great deal of interest due to its rapid training, generality, and simplicity. In the past decade, the use of the recursive least squares (RLS) algorithm fbr training MFNNs has been investigated extensively [2]-[3].

In the RLS pr。미em, the underlying assumption is that we know the input vector exactly and all the errors are confined to the observation vector. Unfortunately, this assumption is frequently not true. Because quantization errors, human errors, modeling errors, and instrument errors

may preclude the possibility of knowing the input vector exactly [4]. Particularly, in the training of MFNNs, the inputs of the hidden and output neurons of the net contains an additive noise because of the use of quantization.

To overcome this problem, the total least squares (TLS) method has been devised. This method compensates for the errors of input vector and the errors of observation vector, sim니taneously. Most ^-dimensional TLS solutions have been obtained by computing a singular value decomposition(SVD), generally requiring ON) multiplications.

In this paper, we present an efficient recursive total least squares (RTLS) algorithm for training multilayer feedforward neural networks. This RTLS algorithm was first presented by Nakjin Choi et 이.[5][6]. This algorithm recursively calculates and tracks the eigenvector corresponding to the minimum eigenvalue, the estimated synaptic weights, from the inverse correlation matrix of the augmented sample matrix. Then, we demonstrate that this algorithm outperforms the RLS algorithm in training MFNNs thiv ugh computer simulation results. Moreover, we show that the recursive TLS algorithm involves approximately the same order of computational complexity O(A戸)as RLS algorithm.

In Section II, we will explain the training of MFNNs and in Section III, we derive an efficient RTLS algorithms fbr MFNNs training. In Section IV, the results and analysis

(2)

-of our experiment are given followed by our conclusions in Section V.

2. The Training of Multilayer Feedforward

NeuNetworks

The MFNNs have attracted a great deal of interest due to its rapid training, generality, and simplicity. The architectural graph in Fig. 1 illustrates the layout of a m니tiplay이" feedforward neural network fbr the case of a single hidden layer.

Fig. 1. MFNN with one hidden layer

Input layer Layer of Layer of

ofsource hidden output

nodes neurons neurons

An MFNN is trained by adjusting its weights according to an ongoing stream of input-output observations {x^n),

血):仁 1,…M 妇1,…/?} where denotes the number of training samples sequences. The objective is to obtain a set of weights such that the neural networks will predict future outputs accurately. This training process is equivalent to the process of parameter estimation. Thus, training MFNN can be considered as a nonlinear identification problem where the weight values are unknown and need to be identified for the given set of input-output vectors.

(aj Output measurement noise

I■(끼

(b) Input and output measurement noise

Fig. 2. Training of Neural Networks

Consider the training of neural networks in Fig. 2. Here, the training problem can be posed as a parameter identification and the dynamic equation fbr the neural network can be cast into the following fbnn

y(n) = f[w(n),x(n)] + £(M)

where w(n) is the parameter of the MFNN at time step n; is the observed output vector of the networks; £(n) is white Gaussian noises with zero mean. The input vector x(/i) fbr the training of neural network is usually taken to be clear like Fig. 2 (a). However, the fact that the unknown system input must be sampled and quantized (along with the desired signal) will result in a broad-band quantization noise contaminating the neural network input. So, Fig. 2(b) model which contains input and output measurement noise is more practical than Fig. 2(a) model which contains input measurement noise only.

The total least squares (TLS) method can compensates for the errors of input vector x(«) and observation vector y(/i), simultaneously.

3. Efficient RTLS Algorithm for MFNNs

The TLS problem is a minimization problem described as follows

mimmize|||E| rj||F subject to (A + E)x = b + r ⑵ where A is an MxN input matrix, b is an Mx] output vector. E, r are the MxN input error matrix and M 시

output error vector respectively. Once a minimization E, r are found, then any x satisfying

(A + E)x그 b + r (3) is assumed to be able to solve the TLS problem in eq.(2).

(3)

Let's define the Mx(N+l) augmented data matrix A as follows

A = [A|b], (4)

Golub and Van Loan proved that the TLS s이uti이] xtls involves the most right singular vector of the A as follows [4] ;vn+i,i *TLS = " , (5) VN+1,1 vN+i =: where 云心_

Now we develop the adaptive algorithm which calculates the TLS s이ution recursively.

We know that the most right singular vector Vn+i(«) can be obtained from the SVD of the inverse correlation matrix P(w) which is defined as follows

P(n) = R *(n) (6)

where R(^) = Ah(h)A(«)

The minimum eigenvector vn+i(«-1) at time index n-\ can be represented as a linear combination of the orthogonal eigenvectors v血),…vn+血)at time index n.

V N+1 (» -1) = C| (») v, (») + • • • + CN+1 (n)v N+1 (n) (7) Because P(n) and P(w-1) are highly correlated, the coefficient cn+i (끼 by which the minimum eigenvector vn+i(«) is multiplied, produces a larger value than any other coefficients. That is,

Cnh(")2Cn(")2“・ZC|(") (8) Now, consider the new vector v(w) which is defined as follows

v(n)

= P(«)v

N+i(n-l) (9) By using the singular value decomposition (SVD) Eq. (7), we can rewrite the vector v(z?) as Eq. (10)

V(시 = 여 -I)

攜念 g쎄 如財

N«l , \

» I

a

i(“

)

(10)

From the Eq. (8) and the property of SVD, the order of the magnitudes of the coefficients v((n) becomes

0n+i 어) > cnW > > 9 어)

^N+iW ^N(rt)~ 片(“) (n)

So v(/i) in Eq.(10) converges to the scaled minimum eigenvector as follows

v(w) r” 2 ,Cn+1(")

vV 聞 n+Q)

bN如) (12)

Therefore, the minimum eigenvector vn+i(«) can be approximately calculated from the v(n) as in Eq. (B).

.z 、 v(n)

")奇헤 (13)

4. Simatii

The MFNN model to identify the unknown system is connected with one hidden layer and one output layer. And it has 1 source node, 25 hidden neurons, and 1 output neuron. The bias of hidden layer ranges from -6.0 to 6.0 which is the dynamic range of input data x. The sampling frequency is 2Hz, which satisfies the Nyquist sampling rate. Also, the synaptic weights of the first hidden layer is fixed to 1. This model can be found in [7].

The unknown system is modeled as a nonlinear system. In this simulation, we assume that the relationship between input x and y of the unknown system is

y = sin(3x), -6 < x < 6 (14) In this MFNN model, the output of the first hidden layer is

W드[冰Jl6.0) 伊2(x-5.5) … 啊(x + 6.0)]「 (15)

and the output of total MFNN model is

y = yvH(p where W = R w2 … 叫」' (16)

The input and output measurement noise in Fig. 2 is a zero-mean white noise process, ind은pendent of the input vector.

It is assumed that there is a 1 크 14. and 16 bits 丄니aw pulse code modulation (PCM) quantizer which '、기I produce SNRs of 73.8, 85.8, and 97.8 dB respectively. Fig >

아lows the errors derived from RLS and RTLS Eih chest SNRs. It is observed that the presented RTLS outpertorms conventional RLS and provides better performance

(4)

-partic미 ariy at high SNRs.

5. Conclusions

In this paper, an efficient recursive total least squares algorithm is presented. This algorithm was found to outperform the RLS algorithm in neural network training. It involves approximately the same computational complexity as RLS algorithm. In order to validate its performance, we applied this algorithm to training m 이 tilayer feedforward neural network in the various quantization noise conditions. In each case, the result showed to be better than that of the RLS.

References

1. S. Haykin, Neural Networks A Comprehensive Foundation, 2nd Ed., Prentice-Hall, 1999.

2. K. W. Wong and C. S. Leung, uOn-line successive synthesis of wavelet networks/1 Neural Processing Lett., vol. 7, pp. 91-100, 1998.

3. C.S. Leung, A.C. Tsoi, uTwo-regularizers for recursive least squared algorithms in feedforward multilayered neural networks/' IEEE Trans. Neural Networks, vol. 12, pp. 1314-1332, 2001.

4. G. H. Golub and C. F. Van Loan, "An analysis of the total least squares problem", SIAM J. Numer. Anal., vol.

17, no. 6, pp. 883-893, Dec. 1980.

5. Nakjin Choi, J.S. Lim, J. Song, K.M. Sung, "Adaptive System Identi-fication Using an Efficient Recursive T아al Least Squares Algorithm," Journal of the Acoustical Society of Korea, Vol. 22, No3E, pp.93-100, 2003.9.

6. Nakjin Choi, J.S. Lim, J.Song and K.M. Sung,"An Efficient Recursive Total Least Squares Algorithm For Fequency Estimation," ICA, vol. 25. (6A.06.02), 2001. 7. Kevin M. Passino, Biomimicry for Optimization, Control and Automation, Springer-Verlag, London, UK, 2004.

iteration :iurab«x

(a) 12 bit quantization noise (SNR=73.8dB)

(b) 14 bit quantization noise (SNR=85.8dB)

(c) 16 bit quantization noise (SNR=97.8dB)

Fig. 3 .Learning curves(RLS vs. RTLS)

수치

Fig. 1. MFNN with one hidden layer
Fig. 3 .Learning curves(RLS vs.  RTLS)

참조

관련 문서

 알고리즘 담합: 경쟁사업자들이 알고리즘을 이용하여 가격, 공급량 등을 조정하거나 관련 정보를 공유하는 행태를 의미함.. 이는 시장에서 경쟁

– Search through the space of possible network weights, iteratively reducing the error E between the training example target values and the network outputs...

– An artificial neural network that fits the training examples, biased by the domain theory. Create an artificial neural network that perfectly fits the domain theory

In this paper, an adaptive pedestrian detection algorithm using the fusion of CCTV and IR camera is proposed.. One of the drawbacks of the pedestrian

An efficient algorithm for mining association rules in large databases. An effective hash-based algorithm for

generalized least square algorithm using variograms as weighting functions. With the least square algorithm, the estimate becomes unbiased and its variance becomes

Lagrange dual Duality Proof of strong duality Optimality conditions Theorems of alternatives.. Lagrange dual functions Lower bounds on

This section describes the overall structure of the diagnostic algorithm for abnormal situations using ANN. 6 illustrates the functional architecture of the diagnostic