25
Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang, Vijaya K. Devabhaktuni, Changgeng Xi, Qi-Jun Zhang Department of Electronics, Carleton University, Ottawa, Canada, K1S 5B6; e-mail: [email protected]; [email protected]; [email protected]; [email protected] Recei ¤ ed 11 August 1998; re¤ ised 1 December 1998 ABSTRACT: Neural networks recently gained attention as fast and flexible vehicles to microwave modeling, simulation, and optimization. After learning and abstracting from microwave data, through a process called training, neural network models are used during microwave design to provide instant answers to the task learned. Appropriate neural network structure and suitable training algorithm are two of the major issues in developing neural network models for microwave applications. Together, they decide amount of training data required, accuracy that could possibly be achieved, and more importantly developmen- tal cost of neural models. A review of the current status of this emerging technology is presented, with emphasis on neural network structures and training algorithms suitable for microwave applications. Present challenges and future directions of the area are discussed. Q 1999 John Wiley & Sons, Inc. Int J RF and Microwave CAE 9: 216]240, 1999. Keywords: neural networks; structures; training; modeling; RF; microwave I. INTRODUCTION The drive for manufacturability-oriented design and reduced time-to-market in the microwave industry require design tools that are accurate and fast. Statistical analysis and optimization with Ž . detailed physics ] electromagnetic EM models of active and passive components can be an impor- tant step toward a design for first-pass success, but it is computationally intensive. In recent years Ž . a novel computer-aided design CAD approach based on neural network technology has been introduced in the microwave community, for the modeling of passive and active microwave compo- w x w nents 1 ] 5 , and microwave circuit design 2, 4, 6, x 7 . A neural network model for a device ] circuit can be developed by learning and abstracting from measured ] simulated microwave data, through a process called training. Once trained, the neural network model can be used during Correspondence to: Q.-J. Zhang microwave design to provide instant answers to wx the task it learned 1 . Recent work by microwave researchers demonstrated the ability of neural networks to accurately model a variety of mi- crowave components, such as microstrip intercon- w x w x w x nects 1, 3, 8 , vias 3, 9 , spiral inductors 5, 10 , w x FET devices 1, 11 ] 13 , power transistors and w x Ž . power amplifiers 14 , coplanar waveguide CPW wx circuit components 4 , packaging and intercon- w x nects 15 , etc. Neural networks have been used in w x circuit simulation and optimization 2, 13, 16 , signal integrity analysis and optimization of very Ž . large scale integrated VLSI circuit interconnects w x w x 8, 15 , microstrip circuit design 17 , microwave w x Ž . filter design 18 , integrated circuit IC modeling w x w x wx 11 , process design 19 , synthesis 6 , Smith chart wx representation 7, and microwave impedance w x matching 20 . The neural network technologies have been applied to microwave circuit optimiza- tion and statistical design with neural network w x models at both device and circuit levels 2, 12 . These pioneering works helped to establish the Q 1999 John Wiley & Sons, Inc. CCC 1096-4290r99r030216-25 216

Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

  • Upload
    donhu

  • View
    218

  • Download
    2

Embed Size (px)

Citation preview

Page 1: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Neural Network Structures and Training Algorithmsfor RF and Microwave Applications

Fang Wang, Vijaya K. Devabhaktuni, Changgeng Xi, Qi-Jun Zhang

Department of Electronics, Carleton University, Ottawa, Canada, K1S 5B6; e-mail:[email protected]; [email protected]; [email protected]; [email protected]

Recei ed 11 August 1998; re¨ised 1 December 1998

ABSTRACT: Neural networks recently gained attention as fast and flexible vehicles tomicrowave modeling, simulation, and optimization. After learning and abstracting frommicrowave data, through a process called training, neural network models are used duringmicrowave design to provide instant answers to the task learned. Appropriate neuralnetwork structure and suitable training algorithm are two of the major issues in developingneural network models for microwave applications. Together, they decide amount of trainingdata required, accuracy that could possibly be achieved, and more importantly developmen-tal cost of neural models. A review of the current status of this emerging technology ispresented, with emphasis on neural network structures and training algorithms suitable formicrowave applications. Present challenges and future directions of the area are discussed.Q 1999 John Wiley & Sons, Inc. Int J RF and Microwave CAE 9: 216]240, 1999.

Keywords: neural networks; structures; training; modeling; RF; microwave

I. INTRODUCTION

The drive for manufacturability-oriented designand reduced time-to-market in the microwaveindustry require design tools that are accurateand fast. Statistical analysis and optimization with

Ž .detailed physics]electromagnetic EM models ofactive and passive components can be an impor-tant step toward a design for first-pass success,but it is computationally intensive. In recent years

Ž .a novel computer-aided design CAD approachbased on neural network technology has beenintroduced in the microwave community, for themodeling of passive and active microwave compo-

w x wnents 1]5 , and microwave circuit design 2, 4, 6,x7 . A neural network model for a device]circuit

can be developed by learning and abstractingfrom measured]simulated microwave data,through a process called training. Once trained,the neural network model can be used during

Correspondence to: Q.-J. Zhang

microwave design to provide instant answers tow xthe task it learned 1 . Recent work by microwave

researchers demonstrated the ability of neuralnetworks to accurately model a variety of mi-crowave components, such as microstrip intercon-

w x w x w xnects 1, 3, 8 , vias 3, 9 , spiral inductors 5, 10 ,w xFET devices 1, 11]13 , power transistors and

w x Ž .power amplifiers 14 , coplanar waveguide CPWw xcircuit components 4 , packaging and intercon-

w xnects 15 , etc. Neural networks have been used inw xcircuit simulation and optimization 2, 13, 16 ,

signal integrity analysis and optimization of veryŽ .large scale integrated VLSI circuit interconnects

w x w x8, 15 , microstrip circuit design 17 , microwavew x Ž .filter design 18 , integrated circuit IC modeling

w x w x w x11 , process design 19 , synthesis 6 , Smith chartw xrepresentation 7 , and microwave impedance

w xmatching 20 . The neural network technologieshave been applied to microwave circuit optimiza-tion and statistical design with neural network

w xmodels at both device and circuit levels 2, 12 .These pioneering works helped to establish the

Q 1999 John Wiley & Sons, Inc. CCC 1096-4290r99r030216-25

216

Page 2: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 217

framework of neural modeling technology in mi-crowave applications. Neural models are muchfaster than original detailed EM]physics modelsw x1, 2 , more accurate than polynomial and empiri-

w xcal models 21 , allow more dimensions than tablew xlookup models 22 , and are easier to develop

w xwhen a new device]technology is introduced 16 .Theoretically, neural network models are a

kind of black box models, whose accuracy de-pends on the data presented to it during training.A good collection of the training data, i.e., datawhich is well-distributed, sufficient, and accu-rately measured]simulated, is the basic require-ment to obtain an accurate model. However, inthe reality of microwave design, training datacollection]generation may be very expen-sive]difficult. There is a trade-off between theamount of training data needed for developingthe neural model, and the accuracy demanded bythe application. Other issues affecting the accu-racy of neural models are due to the fact thatmany microwave problems are nonlinear, nons-mooth, or containing many variables. An appro-priate structure would help to achieve highermodel accuracy with fewer training data. For ex-ample, a feedforward neural network with smoothswitching functions in the hidden layer is good atmodeling smooth, slowly varying nonlinear func-tions, while a feedforward neural network withGaussian functions in the hidden layer could bemore effective in modeling nonlinear functionswith large variations. The size of the structure,i.e., the number of neurons is also an importantcriteria in the development of a neural network.Too small a network cannot learn the problemwell, but too large a size will lead to overlearning.An important type of neural network structure isthe knowledge-based neural networks where mi-crowave empirical information is embedded into

w xneural network structures 1 , enhancing reliabil-ity of the neural model and reducing the amountof training data needed.

Training algorithms are an integral part ofneural network model development. An appropri-ate structure may still fail to give a better model,unless trained by a suitable training algorithm. Agood training algorithm will shorten the trainingtime, while achieving a better accuracy. The mostpopular training algorithm is backpropagationŽ .BP , which was proposed in the mid 1980s. Later,a lot of variations to improve the convergence ofBP were proposed. Optimization methods such assecond-order methods and decomposed optimiza-tion have also been used for neural network train-

ing in recent years. A noteworthy challenge en-countered in the neural network training is theexistence of numerous local minima. Global opti-mization techniques have been combined withconventional training algorithms to tackle thisdifficulty.

In Section II, the problem statement of neuralbased microwave modeling is described. In Sec-tion III, a detailed review of various neural net-work structures useful to microwave design area

Žsuch as standard feedforward networks multi-.layer perceptrons, radial basis functions , net-

Žworks with prior knowledge electromagnetic arti-ficial neural networks or EM-ANN, knowledge

.based neural networks , combined networks,wavelet and constructive networks, is presented.Section IV discusses various training algorithmsof interest such as backpropagation, second-order

Žalgorithms conjugate gradient, quasi-Newton,.Levenberg]Marquardt , decomposed optimiza-

tion algorithms, and global training algorithms.Section V presents a comparison of neural net-work structures and training algorithms, throughmicrowave examples. Finally, Section VI containsconclusions and future challenges in this area.

II. NEURAL BASED MICROWAVEMODELING: PROBLEM STATEMENT

Let x be an N -vector containing parameters of axgiven device or a circuit, e.g., gate length and gatewidth of a FET transistor; or geometrical andphysical parameters of transmission lines. Let ybe an N -vector containing the responses of theydevice or the circuit under consideration, e.g.,drain current of a FET; or S-parameters of trans-mission line. The relationship between x and ymay be multidimensional and nonlinear. In theoriginal EM]circuit problems, this relationship isrepresented by,

Ž . Ž .y s f x . 1

Such a relationship can be modeled by a neuralnetwork, by training it through a set of x y ysample pairs given by,

Ž . Ž .x , d , p s 1, 2, . . . , N , 2� 4p p p

where x and d are N - and N -dimensionalp p x yvectors representing the pth sample of x and y,respectively. This sample data called the training

Page 3: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.218

data is generated from original EM simulationsor measurement. Let the neural network model

Ž .for the relationship in 1 be represented by,

Ž . Ž .y s y x, w , 3

where w is the parameter of the neural networkmodel, which is also called the weight vector inneural network literature, and x and y are calledthe inputs and outputs of the neural model. Thedefinition of w, and how y is computed through xand w determine the structure of the neural net-

Ž .work. The neural model of 3 does not representŽ .the original problem of 1 , unless the neural

Ž .model is trained by data in 2 . A basic descrip-tion of the training problem is to determinew such that the difference between the neuralmodel outputs y and desired outputs d fromsimulation]measurement,

N Np y1 2Ž . Ž . Ž .E w s y x , w y d 4Ž .Ý Ý pk p pk2 ps1 ks1

Ž .is minimized. In 4 d is the kth element ofpkŽ .vector d , y x , w is the k th output of thep pk p

neural network when the input presented to thenetwork is x . Once trained, the neural networkpmodel can be used for predicting the output val-ues given only the values of the input variables. Inthe model testing stage, an independent set ofinput]output samples, called the testing data isused to test the accuracy of the neural model.Normally, the testing data should lie within thesame input range as the training data. The abilityof neural models to predict y when presentedwith input parameter values x, never seen duringtraining is called the generalization ability. Atrained and tested neural model can then be usedonline during microwave design stage providingfast model evaluation replacing original slowEM]device simulators. The benefit of the neuralmodel approach is especially significant when themodel is highly repetitively used in design pro-cesses such as, optimization, Monte Carlo analy-sis, and yield maximization.

When the outputs of neural network are con-tinuous functions of the inputs, the modelingproblem is known as regression or function ap-proximation, which is the most common case inmicrowave design area. In the next section, adetailed review of neural network structures usedfor this purpose is presented.

III. NEURAL NETWORK STRUCTURES

In this section, different ways of realizing y sŽ .y x, w are described. The definition of w and how

y is computed from x and w in the model deter-mine different neural model architectures.

A. Standard FeedforwardNeural Networks

Feedforward neural networks are a basic type ofneural networks capable of approximating genericclasses of functions, including continuous and in-

w xtegrable ones 23 . An important class of feedfor-ward neural networks is multilayer perceptronsŽ .MLP . Typically, the MLP neural network con-sists of an input layer, one or more hidden layers,and an output layer, as shown in Figure 1. Sup-pose the total number of hidden layers is L. Theinput layer is considered as layer 0. Let the num-ber of neurons in hidden layer l be N , l sl1, 2,, . . . , L. Let w l represent the weight of thei jlink between the jth neuron of the l y 1th hid-den layer and ith neuron of lth hidden layer, andu l be the bias parameter of ith neuron of lthihidden layer. Let x represent the ith input pa-i

lrameter to the MLP. Let y be the output of ithineuron of lth hidden layer, which can be com-puted according to the standard MLP formulas

Ž .Figure 1. Multilayer perceptron MLP structure.Typically, the network consists of an input layer, one ormore hidden layers, and an output layer.

Page 4: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 219

as,

Nly1ll ll ly1 ly s s w ? y q u ,Ýi i j j iž /js1 Ž .5

i s 1, . . . , N , l s 1, . . . , Ll

0y s x ,i i Ž .6i s 1, . . . , N , N s N ,x x 0

Ž .where s . is usually a monotone squashing func-tion. Let n represent the weight of the linkk ibetween the ith neuron of the Lth hidden layerand the k th neuron of the output layer, and bkbe the bias parameter of the kth output neuron.The outputs of MLP can be computed as,

NLL Ž .y s n ? y q b , k s 1, . . . , N . 7Ýk k i i k y

is1

For function approximation, output neurons canŽ .be processed by linear functions as shown in 7 .

The most commonly used function for hiddenŽ .neurons s . , also called the activation function,

is the logistic sigmoid function given by,

1Ž . Ž .s t s , 8ytŽ .1 q e

which has the property of,

1, as t ª q`,Ž . Ž .s t ª 9½ 0, as t ª y`.

Ž .Other possible candidates for s . are the arctan-gent function given by,

2Ž . Ž . Ž .s t s arctan t , 10ž /p

and the hyperbolic tangent function given by,

Ž t yt .e y eŽ . Ž .s t s . 11t ytŽ .e q e

All these functions are bounded, continuous,monotonic, and continuously differentiable.Training parameters w includes,

lw s w , j s 1, . . . , N , i s 1, . . . , N ,i j ly1 l

l s 1, . . . , L; u l , i s 1, . . . , N ,i l

l s 1, . . . , L;n , i s 1, . . . , N , k s 1, . . . , N ;k i L y

Ž .b , k s 1, . . . , N . 12k y

ŽIt is well known that a two-layered MLP no.hidden layers is not capable of approximating

w xgeneric nonlinear continuous functions 24, 25 .w xThe universal approximation theorem 26, 27

states that a three-layer perceptron, with onehidden sigmoidal layer, is capable of modelingvirtually any real function of interest to any de-sired degree of accuracy, provided sufficientlymany hidden neurons are available. As such, fail-ure to develop a good neural model can be at-tributed to inadequate learning, inadequate num-ber of hidden neurons, or the presence of astochastic rather than a deterministic relation be-

w xtween input and output 27 .However, in reality a neural network can only

have a finite number of hidden neurons. Usually,three- or four-layered perceptrons are used inneural modeling of microwave circuit compo-nents. Neural network performance can be eval-uated based on generalization capability and

w xmapping capability 28 . In the function approxi-mation or regression area, generalization capabil-

w xity is a major concern. It is shown in 29 thatfour-layered perceptrons are not preferred in allbut the most esoteric applications in terms ofgeneralization capability. Intuitively, four-layeredperceptrons would perform better in defining thedecision boundaries in pattern classification tasksbecause of an additional nonlinear hidden layerresulting in hierarchical decision boundaries. This

w xhas been verified in 28 for the mapping capabil-ity of the network.

Feedforward neural networks which have onlyone hidden layer, and which use radial basis acti-vation functions in the hidden layer, are called

Ž .radial basis function RBF networks. Radial ba-sis functions are derived from the regularizationtheory in the approximation of multivariate func-

w xtions 30, 31 . Park and Sandberg showed thatRBF networks also have universal approximation

w xability 32, 33 . Universal convergence of RBFnets in function estimation and classification has

w xbeen proven by Krzyzak, Linder, and Lugosi 34 .The output neurons of RBF networks are also

linear neurons. The overall input]output transferfunction of RBF networks is defined as,

N1

Ž 5 5. Ž .y s n s x y u , j s 1, . . . , N , 13Ýj ji i yis1

where u is the center of radial basis function ofithe ith hidden neuron, n is the weight of theji

Page 5: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.220

link from the ith hidden neuron to the jth outputneuron. Some of the commonly used radial basis

w xactivation functions are 35 ,

2tŽ . Ž .s t s exp y , 14ž /l

b2 2Ž . Ž . Ž .s t s c q t , 0 - b - 1, 15

Ž . Ž .where 14 is the Gaussian function, 15 is themultiquadratic function and, l, c, and b are thefunction parameters. Training parameters w in-cludes u , n , i s 1, . . . , N , j s 1, . . . , N and li ji 1 yor c and b.

Although MLP and RBF are both feedforwardneural networks, the different nature of the hid-den neuron activation functions makes them be-have very differently. First, the activation functionof each hidden neuron in an MLP computes theinner product of the input vector and the synapticweight vector of that neuron. On the other hand,the activation function of each hidden neuron ina RBF network computes the Euclidean normbetween the input vector and the center of thatneuron. Second, MLP networks construct globalapproximations to nonlinear input]output map-ping. Consequently, they are capable of generaliz-ing in those regions of the input space where littleor no training data is available. On the contrary,RBF networks use exponentially decaying local-

Ž .ized nonlinearities e.g., Gaussian functions toconstruct local approximations to nonlinear in-put]output mapping. As a result RBF neuralnetworks are capable of faster learning and ex-hibit reduced sensitivity to the order of presenta-

w xtion of training data 36 . Consequently, a hiddenneuron influences the outputs of the network onlyfor inputs near to its center, and an exponentialnumber of hidden neurons are required to cover

w xthe entire domain. In 37 , it is suggested thatRBF networks are suited for problems with asmaller number of inputs.

The universal approximation theorems for bothMLP and RBF only state that there exists such anetwork to approximate virtually any nonlinearfunction. However, they did not specify how largea network should be for a particular problemcomplexity. Several algorithms have been pro-posed to find proper network size, e.g., construc-

w x w xtive algorithms 38 , network pruning 39 . Regu-w xlarization 40 is also a technique used to match

the model complexity with problem complexity.Rational functions have also been proven to uni-versally approximate any real-valued functions. In

w x41 , a network architecture that uses a rationalfunction to construct a mapping neural networkhas been proposed. The complexity of the archi-tecture is still considered as a major drawbackof the rational function approach, although itrequires fewer parameters than a polynomialfunction.

B. Neural Network Structureswith Prior Knowledge

Since MLP and RBF belong to the type of blackbox models structurally embedding no problemdependent information, the entire informationabout the application comes from training data.Consequently, a large amount of training data isusually needed to ensure model accuracy. In mi-crowave applications, obtaining a relatively largerset of training data by either EM]physics simula-tion, or by measurement, is expensive andrordifficult. The underlying reason is that simula-tion]measurement may have to be performed formany combinations of geometrical]material]process parameters. On the contrary, if we try toreduce the training data, the resulting neuralmodels may not be reliable. Neural networkstructures with prior knowledge address this prob-lem. There are two approaches to the use of priorknowledge during neural model development pro-cess. In the first approach, the prior knowledge isused to define a suitable preprocessing of thesimulation]measurement data such that the in-put]output mapping is simplified. The firstmethod in this approach, a hybrid EM-ANN

w xmodel was proposed in 3 . Existing approximatemodels are used to construct the input]outputrelationship of the microwave component. TheEM simulator data is then used to develop aneural network model to correct for the differ-

Žence between the approximate model source.model and the actual EM simulation results, and

as such the method is also being called the dif-ference method. In this way, the complexity of theinput]output relationship that neural network hasto learn is considerably reduced. This reductionin complexity helps to develop an accurate neural

w xnetwork model with less training data 3 . Thesecond method in this approach, is the prior

Ž . w xknowledge input PKI method 9 . In this method,the source model outputs are used as inputs forthe neural network model in addition to the origi-nal problem inputs. As such, the input]outputmapping that must be learned by a neural net-work is that between the output response of the

Page 6: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 221

source model and that of the target model. Inthe extreme case, where the target outputs arethe same as the source model outputs, the learn-ing problem is reduced to a one-to-one mapping.

w xIt was shown 9 that PKI method performs betterthan the difference method for 2-port GaAs mi-crostrip ground via.

The second approach to incorporate priorknowledge, is to incorporate the knowledge di-rectly into neural network internal structure, e.g.,w x42 . This knowledge provides additional informa-tion of the original problem, which may not beadequately represented by the limited trainingdata. The first method of this approach, usessymbolic knowledge in the form of rules to estab-lish the structure and weights in a neural networkw x42]44 . The weights, e.g., the certainty factor

w xassociated with rules 45 or both the topologyand weights of the network can be revised during

w xtraining 46 . The second method of this approachuses prior knowledge to build a modular neuralnetwork structure, i.e., to decide the number ofmodules needed, and the way in which the mod-

w xules interact with each other 47, 48 . Anotherneural network structure worth mentioning here

Ž . w xis the local model network LMN 49]52 whichis an engineering-oriented network. It is based onthe decomposition of a nonlinear dynamicsystem’s operating range into a number of smalleroperating regimes, and the use of simple localmodels to describe the system within each regime.The third method restricts the network architec-ture through the use of local connections andconstraining the choice of weights by the use of

w xweight sharing 36 . These existing approaches toincorporate knowledge are largely using symbolicinformation which exist in the pattern recognitionarea. In the microwave modeling areas however,the important problem knowledge is usually avail-able in the form of functional empirical modelsw x53, 54 .

w xIn 1 , a new microwave-oriented knowledge-Ž .based neural network KBNN was introduced. In

this method the microwave knowledge in the formof empirical functions or analytical approxima-tions is embedded into the neural network struc-ture. This structure was inspired from the factthat practical empirical functions are usually validonly in a certain region of the parameter space.To build a neural model for the entire space,several empirical formulas and the mechanism toswitch between them are needed. The switchingmechanism expands the feature of the sigmoidal

w xradial basis function 55 into high-dimensionalspace and with more generalized activation func-tions. This model retains the essence of neuralnetworks in that the exact location of each switch-ing boundary, and the scale and position of eachknowledge function are initialized randomly andthen determined eventually during training. The

Ž .knowledge-based neural network KBNN struc-ture is a nonfully connected structure as shown inFigure 2. There are six layers in the structure,namely, input layer X, knowledge layer Z, bound-ary layer B, region layer R, normalized regionlayer RX, and output layer Y. The input layer Xaccepts parameters x from outside the model.The knowledge layer Z is the place where mi-

Ž .Figure 2. Knowledge-based neural network KBNN structure.

Page 7: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.222

crowave knowledge resides in the form of singleŽ .or multidimensional functions c . .

The output of the ith knowledge neuron is givenby,

Ž . Ž .z s c x, w , i s 1, 2, . . . , N , 16i i i z

where x is a vector including neural networkinputs x , i s 1, 2, . . . , N and w is a vector ofi x iparameters in the knowledge formula. The knowl-

Ž .edge function c x, w is usually in the form ofi iempirical or semi-analytical functions. For exam-ple, the drain current of a FET is a function of itsgate length, gate width, channel thickness, doping

w xdensity, gate voltage, and drain voltage 54 . Theboundary layer B can incorporate knowledge inthe form of problem dependent boundary func-

Ž .tions B . ; or in the absence of boundary knowl-edge just as linear boundaries. Output of the ithneuron in this layer is calculated by,

Ž . Ž .b s B x, n , i s 1, 2, . . . , N , 17i i i b

where n is a vector of parameters in B definingi ian open or closed boundary in the input space x.

Ž .Let s . be a sigmoid function. The region layerR contains neurons to construct regions fromboundary neurons,

Nb

Ž .r s s a b q u , i s 1, 2, . . . , N ,Łi i j j i j rjs1

Ž .18

where a and u are the scaling and bias param-i j i jeters, respectively. The normalized region layer RX

w xcontains rational function based neurons 41 tonormalize the outputs of region layer,

riX Ž .X Xr s , i s 1, 2, . . . , N , N s N . 19i r r rNrÝ rjs1 j

The output layer Y contains second-order neu-w xrons 56 combining knowledge neurons and nor-

malized region neurons,

N N Xz rXy s b z r r q b ,Ý Ýj ji i ji k k j0ž /is1 ks1

Ž .j s 1, 2, . . . , N , 20y

where b reflects the contribution of the ithjiknowledge neuron to output neuron y and b isj j0the bias parameter. r is one indicating thatji k

region rX is the effective region of the ith knowl-kedge neuron contributing to the jth output. Atotal of N X regions are shared by all the outputrneurons. Training parameters w for the entireKBNN model includes,

ww s w , i s 1, . . . , N ; n , i s 1, . . . , N ;i z i b

a , u , i s 1, . . . , N , j s 1, . . . , N ;i j i j r b

b , j s 1, . . . , N , i s 0, . . . , N ;ji y z

r , j s 1, . . . , N , i s 1, . . . , N ,ji k y z

x Ž .Xk s 1, . . . , N . . 21r

The prior knowledge in KBNN gives it moreinformation about the original microwave prob-lem, beyond that in the training data. Therefore,such a model has better reliability when trainingdata is limited or when the model is used beyondtraining range.

C. Combining Neural Networks

In the neural network research community, arecent development called combining neural net-works is presently proposed, addressing issues ofnetwork accuracy and training efficiency. Twocategories of approaches have been developed:ensemble-based approach and modular approachw x w x57 . In the ensemble-based approach 57, 58 ,several networks are trained such that each net-work approximates the overall task in its own way.The outputs from these networks are then com-bined to produce a final output for the combinednetwork. The aim is to achieve a more reliableand accurate ensemble output than would beobtained by selecting the best net. Optimal linear

Ž .combinations OLCs of neural networks werew xproposed and investigated in 58]60 , which is

constructed by forming weighted sums of the cor-responding outputs of the individual networks.

w xThe second category, e.g., 36, 61 , features amodular neural network structure that resultsfrom the decomposition of tasks. The decomposi-

Žtion may be either automatic based on the blindapplication of a data partitioning algorithm, such

w x.as hierarchical mixtures-of-experts 62 or ex-Žplicit based on prior knowledge of the task or the

wspecialist capabilities of the modules, e.g., 47,x.48 . The modular neural network consists of sev-

eral neural networks, each optimized to performa particular subtask of an overall complex opera-tion. An integrating unit then selects or combinesthe outputs of the networks to form the final

Page 8: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 223

output of the modular neural network. Thus, themodular approach not only simplifies the overall

w xcomplexity of the problem 63 , but also facilitatesincorporation of problem knowledge into the net-work structure. This leads to improved overallnetwork reliability andror training efficiencyw x48, 64 .

A new hierarchical neural network approachfor the development of the library of microwaveneural models, motivated by the concept of com-

w xbining neural networks, was proposed in 65 . Thisapproach is targeted for massively developingneural network models for building libraries ofmicrowave models. Library development is ofpractical significance, since the realistic power ofmany CAD tools depends upon the richness,speed, and the accuracy of their library models.In this approach, a distinctive set of base neuralmodels is established. The basic microwave func-tional characteristics common to various modelsin a library are first extracted and incorporatedinto base neural models. A hierarchical neuralnetwork structure, as shown in Figure 3, is con-structed for each model in the library with lowerlevel modules realized by base neural models.The purpose of this structure is to construct anoverall model from several modules so that thelibrary base relationship can be maximally reusedfor every model throughout the library. For eachlower level module, an index function selects theappropriate base model, and a structural knowl-edge hub selects inputs relevant to the base model

Figure 3. The hierarchical neural network structure.X and Y represent the inputs and outputs of the overallnetwork. L is the ith low level module with an associ-i

Ž .ated ith knowledge hub U . . u and z represent theiinputs and outputs of low-level modules.

out of the whole input vector based on the con-figuration of the particular library component.The lower level neural modules recall the trainedbase models in the library. The higher level mod-ule realized by another neural network, models amuch easier relationship than the original rela-tionship since most of the information is alreadycontained in the base models in the lower level.For example, even a linear two-layer perceptronmight be sufficient. Consequently, the amount ofdata needed to train this higher level module ismuch less than that required for training standardMLP to learn the original problem.

The overall library development is summarizedin the following steps:

Step 1. Define the input and output spaces of thebase models, and extract basic characteristics fromthe library, using microwave empirical knowledgeif available.

Step 2. Collect training data corresponding toeach base model inputs and outputs.

Step 3. Construct and train base neural modelsincorporating the knowledge from Step 1.

Step 4. Take one unmodeled component from thelibrary. According to the base model input spacedefinition in Step 1, set up the structural knowl-edge hubs, which maps the model input spaceinto base model input space. This automaticallysets up the lower level modules.

Step 5. Collect training data corresponding to themodel in the library.

Step 6. Preprocess the training data by propagat-ing the inputs through knowledge hubs and lowerlevel modules.

Step 7. Train the higher level neural module frompreprocessed training data.

Step 8. If all done, then stop, otherwise proceedto train the next library model and go to Step 4.

The algorithm described above permits the hier-archical neural models to be developed systemati-cally, and enables the library development pro-cess to be maximally automated. Examples oftransmission line neural model libraries, usefulfor the design of high-speed VLSI interconnects,

w xwere developed in 65 . Compared to standard

Page 9: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.224

neural model techniques, the hierarchical neuralnetwork approach substantially reduces the costof library development through reduced need fordata generation and shortened time of training,while yielding reliable neural models.

D. Other Neural Network Structures

As mentioned earlier, one of the major problemsin the construction of neural network structure, isthe determination of the number of hidden neu-rons. Many techniques have been proposed toaddress this problem. In this section, some of thestructures are reviewed, with emphasis on waveletnetworks which is a very systematic approachalready used in microwave area, cascade-correla-tion network, and projection pursuit network, thelater two networks being popular constructivenetworks.

The idea of combining wavelet theory withneural networks has been recently proposedw x66]68 . Though the wavelet theory has offeredefficient algorithms for various purposes, theirimplementation is usually limited to wavelets ofsmall dimension. It is known that neural networksare powerful tools for handling problems of largedimension. Combining wavelets and neural net-works can hopefully remedy the weakness of eachother, resulting in networks with efficient con-structive methods and capable of handling prob-lems of moderately large dimension. This resultedin a new type of neural networks, called waveletnetworks which use wavelets as the hidden neu-ron activation functions. Wavelet networks arefeedforward networks with one hidden layer. Thehidden neurons are computed as,

1 Ž Ž .. Ž .y s c d x y T , i s 1, . . . , N , 22i i i 1

Ž .where c . is the radial type mother waveletfunction, d are dilation parameters, and T arei itranslation vectors for the ith hidden waveletneuron. Both d and T are adapted together withi i

Ž .n and b of Eq. 7 during training.k i kDue to the similarity between adaptive dis-

cretization of the wavelet decomposition andone-hidden-layer neural networks, there is an ex-plicit link between the network parameters suchas d , T , and n , and the decomposition formula.i i k iAs such, the initial values of network parameterscan be estimated from the training data usingdecomposition formulas. However, if the initial-ization uses regularly truncated wavelet framesw x67 , many useless wavelets on the wavelet lattice

may be included in the network, resulting inlarger network size. Alternative algorithms forwavelet network construction were proposed inw x66 to better handle problems of large dimension.

Ž .The number of wavelets hidden neurons wasconsiderably reduced by eliminating the waveletswhose supports do not contain any data points.Some regression techniques, such as stepwise se-lection by orthogonalization and backward elimi-nation, were then used to further reduce thenumber of wavelets. The wavelet networks withradial wavelet functions can be considered asRBF networks, since both of them have the local-ized basis functions. The difference is that thewavelet function is localized both in the input and

w xfrequency domains 68 . Besides retaining the ad-vantage of faster training, wavelet networks havea guaranteed upper bound on the accuracy of

w xapproximation 69 with a multiscale structure.Wavelet networks have been used in nonparamet-

w xric regression estimation 66 and were trainedbased on noisy observation data to avoid the

w x w xproblem of undesirable local minima 69 . In 14 ,wavelet networks and stepwise selection by or-thogonalization regression technique were usedto build a neural network model for a 2.9 GHzmicrowave power amplifier. In this technique, alibrary of wavelets was built according to thetraining data and the wavelet that best fits thetraining data was selected. Later in an iterativemanner, wavelets in the remainder of the librarythat best fits the data in combination with thepreviously selected wavelets were selected. Forcomputational efficiency, later selected waveletswere orthonormalized to earlier selected ones.

Besides wavelet network, there are a numberof constructive neural network structures, themost representative among them being the cas-

Ž . w xcade correlation network CasCor 70 . A CasCorŽnetwork begins with a minimal network without

.hidden neurons , then automatically adds hiddenneurons one-by-one during training. Each newlyadded hidden neuron receives a connection fromeach of the network’s original inputs and alsofrom every pre-existing hidden neuron, thus re-sulting in a multilayer network. For regression

w xtasks, a CasPer algorithm 71 which constructs aneural network structure in a similar way as Cas-Cor was proposed. CasPer does not use the maxi-mum correlation training criterion of CasCor,which tends to produce hidden neurons that satu-rate, thus making Cascor more suitable for classi-fication tasks rather than regression tasks. How-ever, both CasCor and CasPer may lead to very

Page 10: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 225

deep networks and high fan-in to the hiddenneurons. A towered cascade network was pro-

w xposed in 72 to alleviate this problem.Another constructive neural network structure

Ž .is the projection pursuit learning network PPLN ,which adds neurons with trainable activationfunctions one-by-one, within a single hidden layer

w xwithout cascaded connections 73 . The CasCorlearns the higher order features using cascadedconnection while PPLN learns it using the train-able activation functions. Every time a new hid-den neuron is added to PPLN, it first trains thenew neuron by cyclically updating the output layerweights, the smooth activation function, and theinput layer weights associated with this neuron.Then a backfitting procedure is employed to finetune the parameters associated with the existinghidden neurons. PPLN is able to avoid the curseof dimensionality by interpreting high-dimen-sional data through well-chosen low-dimensionallinear projections.

IV. TRAINING ALGORITHMS

A. Training Objective

A neural network model can be developedthrough a process called training. Suppose the

�Ž .training data consists of N sample pairs, x , d ,p p p4p s 1, 2, . . . , N , where x and d are N - andp p p x

N -dimensional vectors representing the inputsyand the desired outputs of the neural network,respectively. Let w be the weight vector contain-ing all the N weights of the neural network. Forwexample, for MLP neural network w is given byŽ . Ž .12 , and for KBNN w is given by 21 .

The objective of training is to find w such thatthe error between the neural network predictionsand the desired outputs are minimized,

Ž . Ž .min E w , 23w

where

N Np y1 2Ž . Ž .E w s y x , w y dŽ .Ý Ý pk p pk2 ps1 ks1

Np1Ž . Ž .s e w , 24Ý p2 ps1

Ž .and d is the k th element of vector d , y x , wpk p pk pis the k th output of the neural network when theinput presented to the network is x . The termp

Ž .e w is the error in the output due to the pthpsample.

Ž .The objective function E w is a nonlinearfunction w.r.t. the adjustable parameter w. Due to

Ž .the complexity of E w , iterative algorithms areoften used to explore the parameter space effi-ciently. In iterative descent methods, we startwith an initial guess of w and then iterativelyupdate w. The next point of w, denoted as w , isnextdetermined by a step down from the current pointw along a direction vector d,now

Ž .w s w q hd, 25next now

where h is a positive step size regulating theextent to which we can proceed in that direction.Every training algorithm has its own scheme forupdating the weights of the neural network.

B. Backpropagation Algorithmand Its Variants

One of the most popular algorithms for neuralŽ .network training is the backpropagation BP al-

w xgorithm 74 , proposed by Rumelhart, Hinton,and Williams in 1986. The BP algorithm is astochastic algorithm based on the steepest de-

w xscent principle 75 , wherein the weights of theneural network are updated along the negativegradient direction in the weight space. The up-dated formulas are given by,

Ž .­ E wDw s w y w s yh ,now next now

wsw­ w now

Ž .26a

Ž .­ e wpDw s w y w s yh ,now next now

wsw­ w now

Ž .26b

where h called learning rate controls the step sizeŽ .of weight update. Update formula 26b is called

update sample-by-sample, where the weights areupdated after each sample is presented to the

Ž .network. Update formula 26a is called batchmode update, where the weights are updatedafter all training samples have been presented tothe network.

The basic backpropagation, derived from theprinciples of steepest descent, suffers from slowerconvergence and possible weight oscillation. Theaddition of a momentum term to weight update

Ž . Ž . w xformulas in 26a and 26b as proposed by 74 ,

Page 11: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.226

provided significant improvements to the basicbackpropagation, reducing the weight oscillation.Thus,

Ž .­ E wDw s yh q aDwnow old

wsw­ w now

Ž .­ E wŽ .s yh q a w y w ,now old

wsw­ w now

Ž .27a

Ž .­ e wpDw s yh q aDwnow old

wsw­ w now

Ž .­ e wp Ž .s yh q a w y w ,now oldwsw­ w now

Ž .27b

where a is the momentum factor which controlsthe influence of the last weight update directionon the current weight update, and w representsoldthe last point of w. This technique is also known

w xas the generalized delta-rule 36 . Other ap-proaches to reduce weight oscillation have alsobeen proposed, such as invoking a correction

w xterm that uses the difference of gradients 76 ,and the constrained optimization approach whereconstraints on weights are imposed to achievebetter alignment between weight updates in dif-

w xferent epochs 77 .As neural network research moved from the

state-of-the-art paradigm to real-world applica-tions, the training time and computational re-quirements associated with training have become

w xsignificant considerations 78]80 . Some of thereal-world applications involve large-scale net-works, in which case the development of fast andefficient learning algorithms becomes extremely

w ximportant 79 . A variety of techniques have beendeveloped, and among them are two importantclasses of methods. One of them is based onadvanced learning rate and momentum adapta-tion, and heuristic rules of BP, and the other isbased on the use of advanced optimization tech-niques. The latter shall be discussed in SectionIV.C.

An important way to improve efficiency oftraining by backpropagation is to use adaptationschemes that allow the learning rate and themomentum factor to be adaptive during learningw x36 , e.g., adaptation according to training errorsw x81 . One of the most interesting works in thisarea is the delta-bar-delta rule proposed by Ja-

w xcobs 82 . He developed an algorithm based on aset of heuristics in which the learning rate fordifferent weights are defined separately and alsoadapted separately during the learning process.The adaptation is determined from two factors,one being the current derivative of the trainingerror with respect to the weights, and the otherbeing an exponentially weighted sum of the cur-rent and past derivatives of the training error.Sparsity of hidden neuron activation pattern has

w xalso been utilized in 80, 83, 84 to reduce thecomputation involved during training. Variousother adaptation techniques have also been pro-posed, for example, a scheme in which the learn-ing rate was adapted in order to reduce theenergy value of the gradient direction in a close-

w xto-optimal way 85 , an enhanced backpropaga-w xtion algorithm 86 with a scheme to adapt the

learning rate according to values of weights in theneural net, and a learning algorithm inspired fromthe principle of ‘‘forced dynamics’’ for the total

w x w xerror function 87 . The algorithm in 87 updatesthe weights in the direction of steepest descent,but with the learning rate as a specific function ofthe error and the error gradient form. An inter-esting adaptation scheme based on the concept ofdynamic learning rate optimization is presented

w xin 88 , in which the first- and second-orderderivatives of the objective function w.r.t. thelearning rate are calculated from the informationgathered during the forward and backward propa-

w xgation. Another work 76 , which is considered asan extension of Jacob’s heuristics, corrects thevalues of weights near the bottom of the errorsurface ravine with a new acceleration algorithm.This correction term uses the difference betweengradients, to reduce the weight oscillation duringtraining. In general, during neural network train-ing, the weights are updated after each iterationby a certain step size along an updating direction.The standard backpropagation uses learning rateto adjust the step size, with the advantage thatthe method is very simple and does not requirerepetitive computation of the error functions. Adifferent way to determine the step size, is to use

w xline search methods 75 , so that the trainingerror is reduced or optimized along the givenupdating direction. Examples in this category are

w xline search based on quadratic model 89 , andw xline search based on linear interpolation 90, 91 .

One other way to improve training efficiency isw xthe gradient reuse algorithm 92 . The basic idea

of this method is that gradients which are com-puted during training are reused until the result-

Page 12: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 227

ing weight updates no longer lead to a reductionin the training error.

C. Training Algorithms UsingGradient-Based Optimization Techniques

The backpropagation based on the steepest de-scent principle is relatively easy to implement.However, the error surface of neural networktraining usually contains planes with a gentleslope due to the squashing functions commonlyused in neural networks. This gradient is toosmall for weights to move rapidly on these planes,thus reducing the rate of convergence. The rateof convergence could also be very slow when thesteepest descent method encounters ‘‘narrow val-ley’’ in the error surface where the direction ofthe gradient is close to the perpendicular direc-tion of the valley. The update direction oscillatesback and forth along the local gradient.

Since supervised learning of neural networkscan be viewed as a function optimization prob-lem, higher order optimization methods usinggradient information can be adopted in neuralnetwork training to improve the rate of conver-gence. Compared to the heuristic approach dis-cussed in the earlier backpropagation section,these methods have a sound theoretical basis andguaranteed convergence for most of the smoothfunctions. Some of the early work in this area was

w xdemonstrated in 93, 94 with the development ofsecond-order learning algorithms for neural net-

w xworks. Papers 90, 95 reviewed the first- andsecond-order optimization methods for learningin feedforward neural networks.

Let d be the direction vector, h be the learningrate, w be the current value of w, then thenowoptimization updates w such that,

Ž . Ž . Ž . Ž .E w s E w q hd - E w . 28next now now

The principal difference between various descentalgorithms lies in the procedure to determine

Ž . w xsuccessive update directions d 96 . Once theupdate direction is determined, the optimal stepsize could be found by line search,

U Ž . Ž .h s min f h , 29h)0

where

Ž . Ž . Ž .f h s E w q hd . 30now

When downhill direction d is determined fromthe gradient g of the objective function E, suchdescent methods are called gradient-based de-scent methods. The procedure for finding a gradi-ent vector in a network structure is generally

w xsimilar to backpropagation 74 in the sense thatthe gradient vector is calculated in the directionopposite to the flow of output from each neuron.For MLP as an example, this is done by means ofa derivative chain rule starting from output layer,

l­ E ­ E ­ y ­ yŽ .s ? ? , l s L, 31al l l­ y­ w ­ y ­ w

and then through the various layers down towardthe input layer,

lq1 l­ E ­ E ­ y ­ y ­ ys ? ??? ? ,l L l l­ y­ w ­ y ­ y ­ w

Ž .l s L y 1, L y 2, . . . , 1, 31b

where y represents the final outputs of the neurallnetwork, and y represents the outputs of the lth

hidden layer of the neural network.

1. Conjugate Gradient Training Algorithms. Theconjugate gradient methods are originally derivedfrom quadratic minimization and the minimum ofthe objective function E can be efficiently foundwithin N iterations. With initial gradient g sinitial­ Er­ w N , and direction vector d swsw initialinitial

yg , the conjugate gradient method recur-initialw xsively constructs two vector sequences 91 ,

Ž .g s g q l Hd , 32next now now now

Ž .d s yg q g d , 33next next now now

gT gnow now Ž .l s , 34now Td Hdnow now

gT gnext next Ž .g s , 35now Tg gnow now

or,

TŽ .g y g gnext now next Ž .g s , 36now Tg gnow now

where d is called the conjugate direction and H isthe Hessian matrix of the objective function E.

Ž .Here, 35 is called the Fletcher]Reeves formulaŽ .and 36 is called the Polak]Ribiere formula. To

Page 13: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.228

avoid the need of the Hessian matrix to computethe conjugate direction, we proceed from wnowalong the direction d to the local minimum ofnowE at w through line minimization, and then setnext

<g s ­ Er­ w . This g can be used aswswnext nextnext

Ž . Ž .the vector of 32 , and as such 34 is no longerneeded. We make use of this line minimizationconcept to find the conjugate direction in neuralnetwork training, thus avoiding intensive Hessianmatrix computations. In this method, the descentdirection is along the conjugate direction whichcan be accumulated without computations involv-ing matrices. As such, conjugate gradient meth-ods are very efficient and scale well with theneural network size.

Two critical issues have to be considered inapplying conjugate gradient methods to neuralnetwork learning. First, computation requiredduring the exact one-dimensional optimization isexpensive because every function evaluation in-volves the neural network feedforward operationfor a complete cycle of samples. Therefore, effi-cient approximation in one-dimensional optimiza-tion has to be used. Second, since for neuralnetwork training, the error function is not

Ž .quadratic w.r.t. the variable as defined in 24 , theconvergence properties of the method are notassured a priori but depend on the degree towhich a local quadratic approximation can be

w xapplied to the training error surface. In 85 ,inexact line search was proposed and a modifieddefinition of the conjugate search direction wasused to achieve this purpose. To further reducecomputational complexities, a scaled conjugate

Ž . w xgradient SCG algorithm was introduced in 97which avoids the line search per learning iterationby using the Levenberg]Marquardt approach toscale the step size.

2. Quasi-Newton Training Algorithms. Similar tothe conjugate gradient method, the quasi-Newtonmethod was derived from quadratic objectivefunction. The inverse of the Hessian matrix B sHy1 is used to bias the gradient direction, follow-ing Newton’s method. In the quasi-Newton train-ing method, the weights are updated using,

Ž .w s w y hB g . 37next now now now

The B matrix here is not computed. It is succes-sively estimated employing rank 1 or rank 2 up-dates, following each line search in a sequence of

w xsearch directions 98 ,

Ž .B s B q DB . 38now old now

There are two major rank 2 formulas to computeDB ,now

ddT B Dg DgT Bold old Ž .DB s y , 39now T Td Dg Dg B Dgold

or,

DgT B Dg ddTold

DB s 1 qnow T Tž /d Dg d Dg

dDgT B q B DgdTold old Ž .y , 40Td Dg

where

Ž .d s w y w , Dg s g y g , 41now old now old

Ž . Ž .39 is called the DFP Davidon]Fletcher]PowellŽ . Žformula and 40 is called the BFGS Broyden]

.Fletcher]Goldfarb]Shanno formula.Standard quasi-Newton methods require N 2

wstorage space to maintain an approximation ofthe inverse Hessian matrix and a line search isindispensable to calculate a reasonably accuratestep length, where N is the total number ofwweights in the neural network structure. Limited-

Ž .memory LM or one-step BFGS is a simplifica-tion in which the inverse Hessian approximationis reset to the identity matrix after every iteration,

w xthus avoiding the need to store matrices. In 99a second-order learning algorithm is proposedbased on a LM BFGS update. A reasonably ac-curate step size is efficiently calculated in a one-dimensional line search by a second-order ap-proximation of the objective function. Parallelimplementation of second-order gradient-basedMLP training algorithms featuring full and lim-ited memory BFGS algorithms were presented inw x100 . Wavelet neural networks trained by theBFGS algorithm are used for the modeling oflarge-signal hard-nonlinear behavior of power

w xtransistors in circuit design 14 . The quasi-New-ton training algorithm that employs the exact linesearch possesses the quadratic termination prop-erty. Through the estimation of the inverse Hes-sian matrix, quasi-Newton has a faster conver-gence rate than the conjugate gradient method.

3. Le©enberg–Marquardt and Gauss-Newton Train-ing Algorithms. Neural network training is usuallyformulated as a nonlinear least-squares problem.Methods dedicated to least-squares such as

Page 14: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 229

Gauss]Newton can be employed to estimate theneural model parameters. The Gauss]Newtonmethod is a linearization method. Let r be avector containing the individual error terms inŽ . Ž .24 . Let J be the N = N = N Jacobian ma-p y wtrix including the derivative of r w.r.t. w. J has Nwcolumns and N = N rows, where N is thep y pnumber of samples and N is the number ofyoutputs. The Gauss]Newton update formula canbe expressed as,

y1T TŽ . Ž .w s w y J J J r . 42next now now now now now

Ž . TIn the preceding formula 42 J J is posi-now nowtive definite unless J is rank deficient. Thenow

w xLevenberg]Marquardt 101 method can be ap-plied when J is rank deficient and the corre-nowsponding weight update is given by,

y1T TŽ . Ž .w s w y J J q mI J r , 43next now now now now now

where m is a nonnegative number. A modifiedLevenberg]Marquardt training algorithm using adiagonal matrix instead of the identity matrix I inŽ . w x43 was proposed 102 for the efficient trainingof multilayer feedforward neural networks. Inw x103 , to reduce the size of the Jacobian matrix,the training samples are divided into severalgroups called local batches. The training is per-formed successively through these local batches.The computational requirement and memorycomplexity of the Levenberg]Marquardt methodswere reduced by utilizing the deficient Jacobian

w xmatrix 104 . A combined Levenberg]Marquardtand quasi-Newton training technique was used inw x1 to train the KBNN structure. When trainingparameters are far from the local minimum of theerror surface Levenberg]Marquardt algorithmwas used, and when they are close to local mini-mum quasi-Newton was used for faster conver-gence.

It has been proved that the conjugate gradientmethod is equivalent to error backpropagation

w xwith momentum term 90 . The theoretical con-vergence rate and practical performance of sec-ond-order gradient-based methods are generallysuperior to the first-order methods.

D. Training Algorithms UtilizingDecomposed Optimization

As seen in the earlier discussions, implementa-tion of powerful second-order optimization tech-niques for neural network training has resulted in

significant advantages in training. The second-order methods are typically much faster than BPbut could require the storage of the inverse Hes-sian matrix, and its computation or an approxima-tion thereof. For large neural networks, trainingcould be a very large scale optimization. Decom-position is an important way to solve the largescale optimization problems. Several training al-gorithms that decompose the training task bytraining the neural network layer-by-layer have

w x w x Ž .been proposed 105]107 . In 106 , the weights VŽ L.of the output layer and the output vector y of

the previous layer are treated as two sets ofŽ L.variables. An optimal solution pair V, y is first

determined to minimize the sum-squared-errorbetween the desired neural network outputs andthe actual outputs. The current solution y L isthen set as the desired output of the previoushidden layer, and optimal weight vectors of thehidden layers are recursively obtained. In the case

w xof the continuous function approximation, 107optimizes each layer with an objective to increasea measure of linearity between the internal repre-sentations and the desired output.

Linear programming can be used to solve largescale linearized optimization problems. Neuralnetwork training was linearized and formulated as

w xa constrained linear programming in 108 . In thiswork, weights are updated with small local changessatisfying the requirement that none of the indi-vidual sample errors should increase, subject tothe constraint of maximizing the overall reductionin the error. However, extension of such a methodfor efficient implementation in large networks

w xneeds special considerations. In 105 , a layer-by-layer optimization of a neural network with lin-earization of the nonlinear hidden neurons waspresented, which does not rely on the evaluationof local gradients. To limit the unavoidable lin-earization error, a special penalty term is addedto the cost function and layers are optimizedalternately in an iterative process.

A combination of linear]nonlinear program-ming techniques could reduce the degree of non-linearity of the error function with respect to thehidden layer weights and decrease the chances ofbeing trapped in a local minima. As such, a hy-brid training algorithm would be favorable for a

w xlarge-scale problem 52 . For a feedforward neu-ral network with the linear or linearized outputlayer weights, training algorithms were developedby adapting the nonlinear hidden layer weights

w x w xusing BP 109 or BFGS 52 , while employing aŽ .linear least mean square error LMS algorithm

Page 15: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.230

to compute the optimum linear output layerweights.

RBF networks are usually trained by the de-composed process. The nonlinear hidden layerweights of the RBF network can be interpreted ascenters and widths, thus making it possible toorganize these neurons in a manner which re-flects the distribution of the training data. Utiliz-ing this concept, the hidden layer weights can befixed through unsupervised training, such as k-

w xmeans clustering algorithm 110 . The output layerweights can then be obtained through the linearLMS algorithm. On the other hand, the develop-ment of heuristics to initially assign the hiddenlayer weights of MLP, is very hard due to it’sblack box characteristics. Consequently, it is notpossible to train MLP neural networks with thisdecomposed strategy. However, the hybrid lin-

w xear]nonlinear training algorithm presented in 52integrates the best features of the linear LMSalgorithm of RBF, and the nonlinear optimizationtechniques of MLP into one routine. This routinecould be suitable for any feedforward networkwith linear output layer weights. The advantagesof this technique are the reduced number ofindependent parameters and guaranteed global

w xminimum w.r.t. to output layer weights 52 .

D. Global Training Algorithms

Another important class of methods use randomoptimization techniques which are characterizedby a random search element in the training pro-cess allowing the algorithms to escape from localminima and converge to the global minimum ofthe objective function. Examples in this class are,e.g., simulated annealing which allows the opti-mization to jump out of a local minimum throughan annealing process controlled by a parameter

w xcalled temperature 111 ; genetic algorithms whichevolve the structure and weights of the neuralnetwork through generations in a manner that is

w xsimilar to biological evolution 112 ; the LangevinŽ .updating LV rule in multilayer perceptron, in

which noise is added to the weights during train-w xing 113 ; and a stochastic minimization algorithm

w xfor training neural networks 114 , which is basi-cally a random optimization method with no gra-dient information needed. Since the convergenceof pure random search techniques tends to bevery slow, a more general method is the hybridmethod which combines the conventional gradi-ent-based training algorithms with random opti-

w xmization, e.g., 89 . This work introduced a hybrid

algorithm combining the conjugate gradientmethod with line search, and the random opti-mization method to find the global minimum ofthe error function. During training with the con-jugate gradient method, if a flat error surface isencountered, the training algorithm switches tothe random optimization method. After train-ing escapes from the flat error surface, it onceagain switches back to the conjugate gradientalgorithm.

V. EXAMPLES

A. Feedforward Neural Networksand Their Training

Standard feedforward neural networks, MLP andRBF, have been used in many microwave applica-tions. This section demonstrates the use of theseneural model structures in several microwave ex-amples and their training by various training algo-rithms.

MLP and RBF were used to model a physics-based MESFET. Device physical]process param-

Žeters channel length L, channel width W, doping.density N , channel thickness a and terminaldŽ .voltages, i.e., gate-source voltage V and drain-G

Ž .source voltage V , are neural network inputDparameters. Drain-current, i.e., i , is the neuraldnetwork output. The training data and test data

w xwere simulated using OSA90 115 . Three sets oftraining data with 100, 300, and 500 samples, andone set of test data with 413 samples were used.The model accuracy is shown in Table I.

Generally, RBF will need more hidden neu-rons than MLP to achieve similar model accuracydue to their localization nature of activation func-tion. RBF training requires a sufficient amount ofdata. On the other hand, the training of RBF iseasy to converge. The generalization capability ofMLP is better than RBF as seen in the table asthe available amount of training data becomesless.

The most popular training algorithms for stan-dard feedforward neural networks are adaptivebackpropagation, conjugate gradient, quasi-Newton, and Levenberg]Marquardt. Two exam-ples, i.e., 3-conductor microstrip line and physics-based MESFET, were used to illustrate theperformance of these algorithms.

For the microstrip line example, there are fiveinput neurons corresponding to conductor widthŽ . Ž .w , spacing between conductors s , s , sub-1 2

Page 16: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 231

TABLE I. Model Accuracy Comparison Between Multilayer Perceptrons and Radial Basis Functions

Model Type MLP RBFTrainingSample No. of Hidden

Size Neurons 7 10 14 18 25 20 30 40 50

%100 Avg. Test Error 1.65 2.24 2.60 2.12 2.91 6.32 5.78 6.15 8.07300 Avg. Test Error 0.69 0.69 0.75 0.69 0.86 1.37 0.88 0.77 0.88500 Avg. Test Error 0.57 0.54 0.53 0.53 0.60 0.47 0.43 0.46 0.46

Ž . Ž .strate height h , and relative permittivity « asrshown in Figure 4. There are six output neuronscorresponding to the self inductance of each con-ductor l , l , l and the mutual inductance be-11 22 33tween any two conductors l , l , l . There are12 23 13totally 600 training samples and 640 test samples

w xgenerated by LINPAR 116 . A three-layer MLPstructure with 28 hidden neurons is chosen asthe sample structure. The training results areshown in Table II. CPU time is given for PentiumŽ .200 MHz .

The total CPU time used by the Leven-berg]Marquardt method is around 20 min. Theadaptive backpropagation used many epochs andsettled down to good accuracy of 0.252%, witharound 4 h of training time. On the contrary, thequasi-Newton method achieved similar accuracyonly within 35 min. This confirms the faster con-vergence rate of the second-order method. Usu-ally quasi-Newton has a very fast convergence

Figure 4. A 3-conductor microstrip line.

rate when approaching the minimum. However,at the beginning of training, its performance maynot be very strong. Another strategy is to use theconjugate gradient method at the first stage oftraining, then followed by the quasi-Newtonmethod. If we take the MLP network alreadytrained by the conjugate gradient and then con-tinue training by the quasi-Newton method, themodel test error was reduced to 0.167%. Totaltraining time is around 2 h.

For the MESFET example, the inputs to theŽ .neural model include frequency f , channel

Ž . Ž .thickness a , gate-bias voltage V , and drain-gŽ .bias voltage V . This particular MESFET has ad

fixed gate length of 0.55 m and a gate width of1 mm. The outputs include real and imaginaryparts of S , S , S and S . Training and test11 12 21 22

samples are obtained using the simulator OSA90w x115 with the Katibzadeh and Trew model. Thisis a relatively complicated example compared tothe microstrip line example. The sample neuralnetwork structure has 60 hidden neurons. Thetraining results by the four training algorithmsare shown in Table III. This example demon-strates that as the size of the neural networkbecomes large, Levenberg]Marquardt becomes

TABLE II. Comparison of Various Training Algorithms for the Microstrip Line Example

Training No. of Training Error Avg. TestŽ . Ž . Ž .Algorithm Epochs % Error % CPU s

Adaptive 10,755 0.224 0.252 13,724backpropagation

Conjugate 2169 0.415 0.473 5511gradient

Quasi-Newton 1007 0.227 0.242 2034Levenberg] 20 0.276 0.294 1453

Marquardt

Page 17: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.232

TABLE III. Comparison of Various Taining Algorithms for the MESFET Example

Training No. of Training Error Avg. TestŽ . Ž . Ž .Algorithm Epochs % Error % CPU s

Adaptive 15,319 0.98 1.04 11,245backpropagation

Conjugate 1605 0.99 1.04 4391gradient

Quasi-Newton 570 0.88 0.89 1574Levenberg] 12 0.97 1.03 4322

Marquardt

slow as compared to the quasi-Newton method,due to repeated inversion of the large matrix

Ž .in 43 .

B. Knowledge-Based NeuralNetwork and Training

This example demonstrates the knowledge basedŽ .neural network KBNN described in Section III.B

in modeling cross-sectional resistance-inductance-Ž .capacitance-conductance RLCG parameters of

transmission lines for analysis of high-speed VLSIw xinterconnects 16 and its comparison with tradi-

Ž .tional multilayer perceptrons MLP . The outputof the model is the cross-sectional mutual induc-

Ž .tance per unit length , l , between two conduc-12tors of a coupled microstrip transmission line.The inputs of the neural models are width ofconductor x , thickness of conductor x , separa-1 2tion between two conductors x , height of sub-3strate x , relative dielectric constant x , and fre-4 5quency x . There exist mutual inductance empiri-6

w xcal formula, e.g., 53 ,

2Ž .m m 2 xr 0 4 Ž .l s ln 1 q . 4412 24p Ž .x q x1 3

This equation becomes the knowledge to be in-corporated into the knowledge neurons of SectionIII.B as,

Ž .z s c x, wi i i

2Ž .x y w4 i2w i1s ln 1 q e q w xi4 22Ž .x q x y w1 3 i3

q w x q w x q w ,i5 5 i6 6 i7

Ž .i s 1, 2, . . . , N . 45z

Linear boundary neurons were used in the layerB. Notice that this empirical formula is incorpo-

Ž .rated multiple times N times , each with differ-zŽ .ent values of w, w , i s 1, 2, . . . , N . KBNN pro-i z

vides a complete]integrated x y y relationshipincluding those not available in the original em-

Ž .pirical formula e.g., y with respect to x , x , x .2 5 6Five sets of data were generated by EM simu-

w xlation 117 . The first three sets have 100, 300,and 500 samples and were used for training pur-poses. The fourth set of another 500 samples wasgenerated in the same range as the first three setsto test the trained neural models. These testingdata were never used in training. The last set ofdata with 4096 samples were deliberately selectedaround]beyond the boundary of the model effec-tive region in the input parameter space in orderto compare extrapolation accuracy of KBNNand MLP.

Two KBNNs of different sizes were trainedŽand compared with three MLPs with the number

.of hidden neurons being 7, 15, and 20 . All theseneural networks were trained by the Levenberg]Marquardt algorithm. Figure 5 shows the errorfrom individual trainings of KBNN and MLP interms of the average testing error. The overalltendency suggests that the accuracy of KBNNtrained by a small set of training data is compara-ble to that of MLP trained by a larger set oftraining data. A much more stable performanceof KBNN over MLP is observed when making anextrapolation prediction.

C. Hierarchical Neural Networkand Its Training

This example demonstrates the usage of hierar-chical neural networks in the development of alibrary of neural models for N-conductor striplines

w xshown in Figure 6 for different values of N 65 .In the example there are five models in the li-brary, n s 1, 2, 3, 4, 5. In addition, for each nthmodel, N s n. Table IV defines the inputs andoutputs of each model.

Page 18: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 233

Figure 5. Comparison of error from individual train-ings of KBNN and MLP in terms of the average testingerror.

Base Model Selections. Two base models, B1for self inductance and B for mutual inductance2are defined. The inputs to the base models in-clude physical]geometrical parameters such as

Ž . Ž .conductor width w , conductor height g , sub-Ž .strate height h , separation between conductors

Ž . Ž .s , and relative dielectric constant « . The out-rputs of B and B are self and mutual induc-1 2tances, respectively. The training strategy for basemodels is to use the conjugate gradient at the firststage and followed by the quasi-Newton method.The base models B and B are trained to1 2an average testing accuracy of 0.39 and 0.16%,respectively.

Example of Library Model: n s 3. For librarymodel n s 3, we reuse the base models as thelow-level neural modules shown in Figure 3. Thehigh-level neural module is realized by a two-layerperceptron with six inputs and six outputs. Only a

Ž .small amount of training data 15 samples isneeded to train this 3-conductor stripline modelsince the raw]fundamental relationships of themodel have already been captured in the basemodels. Since this is a linear network, the trainingproblem is actually a quadratic minimization. Theconjugate gradient method was used here to findthe unique global minimum. However, with theconventional MLP neural model, even 500 sam-

Figure 6. N-conductor stripline component. For thenth component in the stripline library, N s n.

TABLE IV. Stripline Library Components

Library Component Neural Model Neural Model Reuse of BaseŽ .Component Name Inputs Outputs Models B and B1 2

n s 1 1-conductor w g h « L Br 11 1model

n s 2 2-conductor w w s g h « L , L , L 2 = B , 1 = B1 2 r 11 12 22 1 2model

n s 3 3-conductor w w w s s L , L , L , L , L , 3 = B , 3 = B1 2 3 1 2 11 12 13 22 23 1 2model g h « Lr 33

n s 4 4-conductor w w w w s s L , L , L , L , L , 4 = B , 6 = B1 2 3 4 1 2 11 12 13 14 22 1 2model s g h « L , L , L , L , L3 r 23 24 33 34 44

n s 5 5-conductor w w w w w L , L , L , L , L , 5 = B , 10 = B1 2 3 4 5 11 12 13 14 15 1 2model s s s s g h « L , L , L , L , L ,1 2 3 4 r 22 23 24 25 33

L , L , L , L , L34 35 44 45 55

Page 19: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.234

ples are not enough to achieve a model of similaraccuracy, as shown in Figure 7.

All Library Models. All library models in thelibrary can be developed systematically in a simi-lar way as model 3. It should be noted that effortsin developing those additional models are smalland incremental, since only few training data areneeded, and only the high-level neural modulesneed to be trained.

O¨erall Library Accuracy and De¨elopment Cost.A Comparison. Using standard MLP for eachmodel, the total training time for all library mod-els is 6 h 30 min on SparcStation 5. Using thehierarchical approach, the total training time is36 min. The total amount of training data neededby standard MLP is 2664 samples and by the

Hierarchical neural network approach is only 649Žincluding 564 samples for the base model, and 85

.samples for subsequent library models as shownin Table V. The hierarchical neural networkstructure yields reliable neural models even witha very small amount of training data.

VI. CONCLUSIONS

Neural network technology is an emerging tech-nology in the microwave area for microwave mod-eling, simulation, optimization, and design. Theefficient development of an accurate neural modelrequires a proper neural network structure andsuitable training algorithms, two important as-

Ž .Figure 7. Model accuracy comparison average error on test data between standard MLPand the hierarchical neural network models for the 3-conductor stripline component.

TABLE V. Comparison of the Number of Training Samples Needed and Model Accuracyfor the Stripline Library when Developed by Standard Multilayer Perceptronsand the Hierarchical Neural Network Structure, respectively

Model AccuracyStriplineŽ .No. of Training Samples Needed Average Error on Test Data %Component

Name Standard MLP Hierarchical NN Standard MLP Hierarchical NN

1 2Overhead 0 264 q 3001-conductor 264 0 0.42 0.39

model2-conductor 300 10 1.01 0.56

model3-conductor 500 15 1.25 0.40

model4-conductor 700 25 1.30 0.79

model5-conductor 900 35 0.99 0.63

modelLibrary Total s 2664 Total s 649 Average s 0.99 Average s 0.55

1Base model B training.12 Base model B training.2

Page 20: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 235

pects in successful applications of neural net-works in solving microwave design problems. Thispaper presented a review of the current status ofthis area. The subject of neural network architec-tures and training algorithms in general is largeand we have omitted many advances suitable forother applications such as in signal processing,pattern recognition, and so on. Instead, our paperfocuses on the structures and algorithms whichwe feel are relevant or useful to RF and mi-crowave applications. Standard feedforward neu-ral networks, neural network structures with priorknowledge, combining neural networks and con-structive network structures, are described. Alsodiscussed are various training algorithms includ-ing the backpropagation algorithm and its vari-ants, training algorithms based on classical opti-mization techniques such as conjugate gradientand quasi-Newton algorithms, training algorithmsbased on decomposed optimization, and globalminimization techniques.

Neural networks have a very promising futurein the microwave design area. Benefits of apply-ing neural network technology can be potentiallyachieved at all levels of microwave design fromdevice, components, to circuits and systems, andfrom modeling, simulation, to optimization andsynthesis. From the research point of view, futurework in structures and training algorithms willshift from demonstration of basic significance ofthe neural network technology to addressing chal-lenges from real microwave applications. Model-ing of complicated 3D EM problems is one ofsuch work. In this case, the cost of generatingtraining data by EM simulations is very high. Howto develop a reliable neural model with a verysmall amount of data remains an important re-search. In many practical cases, training datafrom simulation]measurement contains acciden-tal but large errors due to convergence difficultiesin simulators or equipment limits that may hap-pen when data generation goes to extreme pointsin the parameter space. Existing training algo-rithms can be susceptible to such large errors,and consequently the neural model obtained isnot reliable. Robust algorithms automaticallydealing with such cases need to be developed,avoiding manual debugging for clues of modelinaccuracy. Filter design with full EM simulationis one of the important, yet difficult tasks formany engineers. Research on neural networks tohelp for such designs is already underway. Thehighly nonlinear, and nonsmooth relationship in

such microwave models needs to be addressed byan effective neural network structure. Using stan-dard structures, more neurons are typicallyneeded for such cases leading to the requirementof more training data, and higher accuracy isdifficult to obtain. Another scenario leading tothe same challenge is when models contain manyvariables, for example, many geometrical andphysical parameters in a microwave model. Ad-dressing these challenges will be an importantdirection in future research. The potential ofcombining microwave and circuit information withneural networks, continues to motivate researchleading to advanced knowledge-based neuralmodels. Another significant milestone in this areawould be to incorporate the microwave-orientedfeatures and techniques of this emerging technol-ogy into readily usable software tools. These toolswould enable more microwave engineers toquickly benefit from this technology, and theirfeedback could further stimulate advanced re-search in the area. Neural networks, with theirunparalleled speed advantage, and their ability tolearn and generalize wide variety of problems,promise to be one of the powerful vehicles help-ing microwave design today and tomorrow.

ACKNOWLEDGMENTS

The authors thank Marco Antonio Pego Guerra ofCarleton University for his cooperation in running

w xNeuroModeler 118 for Example 1. This work was sup-ported through the Natural Sciences and EngineeringResearch Council of Canada, and Micronet: a Cana-dian Network Centres of Excellence on Microelec-tronic Devices, Circuits and Systems. The authors arewith the Department of Electronics, Carleton Univer-sity, 1125 Colonel By Drive, Ottawa, Canada K1S 5B6.

REFERENCES

1. F. Wang and Q. J. Zhang, Knowledge based neu-ral models for microwave design, IEEE Trans

Ž .Microwave Theory Tech 45 1997 , 2333]2343.2. A. H. Zaabab, Q. J. Zhang, and M. S. Nakhla,

Neural network modeling approach to circuit opti-mization and statistical design, IEEE Trans Mi-

Ž .crowave Theory Tech 43 1995 , 1349]1358.3. P. M. Watson and K. C. Gupta, EM-ANN models

for microstrip vias and interconnects in datasetcircuits, IEEE Trans Microwave Theory Tech 44Ž .1996 , 2495]2503.

Page 21: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.236

4. P. M. Watson and K. C. Gupta, Design and opti-mization of CPW circuits using EM-ANN modelsfor CPW components, IEEE Trans Microwave

Ž .Theory Tech 45 1997 , 2515]2523.5. G. L. Creech, B. J. Paul, C. D. Lesniak, T. J.

Jenkins, and M. C. Calcatera, Artificial neuralnetworks for accurate microwave CAD applica-tions, IEEE Int Microwave Symp Digest, SanFrancisco, CA, 1996, pp. 733]736.

6. M. Vai, S. Wu, B. Li, and S. Prasad, Creatingneural network based microwave circuit modelsfor analysis and synthesis, Proc Asia Pacific Mi-crowave Conf, Hong Kong, Dec. 1997, pp.853]856.

7. M. Vai and S. Prasad, Microwave circuit analysisand design by a massively distributed computingnetwork, IEEE Trans Microwave Theory Tech 43Ž .1995 , 1087]1094.

8. Q. J. Zhang and M. S. Nakhla, Signal integrityanalysis and optimization of VLSI interconnectsusing neural network models, IEEE Int SympCircuits Systems, London, U.K., 1994, pp. 459]462.

9. P. M. Watson, K. C. Gupta, and R. L. Mahajan,Development of knowledge based artificial neuralnetwork models for microwave components, IEEEInt Microwave Symp Digest, Baltimore, MD, 1998,pp. 9]12.

10. G. L. Creech, B. J. Paul, C. D. Lesniak, T. J.Jenkins, and M. C. Calcatera, Artificial neuralnetworks for fast and accurate EM-CAD of mi-crowave circuits, IEEE Trans Microwave Theory

Ž .Tech 45 1997 , 794]802.11. V. B. Litovski, J. I. Radjenovic, Z. M. Mrcarica,

and S. L. Milenkovic, MOS transistor modelingŽ .using neural network, Electron Lett 28 1992 ,

1766]1768.12. A. H. Zaabab, Q. J. Zhang, and M. S. Nakhla,

Analysis and optimization of microwave circuitsand devices using neural network models, IEEEInt Microwave Symp Digest, San Diego, CA, May1994, pp. 393]396.

13. G. Kothapalli, Artificial neural networks as aids inŽ .circuit design, Microelectron J 26 1995 , 569]578.

14. Y. Harkouss, J. Rousset, H. Chehade, D. Barataud,and J. P. Teyssier, Modeling microwave devicesand circuits for telecommunications system de-sign, Proc IEEE Int Conf Neural Networks, An-chorage, AK, May 1998, pp. 128]133.

15. A. Veluswami, M. S. Nakhla, and Q. J. Zhang,The application of neural networks to EM basedsimulation and optimization of interconnects inhigh speed VLSI circuits, IEEE Trans Microwave

Ž .Theory Tech 45 1997 , 712]723.16. Q. J. Zhang, F. Wang, and M. S. Nakhla, Opti-

mization of high-speed VLSI interconnects: A re-view, Int J Microwave Millimeter-Wave CAD 7Ž .1997 , 83]107.

17. T. Horng, C. Wang, and N. G. Alexopoulos, Mi-crostrip circuit design using neural networks,IEEE Int Microwave Symp Digest, Altanta, GA,1993, pp. 413]416.

18. P. Burrascano, M. Dionigi, C. Fancelli, and M.Mongiardo, A neural network model for CADand optimization of microwave filters, IEEE IntMicrowave Symp Digest, Baltimore, MD, 1998,pp. 13]16.

19. M. D. Baker, C. D. Himmel, and G. S. May,In-situ prediction of reactive ion etch endpointusing neural networks, IEEE Trans Comp Packag,

Ž .Manufact Technol A 18 1995 , 478]483.20. M. Vai and S. Prasad, Automatic impedance

matching with a neural network, IEEE MicrowaveŽ .Guided Wave Lett 3 1993 , 353]354.

21. R. Biernacki, J. W. Bandler, J. Song, and Q. J.Zhang, Efficient quadratic approximation for sta-tistical design, IEEE Trans Circuits Syst CAS-36Ž .1989 , 1449]1454.

22. P. Meijer, Fast and smooth highly nonlinear mul-tidimensional table models for device modelling,

Ž .IEEE Trans Circuits Syst 37 1990 , 335]346.23. F. Scarselli and A. C. Tsoi, Universal approxima-

tion using feedforward neural networks: a surveyof some existing methods, and some new results,

Ž .Neural Networks 11 1998 , 15]37.24. B. Widrow and M. A. Lehr, 30 years of adaptive

neural networks: Perceptron, madaline, and back-Ž .propagation, IEEE Proc 78 1990 , 1415]1442.

25. M. Minskey and S. Papert, Perceptrons, MITPress, Cambridge, MA, 1969.

26. G. Cybenko, Approximation by superpositions ofa sigmoidal function, Math Control Signals Syst 2Ž .1989 , 303]314.

27. K. Hornik, M. Stinchcombe, and H. White, Multi-layer feedforward networks are universal approxi-

Ž .mators, Neural Networks 2 1989 , 359]366.28. S. Tamura and M. Tateishi, Capabilities of a

four-layered feedforward neural network: fourlayer versus three, IEEE Trans Neural Networks,Ž .8 1997 , 251]255.

29. J. de Villiers and E. Barnard, Backpropagationneural nets with one and two hidden layers, IEEE

Ž .Trans Neural Networks 4 1993 , 136]141.30. F. Girosi, Regularization theory, radial basis func-

tions and networks, From statistics to neural net-works: Theory and pattern recognition applica-tions, V. Cherkassy, J.H. Firedman, and H. Wech-

Ž .sler Editors , Springer-Verlag, BerlinrNew York,1992, pp. 166]187.

31. M. J. D. Powell, Radial basis functions for multi-variate interpolation: a review, Algorithms for ap-

Ž .proximation, J.C. Mason and M.G. Cox Editors ,Oxford Univ. Press, Oxford, U.K., 1987, pp.143]167.

32. J. Park and I. W. Sandberg, Universal approxima-tion using radial-basis-function networks, Neural

Ž .Comput 3 1991 , 246]257.

Page 22: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 237

33. J. Park and I. W. Sandberg, Approximation andradial-basis-function networks, Neural Comput 5Ž .1991 , 305]316.

34. A. Krzyzak, T. Linder, and G. Lugosi, Nonpara-metric estimation and classification using radialbasis function nets and empirical risk minimiza-

Ž .tion, IEEE Trans Neural Networks 7 1996 ,475]487.

35. W. Kaminski and P. Strumillo, Kernel orthonor-malization in radial basis function neural net-

Ž .works, IEEE Trans Neural Networks 8 1997 ,1177]1183.

36. S. Haykin, Neural networks: A comprehensivefoundation, IEEE Press, Piscataway, NJ, 1994.

37. J. Moody and C. J. Darken, Fast learning innetworks of locally-tuned processing units, Neural

Ž .Comput 1 1989 , 281]294.38. T. Y. Kwok and D. Y. Yeung, Constructive algo-

rithms for structure learning in feedforward neu-ral networks for regression problems, IEEE Trans

Ž .Neural Networks 8 1997 , 630]645.39. R. Reed, Pruning algorithms}a survey, IEEE

Ž .Trans Neural Networks 4 1993 , 740]747.40. A. Krzyzak and T. Linder, Radial basis function

networks and complexity regularization in func-tion learning, IEEE Trans Neural Networks 9Ž .1998 , 247]256.

41. H. Leung and S. Haykin, Rational function neuralŽ .network, Neural Comput 5 1993 , 928]938.

42. G. G. Towell and J. W. Shavlik, Knowledge-basedŽ .artificial neural networks, Artif Intell 70 1994 ,

119]165.43. R. Maclin and J. W. Shavlik, Using knowledge-

based neural networks to improve algorithms: re-fining the Chou]Fasman algorithm for protein

Ž .folding, Mach Learn 11 1993 , 195]215.44. L. M. Fu, Integration of neural heuristics into

knowledge-based inference, Connection Sci 1Ž .1989 , 325]340.

45. R. C. Lacher, S. I. Hruska, and D. C. Kuncicky,Back-propagation learning in expert networks,

Ž .IEEE Trans Neural Networks 3 1992 , 67]72.46. J. J. Mahoney and R. J. Mooney, Combining con-

nectionist and symbolic learning to refine cer-Ž .tainty factor rule bases, Connection Sci 5 1993 ,

339]364.47. W. J. Jansen, M. Diepenhorst, J. A. G. Nijhuis,

and L. Spaanenburg, Assembling engineeringknowledge in a modular multi-layer perceptronneural network, Proc IEEE Int Conf Neural Net-works, Houston, TX, June 1997, pp. 232]237.

48. G. G. Lendaris, A. Rest and T. R. Misley, Improv-ing ANN generalization using a priori knowledgeto pre-structure ANNs, Proc IEEE Int Conf Neu-ral Networks, Houston, TX, June 1997, pp.248]253.

49. T. A. Johansen and B. A. Foss, ConstructingNARMAX models using ARMAX models, Int J

Ž .Control 58 1993 , 1125]1153.

50. T. A. Johansen and B. A. Foss, Identification ofnon-linear system structure and parameters using

Ž .regime decomposition, Automatica 31 1995 ,321]326.

51. O. Nelles, S. Sinsel, and R. Isermann, Local basisfunction networks for identification of a tur-bocharger, Proc UKACC Int Conf Control, Ex-eter, U.K., July 1996, pp. 7]12.

52. S. Mcloone, M. D. Brown, G. Irwin, and G. Light-body, A hybrid linearrnonlinear training algo-rithm for feedforward neural networks, IEEE

Ž .Trans Neural Networks 9 1998 , 669]684.53. C. S. Walker, Capacitance, inductance and

crosstalk analysis, Artech House, Boston, MA,1990.

54. P. H. Ladbrooke, MMIC design: GaAs FET’s andHEMTs, Artech House, Boston, MA, 1989.

55. J. R. Tsai, P. C. Chung, and C. I. Chang, Asigmoidal radial basis function neural network forfunction approximation, Proc IEEE Int ConfNeural Networks, Washington, DC, June 1996, pp.496]501.

56. A. Abdelbar and G. Tagliarini, Honest: A newhigh order feedforward neural networks, ProcIEEE Int Conf Neural Networks, Washington,DC, June 1996, pp. 1257]1262.

57. A. J. Sharkey, On combining artificial neural nets,Ž .Connection Sci 8 1996 , 299]314.

58. S. Hashem and B. Schmeiser, Improving modelaccuracy using optimal linear combinations oftrained neural networks, IEEE Trans Neural Net-

Ž .works 6 1995 , 792]794.59. S. Hashem, Algorithms for optimal linear combi-

nations of neural networks, Proc IEEE Int ConfNeural Networks, Houston, TX, June 1997, pp.242]247.

60. S. Hashem, Optimal linear combinations of neuralŽ .networks, Neural Networks 10 1997 , 599]614.

61. A. J. Sharkey, Modularity, combining and artificialŽ .neural nets, Connection Sci 9 1997 , 3]10.

62. R. A. Jacobs, F. Peng, and M. A. Tanner, ABayesian approach to model selection in hierar-chical mixtures-of-experts architectures, Neural

Ž .Networks 10 1997 , 231]241.63. G. Auda, M. Kamel, and H. Raafat, A new neural

network structure with cooperative modules, IEEEInt Conf Neural Networks, Orlando, FL, June1994, pp. 1301]1306.

64. L. C. Wang, N. M. Nasrabadi, and S. Der, Asymp-totical analysis of a modular neural network, ProcIEEE Int Conf Neural Networks, Houston, TX,June 1997, pp. 1019]1022.

65. F. Wang, V. K. Devabhaktuni, and Q. J. Zhang, Ahierarchical neural network approach to the de-velopment of library of neural models for mi-crowave design, IEEE Int Microwave Symp Di-gest, Baltimore, MD, 1998, pp. 1767]1770.

Page 23: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.238

66. Q. H. Zhang, Using wavelet network in nonpara-metric estimation, IEEE Trans Neural Networks 8Ž .1997 , 227]236.

67. Q. H. Zhang and A. Benvensite, Wavelet net-Ž .works, IEEE Trans Neural Networks 3 1992 ,

889]898.68. B. R. Bakshi and G. Stephanopoulos, Wave-net: a

multiresolution, hierarchical neural network withlocalized learning, Amer Inst Chemical Eng J 39Ž .1993 , 57]81.

69. B. Delyon, A. Juditsky, and A. Benveniste, Accu-racy analysis for wavelet approximations, IEEE

Ž .Trans Neural Networks 6 1995 , 332]348.70. S. E. Fahlman and C. Lebiere, The cascade-corre-

lation learning architecture, Advances in neuralŽ .information processing D.S. Truretzky Editors ,

Vol. 2, Morgan Kauffman, San Mateo, CA, 1990,pp. 524]532.

71. N. K. Treadgold and T. D. Gedeon, ExtendingCasPer: a regression survey, Int Conf Neural In-formation Processing, 1997, pp. 310]313.

72. N. K. Treadgold and T. D. Gedeon, Exploringarchitecture variation in constructive cascade net-works, Proc IEEE Int Conf Neural Networks,Anchorage, AK, May 1998, pp. 343]348.

73. J. N. Hwang, S. S. You, S. R. Lay, and I. C. Jou,The cascade-correlation learning: a projectionpursuit learning perspective, IEEE Trans Neural

Ž .Networks 7 1996 , 278]289.74. D. E. Rumelhart, G. E. Hinton, and R. J. Williams,

Learning internal representations by error propa-gation, Parallel distributed processing, Vol. I, D.E.

Ž .Rumelhart and J.L. McClelland Editors , Vol. I,MIT Press, Cambridge, MA, 1986, pp. 318]362.

75. D. G. Luenberger, Linear and nonlinear program-ming, Addison-Wesley, Reading, MA, 1989.

76. K. Ochiai, N. Toda, and S. Usui, Kick-out learningalgorithm to reduce the oscillation of weights,

Ž .Neural Networks 7 1994 , 797]807.77. S. J. Perantonis and D. A. Karras, An efficient

constrained learning algorithm with momentumŽ .acceleration, Neural Networks 8 1995 , 237]249.

78. N. B. Karayiannis and A. N. Venetsanopoulos,Artificial neural networks: learning algorithms,performance evaluation, and applications, KluwerAcademic, Boston, MA, 1993.

79. N. Morgan, Big dumb neural nets: a workingbrute force approach to speech recognition ProcIEEE Int Conf Neural Networks, vol. VII, Or-lando, FL, July 1994, pp. 4462]4465.

80. F. Wang and Q. J. Zhang, A sparse matrix ap-proach to neural network training, Proc IEEE IntConf Neural Networks, Perth, Australia, Nov.1995, pp. 2743]2747.

81. Neural network toolbox: For use with Matlab,MathWorks, Natick, MA, 1993.

82. R. A. Jacobs, Increased rate of convergencethrough learning rate adaptation, Neural Net-

Ž .works 1 1988 , 295]307.

83. A. H. Zaabab, Q. J. Zhang, and M. S. Nakhla,Device and circuit-level modeling using neuralnetworks with faster training based on networksparsity, IEEE Trans Microwave Theory Tech 45Ž .1997 , 1696]1704.

84. F. Wang and Q. J. Zhang, An adaptive and fullysparse training approach for multilayer percep-trons, Proc. IEEE Int Conf Neural Networks,Washington, DC, June 1996, pp. 102]107.

85. R. Battiti, Accelerated backpropagation learning:Ž .two optimization methods, Complex Syst 3 1989 ,

331]342.86. M. Arisawa and J. Watada, Enhanced back-propa-

gation learning and its application to businessevaluation, Proc IEEE Intl Conf Neural Net-works, Vol. I, Orlando, Florida, July 1994, pp.155]160.

87. A. G. Parlos, B. Fernandez, A. F. Atiya, J.Muthusami, and W. K. Tsai, An accelerated learn-ing algorithm for multilayer perceptron networks,

Ž .IEEE Trans Neural Networks 5 1994 , 493]497.88. X. H. Yu, G. A. Chen, and S. X. Cheng, Dynamic

learning rate optimization of the backpropagationŽ .algorithm, IEEE Trans Neural Networks 6 1995 ,

669]677.89. N. Baba, Y. Mogami, M. Kohzaki, Y. Shiraishi,

and Y. Yoshida, A hybrid algorithm for findingthe global minimum of error function of neuralnetworks and its applications, Neural Networks 7Ž .1994 , 1253]1265.

90. P. P. Van-Der-Smagt, Minimisation methods fortraining feedforward neural networks,‘‘ Neural

Ž .Networks 7 1994 , 1]11.91. W. H. Press, B. P. Flannery, S. A. Teukolsky, and

W. T. Vetterling, Numerical recipes: The art ofscientific computing, Cambridge Univ. Press,Cambridge, MA, 1986.

92. D. R. Hush and J. M. Salas, Improving the learn-ing rate of back propagation with the gradientreuse algorithm, Proc IEEE Int Conf Neural Net-works, Vol. I San Diego, CA, 1988, pp. 441]447.

93. D. B. Parker, Optimal algorithms for adaptivenetworks: second order back propagation, secondorder direct propagation and second order heb-bian learning, Proc IEEE First Int Conf NeuralNetworks, Vol. II, San Diego, CA, 1987, pp.593]600.

94. R. L. Watrous, Learning algorithms for connec-tionist networks: applied gradient methods of non-linear optimization, Proc IEEE First Int ConfNeural Networks, Vol. II, San Diego, CA, 1987,pp. 619]627.

95. R. Battiti, First and second-order methods forlearning: between steepest descent and Newton’s

Ž .method, Neural Comput 4 1992 , 141]166.

Page 24: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

NN Structures and Training Algorithms 239

96. J.-S.R. Jang, C.-T. Sun, and E. Mizutani, Deriva-tive-based optimization, Neuro-fuzzy and softcomputing, a computational approach to learningand Machine intelligence, Prentice-Hall, UpperSaddle River, NJ, 1997, pp. 129-172.

97. M. F. Moller, A scaled conjugate gradient algo-rithm for fast supervised learning, Neural Net-

Ž .works 6 1993 , 525]533.98. T. R. Cuthbert Jr., Quasi-Newton methods and

constraints, Optimization using personal comput-ers John Wiley & Sons, New York, 1987, pp.233]314.

99. K. R. Nakano, Partial BFGS update and efficientstep-length calculation for three-layer neural net-

Ž .work, Neural Comput 9 1997 , 123]141.100. S. McLoone and G. W. Irwin, Fast parallel off-line

training of multilayer perceptrons, IEEE TransŽ .Neural Networks 8 1997 , 646-653.

101. A. J. Shepherd, Second-order optimization meth-ods, Second-order methods for neural networks,Springer-Verlag, BerlinrNew York, 1997, pp.43]72.

102. S. Kollias and D. Anastassiou, An adaptive leastsquares algorithm for the efficient training of arti-ficial neural networks, IEEE Trans Circuits Syst

Ž .36 1989 , 1092]1101.103. M. G. Bello, Enhanced training algorithms, and

integrated trainingrarchitecture selection formultilayer perceptron networks, IEEE Trans Neu-

Ž .ral Networks 3 1992 , 864]875.104. G. Zhou and J. Si, Advanced neural-network

training algorithm with reduced complexity basedon Jacobian deficiency, IEEE Trans Neural Net-

Ž .works 9 1998 , 448]453.105. S. Ergezinger and E. Thomsen, An accelerated

learning algorithm for multilayer perceptrons: op-timization layer by layer, IEEE Trans Neural Net-

Ž .works 6 1995 , 31]42.106. G. J. Wang and C. C. Chen, A fast mulitlayer

neural network training algorithm based on thelayer-by-layer optimizing procedures, IEEE Trans

Ž .Neural Networks 7 1996 , 768]775.

107. R. Lengelle and T. Denceux, Training MLPs layerby layer using an objective function for internal

Ž .representations, Neural Networks 9 1996 , 83]97.108. J. S. Shawe-Taylor and D. A. Cohen, Linear pro-

gramming algorithm for neural networks, NeuralŽ .Networks 3 1990 , 575]582.

109. B. Berma, Fast training of multilayer perceptrons,Ž .IEEE Trans Neural Networks 8 1997 , 1314]1320.

110. F. Wang and Q. J. Zhang, An improved k-meansclustering algorithm and application to combinedmulti-codebookrMLP neural network speechrecognition, Proc Canadian Conf Elec Comp Eng,Vol. II, Montreal, Quebec, 1995, pp. 999]1002.

111. S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi,Optimization by simulated annealing, Science 220Ž .1983 , 671]680.

112. J. C. F. Pujol and R. Poli, Evolving neural net-works using a dual representation with a com-bined crossover operator, Proc IEEE Int ConfEvol Comp, Anchorage, AK, May 1998, pp.416]421.

113. T. Rognvaldsson, On Langevin updating in multi-¨Ž .layer perceptron, Neural Comput 6 1994 ,

916]926.114. R. Brunelli, Training neural nets through stochas-

Ž .tic minimization, Neural Networks 7 1994 ,1405]1412.

115. OSA90 Version 3.0, Optimization Systems Associ-ates Inc., P.O. Box 8083, Dundas, ON, CanadaL9H 5E7, now HP EEsof, 1400 FountaingroveParkway, Santa Rosa, CA 95403.

116. A. Djordjevic, R. F. Harrington, T. Sarkar, and M.Bazdar, Matrix parameters for multiconductortransmission lines: Software and user’s manual,Artech House, Boston, MA, 1989.

117. R. F. Harrington, Field computation by momentmethods, Macmillan Co., New York, 1968.

118. NeuroModeler Version 1.0, Professor Q.J. Zhangand his neural network research team, Depart-ment of Electronics, Carleton University, 1125Colonel By Drive, Ottawa, ON, Canada K1S 5B6.

BIOGRAPHIES

Fang Wang received the B.Eng. degreefrom Xi’an Jiaotong University, Xi’an,China in 1993, and the M.Eng. degreefrom Carleton University, Ottawa,Canada in 1995, both in electrical engi-neering and is currently completing thePh.D. degree in the Department of Elec-tronics at Carleton University, Ottawa,Canada. She is a recipient of a 1997]1999

Postgraduate Scholarship from the Natural Science and Engi-Ž .neering Research Council of Canada NSERC . She is also a

two-time recipient of the Ontario Graduate Scholarship in1996 and 1997, respectively. She received a travel fellowshiptwice from the International Conference on Neural Networks in1997 and 1998, respectively. Her research interests includeneural networks, modeling, and their applications in CAD forelectronics circuits. She is a member of the IEEE.

Page 25: Neural Network Structures and Training Algorithms …qjz/ANN_Course/Fang_RFMCAE_NN...Neural Network Structures and Training Algorithms for RF and Microwave Applications Fang Wang,

Wang et al.240

Vijaya Kumar Devabhaktuni received theB.Eng. degree in electrical and electron-ics and the M.Sc. degree in physics, bothfrom Birla Institute of Technology andScience, Pilani, India in 1996. He is cur-rently a Ph.D. candidate in the Depart-ment of Electronics, Carleton Univesity,Ottawa, Canada. His research interestsinclude neural networks, microwave de-

vice and circuit modeling, computer-aided design of VLSI, andtelecommunication circuits. He is a student member of theIEEE.

Changgeng Xi received the B.Eng. de-gree in Information Engineering and theM.Sc. degree in Pattern Recognition andIntelligent Control, from Xi’an JiaotongUniversity, Xi’an, China in 1994 and 1997,respectively. He is currently a Master’sstudent in the Department of Electronicsat Carleton University, Ottawa, Canada.His research intrests include training al-

gorithms of neural networks, microwave modeling, and CADfor VLSI design.

Qi-Jun Zhang. For bio and photo see p. 157.