Anda di halaman 1dari 14

Artificial Neural Network application in Parameter Optimization of Rectangular Microstrip

Patch Antenna
1 2
R.Malmathanraj , S.Thamarai Selvi
1
Lecturer /ECE National Institute of Technology, Tiruchirapalli malmathan@gmail.com
2
Professor and Head /Information Technology, Madras Institute of Technology, Anna University, Chennai.
stselvi@annauniv.edu

Abstract - Printed microstrip antennas and simple rectangular patch fed at the center of
arrays are known to have limitations in terms radiating wall. A microstrip patch antenna is a
of bandwidth and efficiency, all imposed by the radiating patch on one side of a dielectric
very presence of the dielectric substrate. The substrate, which has a ground plane on the
paper deals with the design of a probe fed and underside. The EM waves fringe off the top
edge fed rectangular microstrip patch antenna patch into the substrate, reflecting off the
with the basic parameters W,h,L,ε r,fo to ground plane and radiates out into the air.
achieve better bandwidth and directivity with Radiation occurs mostly due to the fringing
efficient radiation pattern and Gain. The field between the patch and ground. The
analytical results for various possible radiation efficiency of the patch antenna
dimensions and different dielectric values were depends largely on the substrate permittivity
calculated for achieving bandwidth and (εr) of the dielectric[2]. The basic geometry of
directivity without any structural complexities the microstrip patch is shown in fig (1)
.The analytical results were tested by
simulating with basic design software
PCAAD,MSTRIP40. To obtain an optimum
value for the design parameters of the
microstrip antenna Support Vector Machines
(SVM), Generalised Regularisation Neural
Network (GRNN) and Back Propagation
Network (BPN) were implemented to train the
network to attain optimized values to yield
wide bandwidth and better directivity with
high Gain. The application of artificial neural
network ensures an optimum design
methodology for microstrip antenna design
which is revealed when comparing the results
with analytical methods and the results of the Figure 1.Microstrip Patch Antenna Geometry
simulation softwares.

1. Introduction
Microstrip patch antennas have been
attractive due to their conformal properties.
Mathematical modeling of the basic microstrip
radiator was initially carried out by the Ideally, a thick dielectric is preferred
application of transmission-line analogies to for broadband purposes. Small values of width
W of patch result in low antenna efficiencies
while large W values lead to higher order various complex structures adopted for the
modes. Substrate thickness should be chosen as enhancement of bandwidth, Directivity and
large as possible to maximize bandwidth and Gain. The size of the probe is selected as 0.2
efficiency, but not so large as to risk surface- mm and the various feeding positions were
wave excitation. The patch length is considered for the calculation[5]. The
determined by condition for resonance. This dimensions of the patch antenna along with the
occurs when the input impedance is purely substrate permittivity and the probe position is
real. The bandwidth of the patch is defined as varied for different operating frequencies and
the frequency range over which it is matched the numerical results were arrived using the
with that feed line within specified limits. In basic design formulas of the microstrip patch
other words, the frequency range over which listed below,
the antenna will perform satisfactorily. This
means the channels have larger usable The width of the Microstrip patch antenna is
frequency range and thus results in increased given by
transmission. The bandwidth of an antenna is
c
usually defined by the acceptable standing W= -(1)
wave ratio (SWR) value over the concerned 
r 1
2 f0
frequency range[3,4]. Dimensions of the top 2
patch were calculated to get the required Effective dielectric constant is given by
bandwidth and the impedance matching.
The advantages of microstrip antenna 
1
1 r 1  h 2
is that they are low-cost, conformable, reff = r  1 12  -(2)

lightweight and low profile, while both linear 2 2  W
and circular polarization easily achieved.
Disadvantages of microstrip antenna include Effective length ( Leff ) is given by
such as a narrow bandwidth, a low gain (~6
C
dB) and polarization purity is hard to achieve. Leff  -(3)
Several methods were reported in literature to 2 f 0 reff
improve impedance bandwidth including
employing wide band impedance matching, Length extension ( L ) is given by
stacked patches and utilizing thicker
substrates[9].  0.3
reff
W
 0.264

L 0.412h h  -(4)

2.Design Methodology reff 0.258W 


 0.8 
h 
2.1 Basic Rectangular Microstrip
The major design task of this paper is
optimization of the dimensions of the probe
fed rectangular microstrip patch antenna. The Actual length of patch ( L ) is given by
simplest approach is adopted to demonstrate
how effectively the Artificial neural network L= Leff 2L -(5)
can be used to train and optimize the various
parameters involved in the design of microstrip Ground plane dimensions ( Lg and W g) is given
patch antenna. This work concentrates only on by
the basic geometry of the microstrip ignoring Lg = 6h +L -(6)
Wg = 6h + W -(7) 2.2 Artificial Neural Network Design
for Rectangular Microstrip Patch Antenna
Height of the substrate is given by .
A microstrip rectangular patch
0.3C
h antenna (Fig. 1) can be viewed as a matrix
2fo r
-(8) with X variables and four unknown the bs,
such that
The bandwidth of a rectangular patch is given
by AX=b
BW 3.77  
r 1
/ 2r  
W / L
h / 0  --(9)
The unknown are the resonant frequency (RF),
bandwidth (BW), gain(G), and polarization
(PL). The variables X, are the patch length
Where f 0 is the resonant frequency,  r is the (L1), patch width(L2), substrate height (H 1) ,
relative substrate permittivity, C is the speed of substrate relative permittivity (r) and the
light 3 108 m/s. feeding positions(Xf) and (Yf)

[A][ L1 , L2 , r, H1 , X,Y, Xf Yf]T=[G,BW,S 11,PL]T


processing structure of rectangular microstrip
antennas, where the input neuron units are (L
G=f(L1 , L2 , r, H1 , X,Y, Xf , Yf); ,W, εr ,h ,P ,fo )and the output units are (BW ,D
,G ,RP ). The learning paradigm on the
BW=f(L1 , L2 , r, H1 , X,Y, Xf , Yf); microstrip is supervised learning, where the
mapping function between the inputs and
S11=f(L1 , L2 , r, H1 , X,Y, Xf , Yf); outputs is the matrix A. The inputs are
weighted and the effect that each input has at
PL=f(L1 , L2 , 
r, H 1 , X,Y, Xf , Y f );
decision making is dependent on the weight of
the particular input. The weight of an input is a
In the neural network design the inputs are L1 , number that when multiplied with the input
L2 , r, H1 , X,Y, Xf , Yf and the output is gives the weighted input. Their calculation is
G,BW,S 11,PL. based on the method of moments. These
An artificial neural network is an weighted inputs then generate the unknowns.
information processing paradigm. That is Those unknowns are then compared to stored
inspired by the way biological nervous information that gives the desired bandwidth, -
systems, such as the brain, process directivity along with radiation pattern and
information. The key element of this paradigm gain. The gain is expected to be greater than
is the novel structure of the information 3 db and polarization should be linear or
processing system .It is composed of a large circular [1,2].
number of highly interconnected processing A good paradigm of supervised
elements (neurons), working in union to solve learning that is of interest to microstrip
specific problems. Artificial neural network is antenna designer is error correcting learning
like a normal human, it learns by that is minimization of error between the
examples.[2,6] desired and computed values. In this learning
A neural network with feedback is an paradigm, the set of weights that minimizes the
adequate representation of the information error between the teaching input and the
weighted inputs is obtained. Neural networks mathematically. The structure of the human
are general function approximators. One brain and the learning process is known, but
important characteristic is that they can learn the main difference between both networks is
any Input-Output (IO) mapping by using the the effciency. The human brain is capable of
information contained in a given input-output recognizing a familiar face in approximately
data set without needing a structure definition 100-200 ms, where conventional computers
of the IO-mapping. The type of network, can take hours or days to fulfill less complex
parameter settings, number of hidden neurons, tasks. The biological neural network is still
and the connectivity of the neural network much faster then the artificial neural network
define the structure of the approximated IO- (ANN or NN), but the capabilities of the
mapping. The only additional information a neural networks are promising. In this section,
neural network needs, besides the input-output the general structure of a neural network will
data, is the definition of the input and output be explained. First the concept of neurons is
parameters, the relevant parameters which treated with a special attention to the several
span the IO-mapping. The ideal use of neural activation functions. The possible
networks is antenna model parameters combinations of several neurons in layers is
optimization. Without knowing the IO- handled in the discussion about the
mapping structure of the model, the neural architecture of networks. Some example
network can learn to mimic the IO-mapping networks are used to illustrate the effect of the
[7,8]. parameters of the networks. The primary
In the IO-mapping problem a feed- element in a neural network is the neuron, an
forward neural network is used because the information-processing unit. A mathematical
antenna model is a static mapping (except at model of an artificial neuron in given in Fig.
the very moment the design failure occurs). (2). The structure is similar to the human
Also the ’tangent sigmoidal’, e.g. ’tansig’, neuron; more information about the
activation function will be used in the hidden functioning of the human neural network. The
layer of the network because of two reasons. elements xi on the left side are the input signals
One, in the antenna design, IO-mapping is a to the neuron k. These inputs are multiplied by
smooth mapping with little to none the corresponding weights wki and summed
discontinuities. The ’tansig’ is also a smooth together with the bias bk. The results of the
function with the capability of approximating a summation vk is passed through an activation
discontinuity (by squashing the shape with function (vk) producing the output yk.
respect to the input axis). The second reason The mathematics of the neuron given in
for choosing the ’tansig’ function is that this Fig. (2) can start with the weighted inputs
function is a very ’general’ function. After a p
failure, the antenna design IO-mapping has v k   wki x i -(10)
changed into an unknown form. Therefore an i 1
activation function, which can be used to The output can be written as
mimic almost all shapes, is more preferable yk k (vk bk ) -(11)
since in that case all (unknown) IO-mapping
A weighted bias can be included by adding an
can be approximated by the neural network. extra term in the first equation.
Neural networks are based on the
human brain and its enormous capability of p
learning and adapting. Over decades, people v   w xi -(12)
have been trying to model the human brain k ki
i 0
Where the bias is changed to a fixed input of 1
and with a weight of wk0 . The shape of the
output is then only depending on the activation
function of vk.
y k k (v k ) -(13)

The type of activation function has a large


influence on the output of the neuron as can be
seen from the equation (12). In a signal flow
diagram, a neuron can be represented as shown
in Fig. (3).

x0=1

x1 wk,0

wk,1

x2 wk,2 vk (k) yk

.
.
.
xn wk,n

Figure 2. Mathematical
definition of a neural network

X1
wk, bk
1

wk,
X2 2 Σ
.
.
.
Fig 3. Signal flow diagram of a
neuron

A neural network is a directed graph 3.1 Learning


consisting of nodes with interconnecting
synaptic and activation links, and is As the neural network software read
characterized by four properties: the training set, the network learns the data
1. Each neuron is represented by a set of linear patterns in the training set. Learning
synaptic links, an externally applied bias, and a subprograms differ depending on the
possibly nonlinear activation link. The bias is architecture selected. As training progressed,
represented by a synaptic link connected to an statistical graphs furnished by the neural net
input fixed at +1. software, provided a means to monitor training
2. The synaptic links of a neuron weight their progress.
respective input signals. Numerical historical data and repetitive
3. The weighted sum of the input signals examples in which the solution is already
defines the induced local field of the neuron in known are required to train a neural network.
question. While the relationship between variables may
4. The activation link squashes the induced not be known, network results can be
local field of the neuron to produce an output. improved by the addition of more variables.
Data may need different representation, for
3. Architecture Selection example if data has a very large value range,
A variety of Neural network logarithms or other data transformations or
architectures are available to process the data conversions may be necessary.
from the input data set files. A multi layer 3.2 Generalised Regularisation
Backpropagation Network architecture, Neural Network (GRNN)
Generalized Regularization neural network
(GRNN) and Support vector machines (SVM) The GRNN is based on the Nadaraya –
were used for training because of its ability to Watson Kernel regression. GRNN’s feature
generalize well when applied to a wide variety fast training times can model non linear
of applications and also for the ability to have functions and have been shown to perform
better regression. well in noisy environments given enough data.
The primary advantage of the GRNN is the
speed at which the network can be trained.
Training a GRNN is performed in one pass of
the training data through the network, the
training data values are copied to become the
weight vectors between layers. The
architecture of the GRNN is shown in the
figure 4, it has four layers input pattern,
summation and output, with weighted
connections Wij between the input and pattern
layer and

Pattern Units Summation Layer Output Units

Input
S
S
fo D

S
L

W S
BW

ε
r
D

P S- Summation Unit
D- Division Unit

fo

Figure 4.Architecture of GRNN Network

Input
sets or between H1 and H2 we need to
minimize ||w|| with the condition that there are
no data points between H1 and H2
3.3 Support Vector Machines
Traditionally neural networks have w . x –b ≥+1 for yi = +1 -(17)
been used for classification, which is based on w . x –b ≤-1 for yi = -1 -(18)
Empirical Risk Minimization (ERM). SVM
was developed by Vapnik and had become Combining the above two equations,
popular tools for data mining. The formulation
embodies the Structural Risk Minimization yi ( w . x – b ) >= 1 -(19)
(SRM), which is superior to empirical risk
minimization. SRM minimizes the upper So the problem of maximizing the
bound on expected risk as supposed to ERM distance between hyper plane H1 and H2 is
that minimizes the error on training data. So, formulated as min ½ wT w subject to
SVM generalizes much better. There are many
linear classifiers that can separate data, but yi ( w . x – b ) ≥1 -(20)
SVM only maximizes the margin i.e. the
distance between it and the nearest data point This is a convex quadratic problem in w, b in a
in each class. convex set. The solution is found by solving
We have N training data {(x 1,y1), using lagrangian method by introducing
(x2,y2),….. (xN,yN)} Where xi Є Rd and yi Є lagrangian multipliers. It is easier to solve
{+1,-1}. It needs to be classified using a linear using lagrangian dual equation given by
hyper plane classifier
LD = Σi αi - Σi Σj αi αj yi yj xi · xj -(21)
f(x) =sgn (w.x - b) -(14)
The significance of the above equation is that
This hyper plane will have maximum distance the training input vectors appear only as dot
between each class. This hyper plane product. So when the data is not linearly
H : y =w . x – b = 0 and two hyper planes separable it is required to transform the data
parallel to it into a higher dimensional. This causes
complex calculations in neural networks but in
H1 : y =w . x – b= +1 -(15) SVM as data appear only as a dot product all
calculation can be carried explicitly in low
H2 : y =w . x – b = -1 -(16) dimension if a kernel function exists for

With no data points between H1 and H2, and LD = Σi αi - Σi Σj αi αj yi yj Φ(xi) · Φ(xj) -(22)
distance between H1 and H2 maximized.
Some training point will lie on the hyper plane as Φ(xi ) · Φ(xj ) = K(xi , xj )
H1 and H2, they are called support vector Where K is the kernel function. This is
machines because they define the separating equivalent as the dot product in high
plane and the other training points can be dimension is equal to kernel function in input
removed or moved provided they don’t cross space. The common kernel function used is
the planes H1 and H2. The distance between Gaussian kernel,
hyper plane H1 and H2 is 2/ || w||. To
maximize the distance between the two data K (xi , xj) = e - || xi – xj || 2 / σ2 -(23)
(GRNN) and Support vector machines (SVM)
Mercers condition determines whether whose results were in good agreement with the
a function g(x) can be used as a kernel or not, analytical as well as the designed structure
∫g(x)2 dx should be finite. output shown in Table (1). The input output
relations were also checked for the
4.Design Implementation and Result. experimental results. The Backpropagation
The dimensions of the rectangular Network architecture achieves the antenna
patch were selected in a trial and error basis parameter optimization with maximum time
considering the constraints of the design in for convergence. The GRNN and the SVM
selecting the values. The different geometrical neural network achieves optimization with
parameters were designed analytically and the quicker learning time as shown in Fig 11 and
bandwidth given in equation (9) was used to 12. In this research analysis for antenna
calculate the value for the selected dimensions. parameter optimization the GRNN neural
The parameters were used to construct the network produced the accurate result with
structure using the simulation software. The comparatively minimum time for convergence.
bandwidth and directivity along with the gain The computational time was very less in terms
and radiation pattern of the design were of seconds with high accuracy as shown in Fig
obtained. The parameters of the patch 13. The optimized parameters obtained using
equations (1-9) with the feed position for a the training neural networks achieved high
resonant frequency are fed as input to the impedance bandwidth of 7.8%, directivity
networks. The impedance bandwidth and 7.73db without side lobes and offered high
directivity was taken as the output of the gain 8.67 dbi and radiation efficiency 100%
network. The analytical data values are given was attained. The results were comparatively
as input to train the network to obtain an better when compared with the results from
optimized geometry for the probe fed analytical analysis and simulation analysis for
microstrip antenna .The wide range of Microstrip patch antenna using PCAAD and
parameters was used to provide the optimum MSTRIP40.
result and the training steps were increased to
obtain the accuracy.
To train the SVM parameters
The validity of the network was tested
by comparing the analytical results obtained [alpha,b] = trainlssvm ({X, Y, type, gam, sig2,
from the basic formulas for a given set of input
kernel, preprocess})
values. The same parameters were used to
construct a probe fed rectangular patch using
simulation software shown in figure (7) and
Outputs
the output radiation pattern was obtained as
shown in figure (8). The current pattern of the alpha matrix with support values of SVM
designed antenna is also plotted which shows
b vector with bias term(s) of SVM
the even distribution due to proper impedance
matching of the probe feed. The same values
were trained using the three networks
Inputs
Backpropagation Network architecture (BPN),
Generalized Regularization neural network
Fig 5. Antenna output model using MATLAB
Software .

Fig 6 Antenna output model using MATLAB


Software .
Model Trained object oriented representation Gam Regularization parameter
of the SVM model sig2 Kernel parameter (bandwidth in the
Model Object oriented representation of the case of the 'RBF_kernel')
SVM model kernel Kernel type default 'RBF_kernel'
X matrix with the inputs of the Xt inputs of the test data
training data preprocess preprocess
Y vector with the outputs of the
training data Plotting the graph in SVM
Type function estimation plotlssvm({X,Y,type,gam,sig2,'RBF_kernel','p
Gam Regularization parameter reprocess'},{alpha,b});
sig2 Kernel parameter (bandwidth in the
case of the 'RBF_kernel') Inputs
X matrix with the inputs of the training
kernel Kernel type (by default 'RBF_kernel') data
preprocess preprocess'(*) or 'original' Y vector with the outputs of the training
data
Simulating the SVM Type function estimation
Yt=simlssvm({X,Y,type,gam,sig2,'RBF_kerne Gam Regularization parameter
l','preprocess'},Xt); sig2 Kernel parameter (bandwidth in
the case of the 'RBF_kernel')
Outputs kernel Kernel type (by default
Yt matrix with predicted output of test 'RBF_kernel')
data preprocess preprocess
alpha support values obtained from
Inputs training
X matrix with the inputs of the training b Bias term obtained from
data training
Y vector with the outputs of the training 5.Conclusion
The radiation pattern of the designed
data
antenna presented in this paper figure (7)
Type function estimation clearly depicted that it is a wideband antenna
with high directivity, gain with radiation
efficiency. The major attraction of this antenna for mobile communication. The parameter
is size reduction along with bandwidth and optimization using the networks is the major
directivity made it most suitable for satellite attraction of this paper, which highlights the
communication, commercial applications. The simplicity, accuracy and reduction in
size reduction and operating frequency make it computational time for the designers of
suitable interest.

Figure 7.Structure of Probe Fed Microstrip Rectangular


Patch Antenna
Figure 8.Current Distribution
Pattern

Figure 9. Radiation Pattern of the


Optimized Patch Antenna
Figure
10. Plot Showing the Learning Trial of
Back Propagation Network

Fig 1 1. Plot to show the time for convergence


for SVM neural network.
Fig 13. Plot to show the weight surface of SVM

Fig 12. Plot to show the time for convergence for


GRNN neural network.

Figure 15. Output of the Optimized output of the Fig 14. Plot to show the radiation pattern
rectangular patch antenna using MSTRIP using PCAAD
References

1.Dipak K.Neog, Shyam S.Pattnaik, C.Panda, Swapna Devi, Bonomali Khuntia, and Malaya Dutta, “Design of a Wideband Microstrip
Antenna and the use of Artificial Neural Networks in Parameter Calculation”, IEEE Antennas and Propagation Magazine, Vol.47,
No.3, June 2005,pp.60-65.

2. Inder J.Bahl, Prakash Bhartla and Stanislaw S. Stuchly, “ Design of Microstrip Antennas Covered with a Dielectric Layer”, IEEE
Transactions on Antennas and Propagation, Vol. AP-30, No. 2, MARCH 1982, pp. 314-318.

3. Kin-Lu Wong and Yi-Fang Lin, “Small broadband rectangular microstrip antenna with chip-resistor loading”, ELECTRONICS
LETTERS, 1 l September 1997,Vol. 33 No. 79,pp.1593, 1594.

4. S.Lebbar, Z.Guennoun, M.Drissi, and F.Riouch, “A Compact and Broadband Antenna Design Using a Geometrical- Methodology-
Based Artificial Neural Network”, IEEE Antennas and Propagation Magazine, Vol.48, No.2, April 2006,pp.146-154.

5. C. L. Mak, K. M. Luk, Senior Member, IEEE, K. F. Lee, Fellow, IEEE, and Y. L. Chow, “Experimental Study of a Microstrip Patch
Antenna with an L-Shaped Probe,” IEEE Transactions on Antennas and Propagation, VOL. 48, NO. 5, MAY 2000,pp.777-783.

6. R.K.Mishra and Patnaik, “Designing Rectangular Patch Antenna Using the Neurospectral Method”, IEEE Transactions on Antennas
and Propagation,AP-51,8 August 2003,pp.1914-1921.

7. S.S.Pattnaik, D.C.Panda and S.Devi, “Input Impedance of Rectangular Microstrip Patch Antenna Using Artificial Neural Networks”,
Microwave and Optical Technology Letters,32,5,5 March 2002,pp.381-383.

8. S.S.Pattnaik, D.C.Panda and S.Devi, “Radiation Resistance of Coax-Fed Rectangular Microstrip Patch Antenna Using Artificial
Neural Networks”, Microwave and Optical Technology Letters, 34,1,5 July 2002,pp.51-53.

9. D.M.Pozar, “Microstrip Patch Antennas,” in L.C.Godara (ed), Handbook of Antennas in Wireless Communications, New York, CRC
Press, 2001,Chapter 6.

10.Ye Bin Hu Gu Yu , “The analyze and improve TCP performance using a DSR route protocol based on signal strength”, IEEE
Wireless Communications, Networking and Mobile Computing, pp. 846 – 849, 2005.

11. Dongkyun Kim , Hanseok Bae, Jeomki Song, “Analysis of the interaction between TCP variants and routing protocols in
MANETs”, IEEE Parallel Processing, ICPP 2005 Workshops, pp 380-386, 2005.

12. Prabakaran, M. Mahasenan, A. , “Analysis and enhancement of TCP performance over an IEEE 802.11 multi-hop wireless network:
single session case”, IEEE International Conference on Personal Wireless Communications, pp-29- 33, 2005

13. Caihong Kai Yuzhong Chen Nenghai Yu, “An Improvement Scheme Applied to TCP Protocol in Mobile Ad Hoc Networks”, IEEE
International Conference on Mobile Technology, Applications and Systems, pp.1-6, 2005

Anda mungkin juga menyukai