0% found this document useful (0 votes)
7 views

1997 (6)

This paper discusses the application of neural networks for analyzing microstrip antennas, specifically focusing on the optimization of integral equations related to their electromagnetic properties. The authors highlight the advantages of using neural networks, such as stability and efficiency, in overcoming the limitations of traditional integral equation methods. The proposed approach aims to improve the computational efficiency and accuracy of microstrip antenna analysis by leveraging neural network techniques.

Uploaded by

debanjali
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views

1997 (6)

This paper discusses the application of neural networks for analyzing microstrip antennas, specifically focusing on the optimization of integral equations related to their electromagnetic properties. The authors highlight the advantages of using neural networks, such as stability and efficiency, in overcoming the limitations of traditional integral equation methods. The proposed approach aims to improve the computational efficiency and accuracy of microstrip antenna analysis by leveraging neural network techniques.

Uploaded by

debanjali
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

1414 IEEE TRANSACTIONS ON MAGNETICS, VOL. 33, NO.

2, MARCH 1997

Analysis of Microstrip Antennas Using Neural Networks


Lucio Vegni and Alessandro Toscano
Department of Electronic Engineering, Universita di Roma Tre
Via della Vasca Nayale 84, 00146 Rome, Italy

Abstract -
This paper presents the application of tools which can be achieved employing neural
non-linear neural optimization networks to the Neural networks have been found recently [ 131-[
analysis of arbitrarily shaped microstrip Patch very useful tool in the solution of different classes of
antennas with a very general bianisotropic mathemtical optimization problems with general, even non
grounded slab. A neural network approach is
presented to solve in a new and efficient way the
linear, constraints. The developed neural networks have been
integral equation which describes, from the proved to be very fast and robust in application to the
electromagnetic point of view, the planar solution of many technical problems that can be expressed as
integrated structure. the optimization task.
The primary advantage, infact, of using neural networks is
I INTRODUCTION that it is potentially stable. This is due to the parallel
distributed processing structure of neural networks as well as
Microstrip antennas have been extensively investigated the high degree of interconnectivity and feedback. Neural
experimentally, analytically, and numerically for decades. networks also offer the characteristics of generalization and
Many numerical methods have been serving the engineers and error tolerance.
researchers in the analysis and design of these conformal
antennas for many years. Among them the moment method 11. FORMULATION OF THE PROBLE
in conjunction with various Integral Equation (IE)
formulations played a major role [1]-[3]. E formulations for Consider the three-dimensional structure illustrated in Fig. 1
where a microstrip Patch antenna Or array is residing On Or I
the analysis of microstrip structures are the most elaborate
techniques available and provide "almost rigorous" solutions. embedded in a substrate recessed in a ground plane. We are
They are, also, the most computer intensive, requiring interested in the CharaCteriZatiOn Of the Scattering and radiation
considerable computation time and computer memory. They PrOPe*eS ofthis antenna fed by an electric Point-sowce.
can be used as "standard references", providing data against
which results provided by approximations may be validated.
Nevertheless, with the increasing availability of computing
power, they have become quite popular for analyzing
microstrip antennas and microwave, millimetre wave circuits.
However, IE methods are associated with field
representations in which the appropriate Green's function for
the specific geometry must be employed, and this, of course,
limits their versatility. Moreover, IE techniques are usually
formulated on the assumption of an infinite substrate, a
model that obviously deviates from the practical
y=d
configuration, leading to inaccuracies for larger bandwidth Grousd PLane
antennas. Furthermore, in the context of IE methods, antenna
excitations are represented using simplified models that differ Fig.] Geometry of a mcrostrip patch antenna
more or less from the actual configurations. Also, we have to
mention the additional IE complexities due to possible An IE description for a microstrip circuit or an antenna is
substrate anisotropies or inhomogeneities in the antenna obtained by specifying the boundary conditions that must be
substrates. To overcome all these limitations, in this paper satisfied by the electromagneticfield. The conditions apply to
we present a new approach to the solution of IE which makes the total fields, formed by an excitation (the impressed field)
use of neural networks. and a scattered part (the induced field). The scattered field is
A crucial point of the method is the accurate determination produced by a set of unknown induced sources (currents and ,
of dyadic Green's functions associated with the integrated charges on the conductors). Within any linear system, the
structure, sustained by arbitrarily oriented electric point- scattered field can be expressed as a superposition integral of
'
source embedded in the grounded substrate. the unknown sources, weighted by Green's functions G(rlr'),
Expressions for the Green functions for a biisotropic chiral that are the fields produced by elementary point-sources:
medium are set up in a systematic way in [4], well suited,
besides, for the extension to very general bianisotropic
grounded slabs [5]-[ 121. E(r) = G(rlr'>J(r')dv' (1)
The complexity of the electromagnetic problem asks for iv7
sophisticated design optimization and sensitivity analysis
This expression represents the electric field at the point r,
Manuscript received March 18, 1996.
the integration being carried out over the variable r'. The

0018-9464/97$10.00 @ 1997 IEEE

Authorized licensed use limited to: Meerut Institute of Engineering and Technology. Downloaded on February 23, 2009 at 01:14 from IEEE Xplore. Restrictions apply.
1415

integral is evaluated over the volume V' containing the S(rlr') satisfies the differential equation:
sources J(r'). Introducing the integral formulation for the
scattered field into the boundary conditions yields a set of L(E) = P E = jwJ (5)
coupled integral equations for the unknown induced sources,
One or several boundary conditions can be included in the with:
Green's function, thus reducing the number of integral
equations at the cost of greater complexity of the Green's
functions. When a modified Green's function satisfies some
boundary conditions, any field obtained as a superposition
integral with this Green's function satisfies the same
boundary conditions.
The surface currents flowing on the conductors and the
corresponding surface charge distribution are not yet known
and must be determined by imposing the boundary conditions
for the tangential electric field on the surface S' of the upper
conductors:
n
+ [--+1--+ a2
pXxa z2 pZza x2
1 a2 5 r
w2Ey,-w2=lyy
PYY
- +
yx[Ei(r) + E,WI = Z, & J ~ W (2)
where Ei is the incident electric field, J, is the surface current
density on S'induced by Ei, E, is the scattered field sustained
by Js and Z, is the surface impedance of the metallic patch
on S',taking into account the finite conductivity of metallic
sheets:

Z, = (1 + j) (3)
20
w is the operating angular frequency, p is the metal For constructing the Green's function we take the Fourier
permeability and CJ is the effective conductivity of the metal, transform of (6). Thus:
including the possible effect of roughness. The surface
impedance Z,vanishes on a perfect electric conductor.
The electromagneticquantities can be expressed in the two- (7)
dimensional Fourier domain, and subsequent developments
can be carried out in the transformed domain. The analysis
and the solution of the electromagnetic field problem being R = r - r',
becomes simpler if the Fourier transformed domain- or n n

spectral domain- approach is used. This is mainly due to the k = kxx + kyY + kz;,
fact that Green's function convolution integrals in the spectral
domain are turned into algebraic products. In this manner,
also, one avoids the (previously) difficult integration of badly
behaved functions required to obtain the Green's function in
Substitution of (7) in (1) gives an integral representation
the spatial domain. This method has been generalized by the
for the spatial electric field sustained by the integrated planar
authors to study the electromagnetic behavior of microstrip
antennas with planar symmetry realized on a very general structure.
bianisotropic grounded slab [4]-[121. The spatial electric Green's function for integrated
From the electromagnetic point of view, a very general structures with planar symmetry can be recovered in terms of
bianisotropic medium can be characterized by the following the Sommerfeld integrals of the form:
constitutive relations:

D =p E + b H
(4)
B = P E + &H

E, &, k, are the constitutive tensors which can be


where Gi, denotes the spatial domain Green's function and n -
may be 0 or 1 depending on which component of G is to be
represented by four 3x3 matrices. If the medium is lossless evaluated. The spectral domain Green's function has been
the constitutive tensors are Hermitian.
Starting from the Maxwell's equations and the constitutive derived in general form as the sum of TE and TM (to);
relations (4), it can be shown that the dyadic Green's function eigenmodes of propagation no matter where the observation
point is [lo].

Authorized licensed use limited to: Meerut Institute of Engineering and Technology. Downloaded on February 23, 2009 at 01:14 from IEEE Xplore. Restrictions apply.
1416

The spatial electric field excited by point-sources is directly


related with the evaluation of these type of integrals.
A new and computationally efficient solution in terms of
5
k= 1
In@($k),wm)=@,wm) m=l,2,...n (12)
the so-called radial representation of the Sommerfeld problem
involving electric point-sources above a planar interface of In the special case where the expansion functions are
two homogeneous and linear isotropic, biisotropic, identical to the weighting functions, the result is Galerkin's
anisotropic and bianisotropicmedia can be found in [41-[121. method:
As Somerfeld-type integrals being numerically evaluated,
the convergence of the integrand plays an mportant role.
When the source point and observation points are located at
k= 1
different layers (y f y,), the integrand decays exponentially.
Therefore, the integrals can be evaluated without difficulty. If nothing is known about the mathematical properties of
However, for the case with y = y s as wcll as small the operator L other than its linearity, nothing in general can
observabon distance, the integrand converges slowly. be said concerning the convergence of the approximants Jn to
SO,to obtain the inverse Fourier transform of the spectral the solution J . Unfortunately, most of the interesting and
Green's functions, that is to solve integrals (S), two different practical problems involving, for instance, lossy dielectric
techniques have been developed for the two cases of source slabs and metallic sheets with finite conductivity are described
and observation points vertically or horizontally separated 141, by operators that are neither positive nor semi positive-
r111, WI. definite. Therefore, most of the large body of solutions to
It has to be noted that the IE formulation, we proposed, is electromagnetic problems by the moment method lacks any
not developed under the assumption that the ground plane and sort of mathematical convergence criteria.
the substrate have infinite transverse dimensions. By properly The IE of the problem has, then, to be converted into a set
applying the boundary conditions, the size of the antenna can of linear equations by expanding the solution J in terms of an
be finite and the radiation produced by the diffraction on the appropriate and infinite set of basis functions.
edges toward the back of the antenna can be taken into After the termination of the series at the N-th term, the IE
account. It is well evident that this information is required can be approximated by:
when patches are used as freestanding structures and where the
back-to-front ratio must be minimized for interference N
reasons. The diffraction from the edges of the ground plane InL(Jn) = E + n (14)
may, also, cause scallopping of the main beam and slightly n=l
affect the directivity.
An approximate solution of (1) i s given by the following being n the numerical error due to the truncation of ilie series
procedure. at the N-th term.
Flrst consider: The unknown coefficients (11,. IN)^ = I are determined by
solving the NxN matrix equation:

jv, G(rlr')aJ(r')dv' - E(r) = 0 (9) &I.=V+n (15)

or, equivalently: L(J) - E = 0, with LC) =


where &= jv, dv',
G(rmlr')*Jn(r')

v = (E(r1I,... R N ) ) .
Define the linearly independent sets { @k and { wk \g=
where +k and wk are called expansion functions and weighting The matrix equation (15) is an example of a functional
mapping: y=@(x), with x and y N- and M-dimensional real
functions, respectively. Define a sequence of approximants to
vectors, respectively.
J by: The solution of the matrix equation leads to the
determination of the field of the antenna, and, hence, other
Jn(r') = 5
k= 1
Ik@k(r') (10) antenna parameters can be calculated.
111. NEURAL NETWORKS
A matrix equation is formed in (9) by the condition that,
upon replacement of J by Jn, the left side shall be orthogonal In numerical solutions of Maxwell's equations, integral
to the sequence { Wk} . We have: formulations are attractive because they enable researchers to
address problems without discretizing air regions. However,
(L(J,) - E , wm}= o m = I, 2,...n (11) IEformulations inherently lead to dense systems of equations
and, therefore, they are often thought to be computationally
Substitution of (10) into (11) and use of the linearity
property of the inner product gives the matrix equation of the The main difficulties with integral formulations for
moment method: microstrip antennas are the solution time to evaluate the

Authorized licensed use limited to: Meerut Institute of Engineering and Technology. Downloaded on February 23, 2009 at 01:14 from IEEE Xplore. Restrictions apply.
1417

elements of the matrix equation (15), the amount of memory output [15]. The training process for our problem can be
needed to store the system matrix, and the solution time of viewed as having four phases:
the dense system of linear equations. In practice, the storage
requirements may easily exceed the computer's memory, and 1. PRESENTATION PHASE: Present an input training
the solution time may, also, be unacceptable. To make vector (i.e. the vector Jn) and calculate each successive layer's
integral methods more appealing, effort must be devoted to output until the last layer's output is found.
solve large, dense, linear systems efficiently. In addition,
since parallel computing is an obvious way to increase the 2. CHECK PHASE Calculate the network error vector and
computer's memory capacity and computation speed, it is the sum squared error for the input vector:
important to develop efficient linear-equation solvers for both
sequential and parallel computers.
In this section we discuss how to find and implement
efficient solvers by making use of neural networks. Although
these networks were initially intended to perform cognitive
tasks, that have no precise mathematical descriptions, the where: 6@= ypk - Opk. The subscript p refers to the p-th
networks were subsequently found useful in computational exemplar of the network, Opk is the output of the k-th output-
problems and as function approximators. In recent times the layer unit for the p-th exemplar, ypk are the elements of the
term neural network is used for denoting any massively vector V (i.e. the known tangential component of the
parallel computing architectures, that consist of a large incident electric field), and there are M output-layer units.
number of simple "neural" processors. We solved the real- Stop if the sum squared error for all training vectors is less
valued E,obtained by considering the real and imaginary than the error goal or if the specified maximum number of
parts separately, using (Fig.2) the back propagation network iterations has been reached. Otherwise continue.
architecture together with a suitable generalization of the delta
rule [15]. 3. B ACKPROPAGATION PHASE: Calculate the delta
vector for the output and the hidden layers using the target
vector:

the output layer delta: s&= (ypk-4k>@'<ne$k>, (13)


*pi M
..

the hidden layer delta: 6;j = q(ne$,)x 6&'@j.


k=l

We refer to the overall excitation of the k-th unit of the


Fig2 Neural network. Although there iionly one hiddkn layer i n k s output (hidden) layer as its net-input value: net& (ne$&
figure, you can have more than one.
f(net@ is the output function of the same unit.
Backpropagation networks were created by generalizing the
Widrow-Hoff learning rule to multiple layer networks and 4. LEARNING PHASE: Calculate the gradient of the error
nonlinear differentiable transfer functions. In our problem surface with respect to each of the weights on the hidden layer
input vectors and the corresponding output vectors are used to and on the output layer, given, respectively by:
train a network until it can: 1) approximate in the best way a
function, associating an input vector (i.e. the approximants
J,) with a specific output vector (i.e. the vector V); 2)
parallelize all the involved integrals in the matrix equation by
employing a N-point quadrature Gauss algorithm and
splitting the corresponding calculation on N different
processors; 3) accelerate significantly the convergence of the
proposed Galerkin technique, reducing, then the size of the
matrix Z; 4) speedup the solution time of the dense system The reason why we left the derivatives of the output
(15) of linear equations. Trained backpropagation networks functions as "primed" functions instead of explicitly
tend to give reasonable answers when presented with inputs calculating the derivative, is because the value of the
they have never seen. Typically, a new input will lead to an derivative depends on the form of the output function. The
output similar to the correct output for input vectors used in two primary cases of interest are the sigmoid and the identity
training, that are similar to the new input being presented. function. In these two cases, the derivatives of the functions
This generalization property makes it possible to train a for output layer units are @'(net&)= opk(l - opk) for the
network on a representative set of inpudtarget pairs and get
good results for new inputs without training the network on sigmoid, and $(net&) = 1 for the identity function.
all possible input/output pairs. If a vector not in the set of Once the gradients have been calculated, then adjust each
input training vectors is presented to a well trained network, weight value with a small amount (the learning-rate
the network will tend to exhibit generalization in its resulting parameter q) in the direction of the negative of the gradient.

Authorized licensed use limited to: Meerut Institute of Engineering and Technology. Downloaded on February 23, 2009 at 01:14 from IEEE Xplore. Restrictions apply.
1418

The weight-update equations on both layers take on the learning rate is increased. When the learning rate is too high
following forms: to guarantee a decrease in error, it gets decreased until stable
learning resumes.
Networks are, also, sensitive to the number of neurons in
on the output layer: wgJ(t+l)= wg,(t) + q 6& 9,
(17) their hidden layers. Too few neurons can lead to underfitting
awRj as shown in Fig.4. Too many neurons can contribute to
on the hidden layer: wi(t+l) = wh<t)+ q6k,xpi. (18) overfitting, in which all training points are well fit, but the
fitting curve takes wild oscillationsbetween these points.
Then, return to phase 1. All of these details are wrapped out in our MathematicaTM
The architecture of a backpropagation network is not code. The advantage of using backpropagation networks is
completely constrained by the problem to be solved. The shown in Table 1. The integral equation (9) is solved on a
number of network inputs to the network is constrained by PC Pentium 90 MHz. This is done using the numerical
the problem. The number of neurons in the output layer is solver with or without the backpropagation network
constrained by the number of layers between network inputs, architecture implemented. The numerical code with the
and the output layer and the sizes of the layers are up to the backpropagation network architecture implemented is an order
designers. The 2-layer sigmoidlinear network has been of magnitude faster than the other one.
proven to be able to represent any functional relationship

II
between inputs and outputs, if the hidden layer has enough Number of Equations 25 50
neurons. Backpropagation follows gradient descent on the without backpropagation
- _ _ 5.1 30.3 93.5 375
loo 250
error surface to minimize network error. Local minima may with backpropagation 0.6 3.8 10.7 42.2
trap the network. The more neurons in intermediate layers the Table 1 Solution tlme.
more freedom a network has (i.e. more variables to optimize).
The additional neurons increase the chance that even a local This is mainly due to the fact that with the back
minimum will yield a low error. The error surface of a propagation network the unknown coefficients I,, in (10) are
backpropagation neural network may be very complicated. To determined by calculating the gradient of a proper error
illustrate this complexity a typical error surface is shown in surface, i.e. by computing derivatives and not integrals as for
Fig.3. The problem is that nonlinear transfer functions in the Moment Method.
multilayered networks introduce many local minima in the
error surface. As gradient descent is performed on the error IV. CONCLUSIONS
surface, it is possible for the network solution to become
trapped in one of these local minma. This paper investigates a new approach utilizing an
To decrease the training time and the probability that the artificial neural network for solving scattering problems that
network will get stuck in a shallow minimum in the error can be cast in the form of integral equation. Neural networks
surface we added a term, called momentum, to the weight- offer the advantage of superior computational ability due to
update equations. This term has a significant effect on the the high degree of parallelism and interconnectivity. This
learning speed, in terms of the number of iterations required. ability makes neural networks attractive in many other
The idea behind momentum in a neural network is applications in engineering and sciences.
straightforward. After the code adjusts the weights during one The integral equation is cast in a discretized form by
training iteration, it saves the value of that adjustment; when expressing the unknown induced current on the metallic patch
calculating the adjustment for the next iteration, the code adds as a sum of weighted basis functions. The formulation is,
a fraction of the previous change to the new one. In terms of then, developed in terms of the incident field samplings.
an equation (in this case, for the hidden layer weights): In this paper we restrict our attention to the class of
backpropagation networks. The design of the backpropagation
network for the scattering problem proposed has been
wJI(t+l)= wJ1 + q $Jxpl + wJl(t) (19) completely carried out.
The resulting algorithm has been computed exhibiting
where a is called the momentum term, typically a positive considerable speedups.Due to the speedups, the computation
number less than one, and times become tolerable, allowing for the problem size, hence
the accuracy, to be increased.
The implementation of neural networks is finding
increasing interest with the emergence of optical and VLSI
Training time can, also, be decreased by the use of an techniques [151. With the growing interest in implementation
adaptative learning rate which attempts to keep the learning issues, we anticipate the neurocomputer to become
rate as large as possible while keeping learning stable. commonplace in the near future, making the proposed
Adding an adaptative learning rate can decrease training time approach, for solving scattering problems, very promising.
further.
This procedure increases the learning rate, but only to the
extent that the network can learn without large error REFERENCES
increases. Thus, a near optimal learning rate is obtained, [l] K R. Carver and J. W. Mink, "Microstrip Antenna Technology,"
When a larger learning rate could result in stable learning, the IEEE Trans. on Antennas Propagat., vol. AP-29, pp. 2-24, January 1981.
1419

(2.1 R. J. Mailloux, J;. F. McIlvenna, and N. P. Kernweis, "Microstrip Technology Letters, vol. 7, no.5, April 5 1994,pp. 247-250.
Array Technology," IEEE Trans. on Antennas Propagat., vol. AP-29, pp. [9] A. Toscano, and L. Vegni, "Spectral Electric Green's Dyad for a
25-37, January 1981. Grounded Bianisotropic Slab Fed by a Three-Dimensional Point-Source,"
Microwave and Optical Technology Letters, vol. 7 , no. 10, July 1994, pp.
[3] J. R. James and P. S. Hall, Handbook for hkrostrip Antennas, 448-450.
London, Peregrinus, 1989.
[lo] A. Toscano and L. Vegni, "Electromagnetic Waves in Planar
[4] L. Vegni, and A. Toscano, "Spatial Electromagnetic Fields in Chiral Integrated Pseudochiral a
Structures," Progress in Electromagnetic
Integrated Structures via Sommerfeld Integrals," Japanese Transactions on Research, special issue on Wave Interaction with Chiral and Complex
Electronics, Special Issue on Electromagnetic Theory, vol. E78-C, no.10, Media, vol. 9, December 1994, pp. 181-216.
October 1995, pp. 1391-1401.
[ l l ] B. Popovski, A. Toscano and L. Vegni, "Radial and Asymptotic
[5] A., Toscano, L. Vegni, "Spectral Dyadic Green's Function Closed Form Representation of the Spatial Microstrip Dyadic Green's
Formulation for Planar Integrated Structures with a Grounded Chiral Slab," Function," Journal of Electromagnetic Waves and Applications, vol. 9, no.
Joumal of Electromagnetic Waves and Applications, special issue on Wave 112, January 1995, pp. 97-126.
Interaction with C h i d and Complex Media, vol. 6, no. 5/6, May 1992, pp.
751-769. [12] B. Popovski, A. Toscano and L. Vegni, "Asymptotic Closed Form
Representation of the Spatial Microstrip Dyadic Green's Function,"
[6] A. Toscano, L. Vegni, "ElectromagneticField Computation in Planar Microwave and Optical Technology Letters, vol. 8, no. 2, February 1995,
Integrated Structure with a Biaxial Grounded Slab," IEEE Trans Magn., pp. 103-106.
vol. 29, no. 2, March 1993, pp. 17261729.
[13] D. W. Tank and J. Hopfield, "Simple Neural Optimization
[7] A. Toscano, L. Vegni, "Spectral Electromagnetic Modeling of a Networks," IEEE Trans. CAS, vol. 33, 1986, pp. 533-541.
Planar Integrated Structure with a General Grounded Anisotropic Slab,"
IEEE Trans. AntennasPropagat., vol. 41, no. 3, March 1993, pp. 362-370. [14] M. Kennedy and L. Chua, "Neural Networks for non Linear
Programming,"IEEE Tram. CAS, vol. 35, 1988, pp.554-562.
[8] A. Toscano and L. Vegni, "Novel Characteristics of Radiation
Patterns of a Pseudochiral Point-Source Antenna," Microwave and Optical [15] S . Haykin, eds., Neural Networks, IEEE Press, USA, 1994.

4
Fig.3 Typical error surface graph.

2 4 6 8 10 iterations
I
Fig.4 Network error with underfitting Fig.5 Network error without underfitting.

Authorized licensed use limited to: Meerut Institute of Engineering and Technology. Downloaded on February 23, 2009 at 01:14 from IEEE Xplore. Restrictions apply.

You might also like