Neural Nets For Optimization
Neural Nets For Optimization
Purdue e-Pubs
ECE Technical Reports Electrical and Computer Engineering
11-1-1991
Stefen Hui
San Diego State University Department of Mathematical Sciences
Stanislaw H. Zak
Purdue University School of Electrical Engineering
Lillo, Walter E.; Hui, Stefen; and Zak, Stanislaw H., "NEURAL NETWORKS FOR CONSTRAINED OPTIMIZATION
PROBLEMS" (1991). ECE Technical Reports. Paper 322.
http://docs.lib.purdue.edu/ecetr/322
This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact [email protected] for
additional information.
Neural Networks for
Constrained Optimization
Problems
Walter E. Lillo
Stefen Hui
Stanislaw H. ~ a k
TR-EE 91-45
November 1991
NEURAL NETWORKS FOR CONSTRAINED
OPTIMIZATION PROBLEMS
Stanislaw H. 2ak
Purdue University
This paper is concerned with utilizing neural networks and analog circuits
tion, is then used to solve minimum norm problems subject to linear con-
examples.
Key Words:
1. Introduction
problems seems to have been first proposed by Dennis (1959). Since then, vari-
ous types of "neural" networks have been proposed to obtain solutions to con-
zation network for solving linear programming problems using general princi-
network along with its stability properties were discussed by Smith and Port-
mann (1989). An extension of the results of Tank and Hopfield to more gen-
(1988). In addition, they noted that the network introduced by Tank and
added to account for the dynamic behavior of the circuit. Lillo et. al. (1991)
have shown that the above discussed approach implicitly utilizes the penalty
cussed in section 2 along with its circuit implementation. We show that the
penalty function approach allows one to better control the effects of the physi-
cal constraints of the network's building blocks than the previously proposed
linear neural network model. For a historical account of nonlinear neural net-
works, the reader may consult Grossberg (1988). In section 3 we discuss appli-
subject to Ax 2 b
where p = 1, 2, or oo. The minimum norm problems are important, for exam-
ple, in the context of the control of discrete processes (see Cadzow (1971) or
LaSalle (1986, Chp. 17) for more information related to the issue). The
behavior of the proposed networks are then tested on a numerical example and
5.
minimize f(x)
subject to
g(x) 2 0
h(x) = 0,
Chua and Lin (1984), and later Kennedy and Chua (1988), proposed
canonical nonlinear programming circuits for simulating the constrained
optimization problems of the above type (see Fig. 1 ). They analyzed the case
when only the inequality constraints are present. Their development was
based on the Kuhn-Tucker conditions from mathematical programming theory
(see for example Luenberger (1984) for more information on the Kuhn-Tucker
conditions). The functions dj, j = 1,...,q, on the left side of Fig. 1 are defined
by:
-cI if I 2 0
v = $j (I) = if 1 < 0 .
gj (x)c if g, (x) 5 0
=$j(-gj(x)) = j = 1, ...,q.
if g, (x) > 0
~j
Now applying Kirchhoff's current law (see for example Halliday and Resnick
dxk
Solving for - we obtain
dt
Note that if c +ca then, in the steady state, the Kuhn-Tucker conditions are
satisfied.
In this paper we examine the case where we have equality constraints as
and a general form which more readily lends itself to implementation. This
minimize f(x)
I + CP(X)) ,
zero a t a point if and only if all constraints are satisfied at that point. In this
where
The above penalty function P(x) is often referred to as an exact penalty func-
tion because for a sufficiently large finite value of c the penalty method
approximation, with the above P(x), yields the same global minimizers as the
constrained problem. The exact penalty functions have the drawback that
they are not usually differentiable. Having reviewed the pena1t.y method we
-
now introduce the proposed network. The functions S,,B and S7 in Fig. 2 are
Sa,,(x) =
I
-a
a
a
for x < - p
for -p
forx>p.
5 x IP
Figure 2. The proposed network for constrained optimization.
-
When a = P, we write S,,, as S,. We assume that a > 7. The li, and ij
terms are defined as:
= ic,(h,(x)) .
i,
-
Aj csgn(hj(x)) .
Remark
The Pj terms differ from the ,uj terms in the the canonical dynamical circuit of
Kennedy and Chua (Fig. 1) in that their values are bounded. This
the constraints.
Applying Kirchhoff's current law to the circuits on the right hand side of
Fig. 2 yields:
Note that if
-
then the term S,,@ saturates. If we assume the trajectory is in a region where
and
dxk
In addition, since Ck > 0, we conclude that if c- > p, then - and
h k dt
Thus
This implies that whenever go,8 saturates and the trajectory is in the region
where P(x) is differentiable, then P(x) is decreasing along that trajectory. Note
that the set of points where P(x) is not differentiable has an n-dimensional
Lebesgue measure zero and that the circuits are designed so that P is small and
thus go,@ will be saturated at almost all points outside the feasible region.
Thus, one would expect that the penalty function P(x) would decrease along
the trajectories outside the feasible region. Note that if So,@ operates in the
saturated mode, then the bound for the rate of decrease of the penalty function
jectory reaches the feasible region, then the circuit dynamics are governed by
the equations
problems subject to linear equality and inequality constraints the circuit shown
in Fig. 2 could be implemented using a neural network with the structure dep-
icted in Fig. 3. The implementation of the pnode is the same as was proposed
by Kennedy and Chua (1988) and is shown in Fig. 4. The implementations for
the h and x nodes are depicted in Fig. 5 and 6. It should be clear from the
would like to note that one may also use switched-capacitor circuits to imple-
dratic programming we will now examine how a network of this basic structure
mi.
*ip"'
In this section we show how the previously proposed neural network archi-
equation
where ckdRrn,ukeIR1,for k = 1, 2, ..., and F , G are constant matrices with
obtain
implies that we can drive the system to an arbitrary desired state, cd,from an
arbitrary initial state, to.Thus for sufficiently large N, ( N 2 m ) we can find
In the case where N > m there are an infinite number of input sequences which
would drive the system to the desired state. This can be seen more clearly if
A = [G, FG ,...,FNd2G,FN-'G] , T
xT = [uNPl, T
UN-2, ..., uoT I.
c d - ~ N c O =. ~
Ax=b.
completely controllable, N > m the rank of A is m and the null space of A has
dimension n - m > 0. From this it should be clear that the system of equa-
solutions ). Since there are many possible solutions, secondary criteria are often
should be used. Often it is desirable to find the solution which in some sense
minimizes the input x. This is the reason we consider the following con-
minimize 1 1 xll
subject to Ax = b ,
been studied fairly extensively (see for example Cadzow (1971,1973), Kolev
(1975)~or LaSalle (1986)). For the case of p=2, there are algorithms based on
linear algebra which solve this problem. When p =1 or p = oo, the problems
are somewhat more complex. There are algorithms based on results from func-
tional analysis which have been proposed to solve these problems (Cadzow
solution can be obtained is of the utmost importance. I t is for this reason that
we propose the use of analog circuits, or neural networks, which are capable of
given in the previous section can be applied to solving the problems of interest.
The first thing we notice with all these problems is that the constraints are
linear. Thus in the case where p = 2, since the objective function of the
dratic. However, as shown below, the components of the gradient of the objec-
This being the case, a component of the gradient of 1 1 xll can be approximated
dratic. In addition, we can see from the equation below that the components
otherwise
this is done, consider the level surface 1 1 xllm = a, where a > 0. This level
Thus the problem can be viewed as finding the smallest value of a > 0 such
minimize x n + ~
subject to
h(x) = Ax* - b = 0
g1(x) = Xn+l 2 0
In order to test the ideas presented in this paper, simulations of the pro-
where SalPand S, are as defined in Section 2 with a = 12, ,b' = 0.5, and 7 = 6.
The problem which we choose to simulate is taken from Cadzow (1973) and
subject to Ax = b
where p = 1, 2, or m, and
The variables xj, j = 1,...,n, are constrained to be in the interval [-12,121, The
point
which gives 1 1 xll oo = 0.372.
Thus the results of the simulations closely correspond to the analytical solu-
tion.
Another important consideration is the speed with which the network con-
verges to the correct solution. This depends on the value of the time constants
and the initial condition of the network. In the above simulations we assumed
the variables for the three problems we can see that the network converged to
.5670 -
.3780 -
.I890 -
0 .ooo
0.0 1.0 2.0 3.0 9.0
1-Norm case
%
.-I9
.El09
.1YS9
-
-
-
2-Nora c a s e
2-Norm c a s e
.SO28 - 2-Norm c a s e
\
.377 1 -
.i?slV-
.I257 -
2-Norm c a s e
m
0.000 ! I
0.0 1.0 2.0 3.0 r.o
time ( time c o n s t a n t s ) time ( time c o n s t a n t s )
.OW6 - .2?90 -
4
.056Y - .I860 -
.om2 - .OW9 -
2926 -
t, .lmo-
-1950 -
Figure 12. Trajectory of the augmented variable for the case p=m.
5 . Conclusions
Cadzow, J.A. (1973). 9"F'nctional analysis and the optimal control of linear
Chua, L.O., and Lin, G.-N. (1984). "Nonlinear programming without computa-
tion", IEEE Trans., Circuits and Systems, vol. CAS-31, no. 2, pp. 182-188.
for differential o p timization," Int. J. Circuit Theory and Applications, vol. 19,
Halliday, D., and Resnick, R. (1978). Physics, Third Edition, J . Wiley & Sons,
New York.
Hall.
Kennedy, M.P., and Chua, L.O. (1988). "Neural networks for nonlinear pro-
gramming." IEEE Trans. Circuits and Systems, vol. 35, no. 5, pp. 554-562.
Kolev, L. (1975). 'Iterative algorithm for the minimum fuel and minimum
amplitude problems for linear discrete systems." Int. J. Control, vol. 21, no. 5,
779-784.
LaSalle, J.P. (1986). The Stability and Control of Discrete Processes. New
York, Springer-Verlag.
IN.
Massachusetts, Addison-Wesley.
Smith, M.J.S., and Portmann, C.L. (1989). "Practical design and analysis of a
simple "neural1' optimization circuit", IEEE Trans. Circuit and Systems, vol.