Chen1996A Class of Smoothing Functions For Nonlinear and Mixed
Chen1996A Class of Smoothing Functions For Nonlinear and Mixed
Abstract. We propose a class of parametric smooth functions that approximate the fundamental plus function,
(x)+ =max(O. x). by twice integrating a probability density function. This leads to classesof smooth parametric
nonlinear equation approximations of nonlinear and mixed complementarity problems (NCPs and MCPs). For
any solvable NCP or MCP, existence of an arbitrarily accurate solution to the smooth nonlinear equation as well as
the NCP or MCP, is established for sufficiently large value of a smoothing parameter o. Newton-based algorithms
are proposed for the smooth problem. For strongly monotone NCPs, global convergence and local quadratic
convergence are established. For solvable monotone NCPs, each accumulation point of the proposed algorithms
solvesthe smooth problem. Exact solutions of our smooth nonlinear equation for various values of the parameter CL,
generate an interior path, which is different from the central path for interior point method. Computational results
for 52 test problems compare favorably with those for another Newton-based method. The smooth technique is
capable of solving efficiently the test problems solved by Dirkse and Ferris [6], Harker and Xiao [ 111 and Pang &
Gabriel [28].
1. Introduction
where x and y are vectors in R” and the symbol I denotesorthogonality, plays a funda-
mental role in mathematical programming. Many problems can be formulated by using
this complementarity condition. For example, most optimal@ conditions of mathematical
programming [24] as well as variational inequalities [4] and extended complementarity
problems [9, 21, 38-Jcan be so formulated. It is obvious that the vectors x and y satisfy
complementarity condition if and only if
x=(x-y)+,
*This material is based on research supported by Air Force Office of Scientific Research Grant F49620-94-l-0036
and National Science Foundation Grant CCR-9322479.
98 CHEN AND MANGASARIAN
also given and convergence is established. In Section 5 we show that exact solutions of
our smooth nonlinear equation, for various values of the smoothing parameter /3 generate
an interior path to the feasible region, different from the central path of the interior point
method [ 171. We compare the two paths on a simple example and show that our path gives
a smaller error for the same value of the smoothing parameter B. In Section 6, encouraging
numerical testing results are given for 52 problems from the MCPLIB [7] which includes
all the problems attempted in [6, 111 and [28]. These problems range in size of up to 8192
variables. These examples include the difficult von Thiinen NCP model [28, 371 which is
solved here to an accuracy of 1.Oe - 7.
A few words about our notation. For f: R + R and x E R”, the vector f(x) in R” is
defined by the components(f (x))i = f (Xi), i = 1, . . . , n. The support set of f(x), which
is the set of points such that f(x) # 0, will be denoted by supp(f (x)}. The set of m-by-n
real matrices will be denoted by RmMI. The notation 0 and 1 will representvectors with all
components0 and 1 respectively, of appropriate dimension. The infinity, 1, and l2 norms
will be denoted by ]I . Iloo, I] . 111and I] . 112
respectively. The identity matrix of arbitrary
dimension will be -denoted by I. For a differentiable function f: R” + Rm, V f will
denote the m x n Jacobianmatrix of partial derivatives. If F(x) hasLipschitz continuous
first partial derivatives on R” with constant K > 0, that is
a(x) =
11 ifx>O
0 ifxF0
The step function a(x) can in turn be written as, a(x) = [:,6(y) dy, where 6(x) is the
Dirac delta function which satisfiesthe following properties
+CO
S(x) 10, S(y) dy = 1.
s --oo
Figures 1to 3 depict the above functions. The fact that the plusfunction is obtainedby twice
integrating the Dirac delta function, prompts us to propose probability density functions
as a meansof smoothing the Dirac delta function and its integrals. Hence we consider
the piecewise continuous function d(x) with finite number of pieces which is a density
function, that is it satisfies
00
d(x) 10 and d(x) dx = 1. (1)
J --bo
10
-2
-10 4 644 0 2 4 6 8 10
2
44 = S wcy -
1.5 *
0.5 -
0.
-1
-104 d 4 -2 0 2 4 8 8 10
1.2 - w - -
l-
0.4 -
o- * a * .
-104 6 4 -2 0 2 4 6 0 10
,. 1 x
t(x, PI = $ s (2)
0
where p is a positive parameter. When B goesto 0, the limit of i(x, jl) is the Dirac delta
function 6(x). This motivates a classof smooth approximations asfollows:
x
i(t,j?)dtx ff(x)
Sk B>= s--bo
and
(3)
Therefore, we can get an approximate plus function by twice integrating a density function.
In fact, this is the sameasdefining
x
ixx, B) = +O”(x - t)+i(t, /3) dt = (x - t)i(r , p) dt. (4)
/ --co J-cm
This formulation wasgiven in [18] and [ 13, p. 121for a density (kernel) function with finite
support. We will give our resultsin termsof a density function with arbitrary support. This
includes the finite support density function asa specialcase.
Proposition 2.1. Let d(x) be a probability densityfunction and i(x, B> = id($), where
#I is a positive parameter: Let d(x) satisfy the following assumptions:
(Al) d(x) is piecewise continuous withBnite number ofpieces and satisfies (1).
W) W4lcqx, = ./-‘,” 1x1d(x) dx < +CQ.
Then the dejinitions of $(x, @) given by (3) and (4) are consistent.
Proof: By the definition (2) and assumption(A2), we have that p(x, /3) defined by (4)
satisfies
r z
j%x,B)=x s-co’ d(t) dt -p
s ’ td(t) dt
--oo
(5)
By direct computation,
5
P’CX,
B>= J-co d(t) dt
x
= t^(t, /?) dt = sI(x, #?) (6)
s -m
102 CHEN AND MANGASARIAN
Hence the derivatives of 5(x, B) defined by (3) and (4) are the same and the difference
between the two representations of j?(x, p) is a constant, say c. If we let x approach -oo
in both (3) and (4), then j(x, /I) approaches 0 in both, and hence c = 0. Therefore the
definitions of j(x, /3) given by (3) and (4) are consistent. 0
Now we give properties of j(x, ,5) that show that it is an accurate approximation of the
plus function (x)+ as p approaches zero.
Proposition 2.2 (Properties of fi(x, jS>,j? >O). Let d(x) and i(x, p) be as in Proposition
2.1, and let d(x) satisfy (Al) and (A2). Then Z?(x, /3) has thefollowing properties:
(1) 5(x, /I) is continuously differentiable. If, in addition, d(x) is k-times continuously
differentiable, 6(x, t?I) is (k + 2)-times continuously differentiable.
(2) -028 5 Ij(x, B) - (xl+ 5 DIP, where
0
D1 = Ix Id(x) dx (7)
s -co
+bo
D2 = max xd(x) dx, 0 (8)
IS-m
(3) 5(x, p) is nondecreasing and convex. In addition, let d(x) satisfy
Then $(x, /I) is strictly increasing and strictly convex in x for a@ed B P- 0.
(4) 0 5 $‘(x, /3) 5 1 and if(A3) is satisfied, then 0 -C j’(x, #?) c 1.
(5) Let (A3) be satisfied and D2 = 0, then
Proof:
(1) By Eq. (6) in the proof of last proposition, the conclusion follows.
(2) If x 2 0, by using (5), we have that
I. d
BCX, B) - (x>+ = x ’ d(t) dt - /9 ’ td(t) dt -x
s -co s -co
co 00 ce
= -x d(t) dt + /3 td(t) dt - /3 td(t) dt
s s s
=Z$i(t-;)d(t)ci-/!?~~td(t)~w
P
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 103
Therefore
b(x,
B>
-(x>+
2-B
s03 -cm
td(t) dt 2 -f?Dz
and
ItId dt = BDl.
and
B)=s%
y<x, --co
d(t) dt 2 0
and
Therefore b(x, /?) is nondecreasingand convex. If, in addition, (A3) is true, then
b’(x, B> > 0 and <$‘<x, B> - $‘(y, /9)(x - y) > 0, for x # y
and
Therefore
Now we have a systematicway for generatinga classof smoothplus functions. Given any
probability density function d(x) satisfying (Al) and (A2) we define i(x, /I) and b(x, fi)
asin (2) and (3) respectively. The smoothfunction j(x) /3) approximatesthe plus function
with increasing accuracy asthe parameter#l approaches0. The properties of the function
j(x, p) are given in Proposition 2.2 above.
We now give examplesof smoothplus functions. The first example, which will be used
throughout this paper, is basedon the sigmoid function of neural networks [3, 12, 201 and
defined asfollows:
1
s(x, a) = a!>0
1 + e--ax ’
This function approximatesthe stepfunction o (x) as(II tendsto infinity. Since the derivative
with respect to x of this function tendsto the Dirac delta function as(Ytends to infinity, it
follows that a! plays the role of $ and we shall therefore take
1
a! = -.
B
Example 2.1 (Neural Networks SmoothPlus Function [3]). Let
e-I
d(x) =
(1 + e+)2
Letting IX = i, we have
Figures 4 to 6 depict the functions p(x, 5), s(x, 5) and t (x, 5) respectively.
10 F -
P(G 4
5
-10 -5 6 4 -2 0 2 4 5 5 10
2
8(W) --
8(W)
1.5 '
1 -
0.5 -
00.
0.5 .
-1L
-1 - a * *
-10
-10 .5
d 4 -4 -2 0 2 4 5 5 10
1.2
0.6
0.6
0.4
0.2
e
--lo d -6 -4 -2 0 2 4 6 8 10
Example 2.2 (Chen-Harker-Kanzow-Smale Smooth Plus Function 121, [15] and 1361).
Let
2
d(x) =
(x2 + 4)f
1 if05xll
d(x) =
0 otherwise
0 ifx CO
b(x,B)= 5 ifOlx (fi
i x- $ ifx>/3
This function can also be obtained by applying the Moreau-Yosida regularization [13,
p. 131to the plus function.
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 107
1 if-isxs$
d(x) =
0 otherwise
1
0 if xc-g
jxx, B) = $j(x + $)” if 1x15 $
x if x>$
Note that in Examples 2.3 and 2.4 above, the density function d(x) hascompact support
while the smoothfunction F(x, B) is only oncecontinuously differentiable. In Examples 2.1
and 2.2, d(x) hasinfinite support while the functions p(x, cr) and b(x, B) are differentiable
infinitely often.
We summarizethe various functions introduced asfollows:
1 1
pb, a) = @ x, A ,I s(x,a) =i x,- -A- t(x,cx) =; x,-
( > ( (Y> ( a>
Becauseof our favorable experience with the function p(x, o) [3] on linear complemen-
tarity problems and linear and convex inequalities, we chose it for our numerical experi-
ments. Further comparisonsusing different approximations to the plus function are left for
future work.
0 5 x I F(x) 2 0 (9)
108 CHEN AND MANGASARIAN
Here F(x) is a differentiable function from R” to R”. By using the smoothfunction 3(x, B)
introduced in last section, we considerthe smoothnonlinear equation
x = (x - F(x))+ (11)
We first show that a natural residual for the NCP is easily bounded by a corresponding
residual for the nonlinear Pq. (10).
Lemma 3.1.
Proof:
The above result is also true for any monotone norm [25].
We first consider the strongly monotone NCP, that is there exists a k > 0 such that for
anyx, y E R”
Since the NCP is strongly monotone, it has a unique solution [lo]. The following error
bound for the strongly monotone NCP is given asTheorem 3.2.1 in [31].
Lemma 3.2. Let the NCP be strongly monotoneand let F(x) be Lipschitz continuous.
Thenfor any x E R”
where X is the unique solution of the NCP and C, is a condition constant of F independent
ofx.
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 109
Now, we give an error bound for the NCP by using the smooth function j?(x, ,3). By
Lemma 3.1 and Lemma 3.2, it is easy to get the following lemma.
Lemma 3.3. Let the NCP be strongly monotone and let F(n) be Lipschitz continuous.
Then for any x E R”
where X and C, are defined in Lemma 3.2, and y!, and the constants Dt, 02 are dejined in
Lemma 3.1.
Let the residual f(x) of the nonlinear Eq. (10) be defined as follows
We now prove that if x is a stationary point of f(x) for a monotone F(x), then x must be a
solution of the nonlinear Eq. (lo), and hence by (15) x is an approximate solution of the NCP.
Proposition 3.1. Suppose that d(x) satisjes (Al)-(A3) and p(x, /?) is defined by (4).
For any monotone NCP, we have that V R(x) is positive definite. In addition, let x be a
stationary point of f(x), then x mustbe a solution of the nonlinear Eq. (IO).
Proof: By definition,
By (4) of Proposition 2.2, we have 0 c j’(x, B) < 1 and hencethe diagonal matricesabove
are positive definite. Since V F (x) is positive semidefinite,it follows that V R(x) is positive
definite. Let x be a stationary point of f(x), then
Vf(x) = VR(x)rR(x) = 0.
When F(x) is strongly monotone and Lipschitz continuous, then the level setsof f(x)
are compact. We state this result asthe following proposition.
Proposition 3.2. Consider the strongly monotoneNCP with Lipschitz continuous F(x).
Then f(x) defined by (16) has compact level sets.
Proof: Supposenot, then there exists a sequence{xk} c R” and a positive number M such
that llxk 112
--+ co ask + co, and llxk - $(xk - F(xk), 8)112_( M. Then by Lemma 3.3,
where X is the unique solution of the NCP. Let k ----f 00, the left hand side of the above
inequality goesto 00 and the right hand side staysfinite. This is a contradiction. Hence the
level setsof f(x) are compact. 0
We now show that, for the strongly monotone NCP with Lipschitz continuous function
F(x), the nonlinear Eq. (10) always hasa unique solution.
Theorem 3.1. Supposethat d(x) safisJes(Al)-(A3), j(x, /?) is deBned by (4), F(x) is
strongly monotone and Lipschitz continuous. Then the nonlinear Eq. (10) has a unique
solution.
Proof: By Proposition 3.2, the level setsof f(x) are compact. So minXeRaf(x) must
have a solution X, which is a stationary point of f(x). By Proposition 3.1 we get that x
satisfies(10). If y is another solution of (lo), then
for somet E [0, 11. Since VR is positive definite by Proposition 3.1, it follows that x = y.
Therefore Eq. (10) has a unique solution. 0
Let x(p) be a solution of (10). Then x(p) = @(x(B) - F(x@)), p). By Lemma 3.3, we
have the following theorem which boundsthe distancebetween the solution x(/l) of (10)
and the solution point of the original NCP (9).
Theorem 3.2. Consider a strongly monotone NCP with Lipschitz continuous F(x). Let
x(B) be a solution of (10). Then, for the solution X of the NCP (9), we have that
Here C, is the condition constant defined in Lemma 3.2, yp and D1, 02 are constants
de$ned in Lemma 3.1.
By the above result, we know that if p is sufficiently small, a solution of (10) can
approximate the solution of NCP to any desiredaccuracy. Hence we can solve (10) to get
an approximate solution of the NCP.
For most part of this remaining section, we consideronly the function p(x, 01)defined in
Example 2.1. We explore further the property of real numbersx and y that approximately
satisfy the following equation
x=p(x-Yea)
which is related to Eq. (10) that generatesan approximate solution of the NCl? We claim
that suchx and y will approximately satisfy
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 111
In order to prove this fundamental fact we establishthe following lemma,the proof of which
is relegatedto the Appendix 1.
max h(x) 5 2.
rqo,--logSI
max h(x) 5 2.
XC[O,~) e
-2 5 x - p(x - y, a) 5 0,
a!
where 61 > 0, then the complementarity condition 0 5 x I y 2 0 is approximatdy
C(h)
t-x)+ 5 19 C-y)+ I ;, (XY)+ I ---p
where C(&) is the constant defined in Proposition 3.3. Note that asu + 00, the comple-
mentarity condition 0 I x I y 1. 0 is exactly satisfied.
--615x-p(x-y,a)10,
CY
where 61 > 0. Then
CC&)
t-x>+ I ;, t-Y)+ 5 2, by)+ I Q2’
where
-1 ( x - (x - y) - L log(1 + e-+y)) _( 0,
a! (II
we have
we have x p - %. Hence
61
(-xl+ 5 -.
a
Similarly,
C-y)+ 5 ;.
Therefore
and
(18)
CaseI. 0 -c 6 -c 1.
If y > 0, then by (17),
x ( -‘log 6.
a!
Hence
Otherwise y 5 0, then
x&ogs50.
a!
Case 3. 6 = 0.
In this case, we have x 1 0 and y > 0. By using (iv) of Lemma 3.4,
1
(XY)+ i - max h(x) 5 -$a
a2 xaca)
C(h)
(XY)+ I -. q
a!2
always has a solution for #l > 0. In particular, for the p(x, o) defined in Example 2.1 we
have that for all 61 > log 2, the following system of inequalities
always has a solution. Hence by Proposition 3.3, a solution of (20) will approximately
satisfy the NCP condition
x 2 0, F(x) 2 0, XTF(X) = 0.
Consequently, Proposition 3.3 can be used to establish the following useful theorem.
c-x>+ 5 -1,
61 (-F(x))+ 5 -1,
f-5 (XTF(X))+ 5 9,
a! (Y
We now specify our computational algorithm for solving the NCP by smoothing. The
algorithm consists of a Newton method with an Armijo line search with parameters 6 and
asuchthatO<S<landO<a<i.
dk = -VR(x$‘R(xk)
k = k + 1 go to step (I).
The above algorithm is well defined for a monotone NCP with a continuously differen-
tiable F(x). We will state the following convergence theorem [5]. We omit the proof that
is similar to the proof of Theorem 4.3.
F(x)-w+u=O
05x-ZlwzO (21)
o~ulu-x~o
This MCP model includes many classes of mathematical programming problems, such
as nonlinear equations, nonlinear programming, nonlinear complementarity problems and
variational inequalities.
By using the smooth function j(x, /3) instead of the plus function, we reformulate the
MCP approximately as follows.
Fori = l,...,n:
F;:(x) = 0
Xi - li - $(-Xi - li - Fi(x), p) = 0
I;;:(x)-wi+ui=O
Xi - li - ,G(Xi - li - Wi, @) = 0
ui-Xi-~(Ui-Xi-Ui,B)=O.
R(x,w,u)=O (23)
Note that the natural residual for the MCP is given by the left hand side of above relation
with the jj function replaced by the plus function. We denote collectively this natural
residual by
r(x, w, u) (24)
116 CHEN AND MANGASARIAN
Now we give a lemmathat boundsthe naturalresidualof the MCP by the residualof Eq. (23)
and vice versa. The proof is a simple application of the properties of the p function.
Lemma 4.1. Let N be number of equationsin (23) and p(x, a) de@ed in Example 2.1.
Then
alog
IINx, w ~)I12 5 lb-(x, w, ~)I12 +
a0
Let f(x, w, u) be the residual function of the nonlinear Eq. (23) defined as follows
Theorem 4.1. Supposethat d(x) sati@es(Al)-(A3) and b(x, /?) is dejined by (4). Con-
sider a solvable mixed complementarity problem (21) with monotone F(x) and 1, u E R”.
The nonlinear Eq. (23) has a solutionfor suflciently small /%
Proof: We shall prove that a level set of f(x, w, u) is nonempty and compact. First we
will prove that the set X = {x I f(x, w, u) 5 C} is compact for all C E R. Since f is
continuous, the level setX is closed. Hence we only needto showthat the set X is bounded.
Supposenot, then there exists ink} E X and there exists 1 5 i 5 n such that xi” goesto
+oo or -co. Without lossof generality, we assumethat xi” goesto +oo. Then the residual
corresponding to the following equation approaches00:
Ui - Xi - $(Ui - Xi - Vi, B) = 0.
This contradicts the fact that xk E X. Let C = fn max{Di, Dz}~~~, where Di and DZ
are nonnegative constants defined in (7) and (8). It is easy to show that the level set
Levc(f) = {(x, w, u) 1 f(x, w, u) 5 C} is not empty. Now we will prove that Levc(f)
is compact for
IIliIll<i<n(Ui - li)
B<
,/Zmai{Dl, DzJ’
We have proven that the x part must be bounded. Therefore, if the level set Levc(f) is
unbounded,thereexists (xk, wk, u’) E Levcf f) suchthat (wk, uk) are unbounded. Without
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 117
Xi - li - j(Xi - li - Wi, B) = 0.
goesto 00 ask + 00. But (xk, wk, uk) E Levc(f), which is a contradiction. Otherwise,
wf + +CS. By the equation
Fi(X) - Wi + Vi = 0
and the fact that xk is bounded,we get that ask -+ 00, I$ + +m. Hence, ask -+ cm,
f(Xk, Wk, U’) 2 i((*i -Xf - b(Uf - Xf - $9 fi))* + (Xf - li - j(Xf - Zi - Wf, 8))“)
--+ k((Ui - ifi)* + (Xi - li)*) 2 f (Ui - li)* > in max{Di, D*}*p* = C
for all
This contradicts (xk, wk, uk) E Levc(f). Hencethere existsalevel set of f(x, w, u) which
is nonempty and compact. Therefore the problem
min f(x, w, u)
x.w,u
VR(x,w,u)= [KY!; $ iJ
Notice that
Vf(x, w, u) = 0 d R(x, w, u) = 0. 0
118 CHEN AND MANGASARIAN
From now on, we only consider the function p(x, cr) defined in Example 2.1. The
following theorem is a direct application of Proposition 3.3. that can be proved by applying
the Proportion 3.3 to each equation.
Theorem 4.2. Consider a solvable mixed complementarity problem (21). Let 61 1 log 2
and a > 0. Then the following system of inequalities, i = 1, . . . , n:
Case 1. lj = --oO and Ui = Co:
-1 < Fi(X) = 0 5 1
a!- cd
-'_(Xj-li-~(Xj-lj-Fj(X),B)iO
cd
&
-- 61 5 F;:(X)-Wi+Vj I-
o! 01
where p(x, a) is defined in Example 2.1, always has a solution (x, w, v) which satisjies
the MCP conditions approximately in thefollowing sense,for i = 1, . . . , n:
Case1. li = -CXJand ui = 00:
(lj-Xi>+
I z7 (--Fi(x)>+
5z, ((Xi -li>Fi(X))+ I: %$9
Case3. li = --oO and Ui < 00:
61
(Xi - Ui)+ I -9 (4(X))+ I ZT (-(Ui - Xi)Fi(X))+ 5 y,
ff
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 119
C(b)
(xi -Ui>+ I $t (-vi>+ I z, ((Ui -Xi>%>+ i -
cd2 ’
Now we state the smooth method for the mixed complementarity problem based on the
Newton Algorithm 3.1 in which the smoothing parameter will be adjusted. In the algorithm,
we adjust the smoothing parameter a! inverse proportion to the natural residual r(x, w, u)
of (24) for the MCP in the following way. Let N be the total number of nonlinear equations
in (23) and let (x, UI, V) be current point. Let
The following smooth algorithm generates an E-accurate solution for the MCP, in the sense
that the natural residual r(x, w, u) of (24) satisfies Ilr(x, w, u)llm 5 6.
In order to get an r-accurate solution for the MCP. We need (Y sufficient large. We will
establish a simple lemma before we get such o.
then
maxI&, XT&J,
bin@, b)l I t
a!
where C(&) is defined in Proposition 3.3.
Proof: Without loss of generality, we assume that a 5 b. If a 2 0, then (ab)+ > a2.
Therefore
Ifa (: 0,
61
Jmin(a, b)J = -a = (-a)+ 5 -.
a!
120 CHEN AND MANGASARIAN
Algorithm 4.1 (Smooth Algorithm for MCP). Input tolerance E, parameter u1 2 v > 1
and initial guess x0 E R”:
(1) Initialization. For1 5 i 5 nofCase4of(22),letw~ = (Ft(xo))+, I$ = (-Ft(xg))+,
k = 0 and (110= a(xo, wg, ~0). Choose amax 2 $
(2) If Ilr(xk, wk, ~k)iico 5 c, stop.
(3) Newton Armijo step. Find (xk+l , wk+l , Ukfl) by a Newton-Armi,io step applied to
R(x, w, u) = 0.
Let Z denote the index set of the Fi of Case 1, J of Case 2, K of Case 3 and L of Case 4
of (22). In order to characterize the nonsingularity of VR, we now give a definition of a
regular MCP Note that the monotone NCP is regular. More generally, an NCP with a PO .
Jacobian matrix is regular.
is nonsingular, for all positive diagonal matrices DJ , Dk and DL, that have the dimensions
of I .Z1, I K I and IL I respectively.
Theorem 4.3. Consider a solvable regular mixed complementarity problem (21) with
F(x) E LCk(R”). Then
(1) The sequence {xk, wk, uk} defined in Algorithm 4.1 exists.
(2) Any accumulation point of the above sequence is an e-accurate solution of the MCP
(21).
NONLINEAR AND MIXED COMPLEMENTARI-R’ PROBLEMS 121
(3) Zf an UCCUmUhtiOn point eXiStS, the whole sequence{xk, wk, vk} converges to an
E-accurate solution quadratically.
(4) Zf; inn&$ion, the level set {(x, w, v) 1 11r (x3 w, @II2 5 Ilr(x0, ~0, v0>112 +
v-l
g ) is compact, the sequence{xk, wk, vk} convergesto an E-accurate solution
at a qu2fratic rate.
AJ
Z -AL, AL,
I- ALA-1 AL2
O
I-AJ
Z-AK
+
0
0
0
122 CHEN AND MANGASARIAN
Therefore
VF,(x) 0’
VF.I(X> 0
VFK(X) 0
+
VFL(X) 0
0 0
0 0. 1 AL? - I ALL _
Since VR(y) is nonsingular,on the compact S, there exists K(S) and k(S) such that
Consequently
2 -hk,Vf(yki)Tdk, - T@(Yk,)‘R(yk,)
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 123
= ah ’ - TAk,) Ivf(yk,)‘dki 1
2(1 -a)
z hk,a(Vf(yki)Td(Yk,)l, if kk, 5 KIK(S)
By the rule of choosing hk,, we have kki 2 Sw, where S is the constant used in
Armijo stepsize. Therefore
= 2as ~~v~(Yk,)‘vR(Yki)-l~~(yki)-~~~(yk~)~
Since yk, - j, we have Of(y) = 0. Thus R(y) = 0 and f(J) = 0. This contradicts
the assumptionf(j) > 0. This casecannot occur.
(3) By the analysis in (2), we have yk, + j, R(y) = 0 and R(y) E LCK, (S). Therefore
2JG
if IlR(~)ll2 5 KIKcSj
Hence, if y is close enoughto y, the Newton step accepted. According to the standard
result of local quadratic convergence for Newton Method, Theorem 5.2.1 in [S], the
conclusion follows.
(4) IAd’i =o, l,..., be the sequenceof different parametersrx usedin Algorithm 4.1.
Let{kj}, i =O, l,..., with kc = 0, be the indicessuchthat the parametercxchanges,
that is for ki 5 k 5 ki+l - l,ok=&. For~eandya,byJkmna4.1,wehave
filog2
IIWYk)llZ I IIWYo)ll2 5 Il~(Y0)112 + do
By Lemma 4.1,
filog2 2Jmog2
ll~(Yk>llZ 5 IIWYdll2 + I ll~(YO>ll2 +
a0 a0
2filog2 + filog2
llwk)112 I II~(Yk~-l)ll2 5 Il~(Yo)ll2 + (yo
vu0
By Lemma4.1,
If the level set (y 1 Ilr(y)ll2 5 Ilr(yo)ll2 + &2<F2) is compact, there exists an accu-
mulation point. By (2) and (3), the whole sequenceconvergesto an e-accurate solution of
MCP (21). D
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 125
We shall give our numerical test results for Algorithm 4.1 in Section 6, after relating our
smooth approach to the central path of the interior point method [ 171 in Section 5.
In this section, we consider the NCP (9). Let the density function d(x) satisfy (Al)-(A3)
and 02 = 0, and let j?(x, B) be defined by (4). If x solves the nonlinear Eq. (10) exactly, then
where the last inequality follows from the fact that j(t, fi) > e, from (5) of Proposition
2.2. Hence
and x belongs to the interior of the feasible region {x 1 F(x) 2 0, x >_ 0) of the NCP Hence
an exact solution of (10) is interior to the feasible region. However the iterates of the smooth
method, which are only approximate solutions of (lo), are not necessarily feasible. For the
function j defined in Example 2.2 [2, 15,361 , the exact solution x of the Eq. (10) satisfies
which is precisely the central path of the interior point method for solving NCP. Methods
that trace this path but allow iterates to be exterior to the feasible region have been proposed
in [2, 15, 361. In [ 161, the relation between Smale’s method 1361 and the central path was
pointed out. For our function b defined in Example 2.1, the solution x of the nonlinear Eq.
(lo), for different values of 8, constitutes another path in the interior of the feasible region
that satisfies:
We now compare our path and the central path of the interior point method by using a very
simple example.
0.8 -
c10.5-
c- Smooth Path
0.1 -
Figure 7. Comparison of the interior smooth path generated by an exact solution of the smooth nonlinear Eq.
(16versus the c&al path for Example 5.1.
0.9 -
0.8 -
0.2 -
Figure 8. Error comparison for the central path versus the smooth path for Example 5.1.
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 127
6. Numerical results
In this section, we give our computational experience with the smooth Algorithm 4.1 for
the MCP. We implemented the smooth Algorithm 4.1 with an SOR preprocessor if all
diagonal elements of the Jacobian matrix are positive. An initial scaling of the function
Fi (x), inversely proportional to the absolute value of the diagonal elements of the Jacobian
matrix, is performed if 1Vi Fi (xc) 1 2 100. The details of implementing the smooth algorithm
are given in Appendix 2. For comparison, we also give the results for the PATH solver [6].
Both algorithms were run on a DECstation 5000/125. Among the 52 test problems, which
includes all the problems attempted in [ 11,281 and [6], 5 1 problems are from the MCPLIB
[7], and one is the generalized von Thtinen model from [28] and [37]. Our smooth algorithm
was run using one set of default parameters and so was the PATH solver. The smooth
algorithm is written in the C language and implemented by using the GAMSKPLIB [8].
A MINOS routine [23] was used to perform a sparse LU decomposition for solving sparse
linear equations. Both algorithms use the same convergence tolerance of E = 1.Oe - 6.
Table 1 gives a simple description of the test problems [7].
The average CPU times taken by PATH solver and smooth algorithm for all small prob-
lems are depicted in Figure 9. Figures 10, 11 and 12 depict the CPU times for all remaining
problems except the von Thtinen model. We note that the PATH solver [6] is faster than
Josephy’s Newton method [14] and Rutherford’s GAMS [8] mixed inequality and linear
equation solver (MILES) [35] which is also Newton-based. Figures 9 to 12 indicate that
our smooth algorithm is faster than PATH solver for the larger problems, whereas PATH
solver is faster on smaller problems.
The newest version of PATH (PATH 2.7) that uses a Newton method on the active set [l]
as a preprocessor, improves solution times on the larger problems. Our smooth method can
be similarly improved by adding the projected Newton preprocessor. We have compared
PATH and SMOOTH with a Newton preprocessor on a Sun SPARCstation 20. The results
are given in Figures 13 to 16. It can be seen that with a Newton preprocessor, the solution
times are very similar for PATH and SMOOTH for larger problems, whereas PATH is still
better for the smaller problems.
As mentioned in [28], the generalized von Thtinen model is an NCP with 106 variables.
This is a very difficult problem that has challenged many of the recently proposed algorithms
[28, 371. In order to guarantee that the function F(x) is well defined, we added a lower
bound of 1.Oe-7 to variables xi to x26 as suggested by Jong-Shi Pang. We used three starting
points. In the first, we set all variables to 1, as suggested by Michael C. Ferris; the second
one is a starting point suggested in [37], while the third is the point suggested in [37] and
modified by Jong-Shi Pang. SMOOTH, with or without the Newton preprocessor, solved
the problem from all the three starting points. Solution times did not change by adding
the Newton preprocessor. We report times for SMOOTH with the preprocessor. Starting
with the first point, SMOOTH took a long time, 95.44 seconds to solve the problem. From
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 129
2soo
WOO-
lsao-
1
I!!
'WOO-
SOO-
smodl Hms
1ooo2ooosooo4ooosoOOwo07wo8ooogoM)
n
14 16 fdhd path e-
=1+ + * + ’ ,
1.8 -
I
I
1.6-
1.4-
0.6
IJ
0.4 t ,t
0.2 I, I’
1 A___* I”\ss- ,,’
’ :\ ,
.--------
_---
-
___---
path time
ol Ad +
0 10 21 D 30 40 60 80 70 60 60 loo
n
Figure 13. Smooth with Newton preprocessor versus PATH 2.7 for small MU?
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 131
Figure 14. Smooth with Newton preprocessor versus PATH 2.7 for optimal control problem(bert-oc.gms).
Figure 15. Smooth with Newton preprocessor versus PATH 2.7 for optimal control problem(opt-cont.gms).
132 CHEN AND MANGASARIAN
0
0 1000 WOO aooo 4000 5wo woo
n
Figure 16. Smooth with Newton preprocessor versus PATH 2.7 for obstacle problems.
the secondpoint, we obtained a solution in 36 iterations and 3.70 secondsand from the
third point, we obtained a solution in 49 iterations and 7.01 seconds.PATH 2.7 solved the
problem 7 times out of 10 from the first starting point, 6 times out of 10 from the second
starting point, and 5 times out of 10 from the third starting point. The average times of
the successfulPATH runs were 2.59, 3.94 and 3.21 secondsfor the first, secondand third
starting points respectively. We note that when the artificially imposedlower boundson x1
to X26were removed, PATH failed from all starting points and SMOOTH failed from the
first and secondstarting points. However from the third starting point, SMOOTH generated
a solution with somesmall negative componentswhich, when setto 1.Oe-9,gave a solution
accurate to 1.79539e-7.
Summing up the numerical experiments with PATH and SMOOTH, we believe that
comparisonsbetween the two methodswithout a Newton preprocessoris more indicative
of their relative effectiveness. With the Newton preprocessor,a lot of the work for the larger
problemsis performed by the Newton preprocessorandhencethe nearly equalperformance
of the two methodson theseproblems.
7. Conclusion
Appendix 1
Lemma 7.1.
(i) Let’t (x) = xemX, then
(ii) Let
x2
g(x) = (1 + J)eX - 1’ 6 L 0, - lo&l+ S) 6 [a, bl
(iii) Let h(x) be de@ed in Lemma 3.4, - log(1 + 6) 6 [u, b], f/Zen
Proof:
1+6
(1 + 6)e” - 1 = yxeX.
134 CHEN AND MANGASARIAN
(iii) Since
Therefore, by (ii),
2 2
5 max h(u),h(D),g(u),g(b),~t(~),- -
i 1+6 t(b)y (1 +qe I * cl
Proof of Lemma 3.4:
we have
2
max h(x) 5 max 61og*S, -& log& - .
xG[O,-logs] i (1 + 6)e I
It is easy to get
2s
maxSlog*S5:, -1ogs 5 2.
6dO.11 s~~1-1+6
26
max h(x) I max Slog2 S, -- ‘) jmax{Slog2S,f].
xG[-log&O] i 1+6 log” (1 + J)e
(iv) If 6 = 0,
For any c > 0, we have 0 = - log(1 + S) $ [c, t-co). And there exists EO> 0 such
that
max h(x) 5 2
XE[E,DO) e
max h(x) I z.
X~[O.co)
Appendix 2
Here is the actual implementation of the smoothAlgorithm 4.1. In the following algorithm
(Xk) t”k, uk) is simply denoted by yk.
Algorithm 7.1 (Smooth Algorithm for MCP). Input tolerance E = l.Oe - 6, and initial
guessx0 E R”
dk = --VR(yd-‘R(Yd
In order to avoid nonsingularity, for Cases 24 of (22), if Vi Ri(yk) < l.Oe - 9, let
ViRi(yk) = l.Oe - 9.
136 CHEN AND MANGASARIAN
where 6 = 0.75.
(5) Parameter update. !?fdyk+l) > ak9 set
ak+l = dYk+lh
ak+l = kk.
Letk=k+l,gotostep@).
For someof the test problems, somefunctions are not well defined outside the feasible
region. In such cases,the line searchstep (5) may fail. If this occurs, we will try to push
the next point inside the feasible region by setting the smoothparametera! to a very small
value, such as l.Oe - 10.
Acknowledgment
We are grateful to Michael C. Ferris and Steven P. Dirkse for letting us use their CAMS
interface routines, their PATH solver for comparison purposesand their Newton prepro-
cessorthat we incorporated into a later version of SMOOTH. We are also grateful to Jong-
Shi Pang for providing us with the generalized von Thiinen model and an initial starting
point.
References
1. D.P Bertsekas, “Projected Newton methods for optimization problems with simple constraints,” SIAM Journal
on Control and Optimization, vol. 20, pp. 221-246.1982.
2. B. Chen and PT. Harker, “A non-interior-point continuation method for linear complementarity problems,’
SIAM Journal on Matrix Analysis and Applications, vol. 14, pp. 1168-l 190, 1993.
3. C. Chen and O.L. Mangasarian, “Smoothing methods for convex inequalities and linear complementarity
problems,” Technical Report 1191, Computer Sciences Department, University of Wisconsin, Madison, Wis-
consin, 53706, November 1993. (To appear in Mathematical Programming, vol. 71, 1995.)
4. R.W. Cottle, E Giannessi, and J.-L. Lions, Variational Inequalities and Complementarity Problems, John
Wiley & Sons: New York, 1980.
5. J.E. Dennis and R.B. Schnabel, Numerical Methods for Unconstrained Optimization and Nonlinear Equations,
Prentice-Hall: Englewood Cliffs, N.J., 1983.
6. S.P. Dirkse and MC. Ferris, “The path solver: A non-monotone stabilization scheme for mixed complemen-
tarity problems,” Optimization Methods and Software, vol. 5, pp. 123-156, 1995.
NONLINEAR AND MIXED COMPLEMENTARITY PROBLEMS 137
7. S.P. Dirkse and M.C. Ferris, “MCPLIB: A collection of nonlinear mixed complementarity problems,” Opti-
mization Methods and Software, vol. 5, pp. 319-345, 1995.
8. S.P. Dirkse, M.C. Ferris, P.V. Preckel, and T. Rutherford, “The GAMS callable program library for variational
and complementarity solvers,” Manuscript, University of Wisconsin, 1993.
9. M.S. Gowda, “On the extended linear complementarity problem,” Technical Report, Department of Mathe-
matics & Statistics, University of Maryland Baltimore Country, Baltimore, Maryland, 1994, Mathematical
Programming, to appear.
10. P.T. Harker and J.-S. Pang, “Finite-dimensional variational inequality and nonlinear complementarty prob-
lems: A survey of theory, algorithms and applications,” Mathematical Programming, vol. 48, pp. 161-220,
1990.
11. P.T. Harker and B. Xiao, “Newton’s method for the nonlinear complementarity problem: A B-differentiable
equation approach,” Mathematical Programming, vol. 48, pp. 339-357, 1990.
12. J. Hertz, A. Krogh, and R.G. Palmer, Introduction to the Theory of Neural Computation, Addison-Wesley:
Redwood City, California, 199 1.
13. J.-B. Hiriart-Urruty and C. Lemarechal, Convex Analysis and Minimization Algorithms I, Springer-Verlag:
Berlin, 1993.
14. N.H. Josephy, “Newton’s method for generalized equations,” Technical Summary Report 1965, Mathematics
Research Center, University of Wisconsin, Madison, Wisconsin, 1979.
15. C. Kanzow, ‘Some tools allowing interior-point methods to become noninterior,” Technical Report, Institute
of Applied Mathematics, University of Hamburg, Germany, 1994.
16. M. Kojima and N. Megiddo, “The relation betweeh the path of centers and smale’s regularization of the linear
programming problem,” Linear Algebra and Its Applications, vol. 152, pp. 135-139, 1991.
17. M. Kojima, S. Mizuno, and A. Yoshise, “A polynomial-time algorithms for a class of linear complementarity
problems,” Mathematical Programming, vol. 44, pp. l-27, 1989.
18. J. Kreimer and R.Y. Rubinstein, “Nondifferentiable optimization via smooth approximation: General analyt-
ical approach,” Annals of Operations Research, vol. 39, pp. 97-l 19, 1992.
19. K. Madsen and H.B. Nielsen, “A finite smoothing algorithm for linear II estimation,” SIAM on Optimization,
vol. 3, no. 2, pp. 223-235, May 1993.
20. O.L. Mangasarian, “Mathematical programming in neural networks,” ORSA Journal on Computing, vol. 5,
no. 4, pp. 349-360, 1993.
2 1. O.L. Mangasarian and J.-S. Pang, “The extended linear complementarity problems,” SIAM Journal on Matrix
Analysis and Applications, vol. 16, pp. 359-368, 1995.
22. Jorge J. More, “Global methods for nonlinear complementarity problems,” Technical Report Preprint MCS-
P429-0494, Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, Illinois,
1994.
23. B.A. Murtagh and M.A. Saunders, “MINOS 5.0 user’s guide:’ Technical Report SOL 83.20, Stanford Uni-
versity, December 1983.
24. K.G. Murty, Linear Complementarity, Linear and Nonlinear Programming, Helderman-Verlag: Berlin, 1988.
25. J.M. Ortega, Numerical Analysis, A Second Course, Academic Press, 1972.
26. J.-S. Pang, “Newton’s method for B-differentiable equations,” Mathematics of Operations Research, vol. 15,
pp. 331-341.1990.
27. J.-S. Pang, “A B-differentiable equation based, globally and locally quadratically convergent algorithm for
nonlinear programs, complementarity and variational inequality problems,” Mathematical Programming,
vol. 51, pp. 101-131, 1991.
28. J.-S. Pang and S.A. Gabriel, “NE/SQP: A robust algorithm for the nonlinear complementarity problem,”
Mathematical Programming, vol. 60, pp. 295-337, 1993.
29. M.C. Pinar and S.A. Zenios “On smoothing exact penalty functions for convex constrained optimization,”
SIAM Journal on Optimization, vol. 4, pp. 486511, 1994.
30. D. Ralph, “Global convergence of damped Newton’s method for nonsmooth equations via the path search,”
Mathematics of Operations Research, vol. 19, pp. 352-389, 1994.
3 1. J. Ren, “Computable error bounds in mathematical programming,” Technical Report 1173, Computer Sciences
Department, University of Wisconsin, Madison, Wisconsin 53706, August 1993.
138 CHEN AND MANGASARIAN
32. SM. Robinson, “Generalized equations and their solution: Part I: Basic theory,” Mathematical Programming
Study, vol. 10, pp. 128-140, 1979.
33. SM. Robinson, “Strongly regular generalized equations,” Mathematics of Operations Research, vol. 5, pp. 43-
62, 1980.
34. S.M. Robinson, “Newton’s method for a class of nonsmooth functions,” Working Paper, Department of
Industrial Engineering, University of Wisconsin, Madison, Wisconsin, 1988.
35. T.F. Rutherford, “MILES: A mixed inequality and nonlinear equation solver,” Working Paper, Department of
Economics, University of Colorado, Boulder, 1993.
36. S. Smale, “Algorithms for solving equations,” In Proceedings of the International Congress of Mathematicians,
pp. 172-195, Ameri. Math. Sot., Providence, 1987.
37. B. Xiao, “Global Newtons methods for nonlinear programs and variational inequalities,” Technical Report,
Ph.D. Thesis, Department of Decision Sciences, The Wharton School, University of Pennsylvania, Philadel-
phia, PA, 1990.
38. Y. Ye, “A fully polynomial-time approximation algorithm for computing a stationary point of the general
linear complementarity problem,” Mathematics of Operations Research, vol. 18, pp. 334-345, 1993.
39. I. Zang, “A smoothing-out technique for mitt-max optimization,” Mathematical Programming, vol. 19, pp.
61-77, 1980.