Modified Limited Memory BFGS Method With Nonmonoto
Modified Limited Memory BFGS Method With Nonmonoto
net/publication/228974276
Modified limited memory BFGS method with nonmonotone line search for
unconstrained optimization
CITATIONS READS
35 223
3 authors, including:
All content following this page was uploaded by Gonglin Yuan on 23 May 2014.
1. Introduction
Consider the following unconstrained optimization problem
(1.1) min f (x),
x∈Rn
c
°2010 The Korean Mathematical Society
767
768 GONGLIN YUAN, ZENGXIN WEI, AND YANLIN WU
which is the dual form of the DF P update formula in the sense that Hk ↔ Bk ,
Hk+1 ↔ Bk+1 , and sk ↔ yk . It has been shown that the BFGS method is
very efficient for solving unconstrained optimization problems (1.1) [11, 14, 47].
For convex minimization problems, the BFGS method are global convergence
if the exact line search or some special inexact line search is used [1, 2, 4,
12, 16, 32, 33, 38] and its local convergence has been well established [9, 10,
17]. For general function f, Dai [5] have constructed an example to show that
the standard BFGS method may fail for non-convex functions with inexact
line search, Mascarenhas [29] showed that the nonconvergence of the standard
BFGS method even with exact line search.
In order to obtain a global convergence of BFGS method without convexity
assumption on the objective function, Li and Fukushima [22, 23] made a slight
modification to the standard BFGS method. Now we state their works as
follows:
(i) A new quasi-Newton equation [22] with following form
Bk+1 sk = yk1∗ ,
yT s
where yk1∗ = yk + (max{0, − kskk kk2 } + φ(kgk k))sk , and function φ : R → R
satisfies: (a) φ(t) > 0 for all t > 0; (b) φ(t) = 0 if and only if t = 0; (c) φ(t) is
bounded if t is in a bounded set.
(ii) A modified BFGS update formula [23] with following form
( T
B s sT B y 1∗ y 1∗ sT y 1∗
Bk − skT kBkksk k + yk1∗ Tks , if ksk k kk2 ≥ φ(kgk k)
(2.5) Bk+1 = k k k
Bk , otherwise.
Then it is not difficult to see that sTk yk1∗ > 0 always holds, which can ensure
that the update matrix Bk+1 inherits the positive definiteness of Bk (see [14]).
The global convergence and the superlinear convergence of these two methods
for nonconvex have been established under appropriate conditions (see [22, 23]
in detail).
In order to get a better approximation of the objective function Hessian
matrix, Wei, Li, and Qi [41] and Zhang, Deng, and Chen [51] proposed modified
quasi-Newtion equations which are given as follows. (i) The equation of Wei,
Li, and Qi [41]:
(2.6) Bk+1 sk = yk2∗ = yk + Ak sk ,
where
2[f (xk ) − f (xk+1 )] + [g(xk+1 ) + g(xk )]T sk
(2.7) Ak = .
ksk k2
They replaced all the yk in (2.3), and obtained the following modified BFGS-
type update formula
T
Bk sk sTk Bk yk2∗ yk2∗
(2.8) Bk+1 = Bk − + .
sTk Bk sk T
yk2∗ sk
MODIFIED LIMITED MEMORY BFGS METHOD 771
Note that this quasi-Newton equation (2.6) contains both gradient and function
value information at the current and the previous step, one may argue that the
resulting methods will really outperform than the original method. In fact, the
practical computation shows that this method is better than the normal BFGS
method (see [41, 42] for detail) for some given problems [30]. Furthermore,
some theoretical advantages of the new quasi-Newton equation (2.6) can be
seen from the following two theorems.
Theorem 2.1 ([42, Lemma 3.1]). Considering the quasi-Newton equation (2.6).
Then we have for all k ≥ 1
1
f (xk ) = f (xk+1 ) + g(xk+1 )T (xk − xk+1 ) + (xk − xk+1 )T Bk+1 (xk − xk+1 ).
2
Theorem 2.2 ([24, Theorem 3.1]). Assume that the function f (x) is suffi-
ciently smooth and ksk k is sufficiently small. Then we have
1
(2.9) sTk Gk+1 sk − sTk yk2∗ − sTk (Tk+1 sk )sk = O(ksk k4 )
3
and
1
(2.10) sTk Gk+1 sk − sTk yk − sTk (Tk+1 sk )sk = O(ksk k4 ),
2
where Gk+1 denotes the Hessian matrix of f at xk+1 , Tk+1 is the tensor of f
at xk+1 , and
n
X ∂ 3 f (xk+1 ) i j l
sTk (Tk+1 sk )sk = s s s .
∂xi ∂xj ∂xl k k k
i,j,l=1
Theorem 2.3 ([51, Theorem 3.3]). Assume that the function f (x) is suffi-
ciently smooth and ksk k is sufficiently small. Then we have
(2.14) sTk (Gk+1 sk − yk3∗ ) = O(ksk k4 )
and
(2.15) sTk (Gk+1 sk − yk ) = O(ksk k3 ).
It is not difficult to deduce that sTk yk2∗ > 0 holds for the uniformly convex
function f (or see [42]). We all know that the condition sTk yk2∗ > 0 can ensure
that the update matrix Bk+1 from (2.8) inherits the positive definiteness of Bk .
Similarly, in order to get the positive definiteness of Bk in (2.13) for each k,
we give a modified BFGS update of (2.13), i.e., the modified update formula is
defined by
T
Bk sk sTk Bk y 4∗ y 4∗
(2.16) Bk+1 = Bk − T
+ k Tk ,
sk Bk sk yk4∗ sk
1
where yk4∗ = yk + A∗k sk , A∗k = 3 max{Āk , 0}. Then the corresponding quasi-
Newton equation is
(2.17) Bk+1 sk = yk4∗ .
From the definition of yk4∗ , we can obtain sTk yk4∗ > 0 if the objective function f
is uniformly convex function (see Lemma 4.1).
There are other modified formulas that can be seen from [7, 34] in detail,
here we do not present anymore.
2.2. One nonmonotone line search
Normally the steplength αk is generated by the following weak Wolfe-Powell
(WWP): Find a steplength αk such that
(2.18) f (xk + αk dk ) ≤ f (xk ) + σ1 αk gkT dk ,
objective function. Numerical results show that this method is more compet-
itive than the normal BFGS method with WWP line search. Recently, the
superlinear convergence of the new nonmonotone BFGS algorithm for convex
function was proved by Yuan and Wei [49].
4. Convergence analysis
This section is devoted to show that Algorithm 2 is convergent on twice
continuously differentiable and uniformly convex function. In order to establish
global convergence for Algorithm 2, we need the following assumptions.
Assumption A. (i) The level set Ω = {x | f (x) ≤ f (x0 )} is bounded.
(ii) The function f is twice continuously differentiable on Ω.
(iii) The function f is uniformly convex, i.e., there exist positive constants
m and M such that
(4.1) mkdk2 ≤ dT G(x)d ≤ M kdk2
holds for all x ∈ Ω and d ∈ Rn , where G(x) = ∇2 f (x). These assumptions are
the same as those in [42, 51].
It is obvious that Assumption A implies that there exists a constant M∗ > 0
such that
kG(x)k ≤ M∗ , x ∈ Ω.
Assumption A (ii) implies that there exists a constant L ≥ 0 satisfying
(4.2) kg(x) − g(y)k ≤ Lkx − yk, x, y ∈ Ω.
Lemma 4.1. Let Assumption A hold. Then there exists a positive number M1
such that
kyk4∗ k2
≤ M1 , k = 0, 1, 2, . . . .
sTk yk4∗
Proof. Following the definition of yk4∗ and the Taylor’s formula, we get
sTk yk4∗ = sTk yk + sTk A∗k sk
T
= max{2[fk − fk+1 ] + 2gk+1 sk , sTk yk }
776 GONGLIN YUAN, ZENGXIN WEI, AND YANLIN WU
T 1
= max{2[−gk+1 sk + sTk G(xk + θ(xk+1 − xk )sk ] + 2gk+1
T
sk ,
2
sTk G(xk + θ1 (xk+1 − xk )sk }
= max{sTk G(xk + θ(xk+1 − xk )sk , sTk G(xk + θ1 (xk+1 − xk )sk },
where θ, θ1 ∈ (0, 1). Combining with Assumption A(iii), it is easy to obtain
(4.3) mksk k2 ≤ sTk yk4∗ ≤ M ksk k.
By the definition of yk4∗ and the Taylor’s formula again, we have
° ½ ¾ °
° 2[f (xk ) − f (xk+1 )] + [g(xk+1 ) + g(xk )]T sk °
4∗ °
kyk k = °yk + max , 0 sk °
ksk k 2 °
½ T
¾
| 2[f (xk ) − f (xk+1 )] + [g(xk+1 ) + g(xk )] sk |
≤ max kyk k + , kyk k
ksk k
| sTk G(xk + θ(xk+1 − xk )sk |
≤ 2kyk k +
ksk k2
≤ 2Lksk k + M ksk k
(4.4) = (2L + M )ksk k,
where θ ∈ (0, 1), the third inequality follows (4.1) and (4.2). By (4.3) and (4.4),
we get
kyk4∗ k2 (2L + M )2 ksk k2 (2L + M )2
≤ = = M1 .
sTk yk4∗ mksk k2 m
The proof is complete. ¤
T
yk4∗ sk
= det(Bk ) ,
sTk Bk sk
where the third equality follows from the formula (see, e.g., [9, Lemma 7.6])
det(I + u1 uT2 + u3 uT4 ) = (1 + uT1 u2 )(1 + uT3 u4 ) − (uT1 u4 )(uT2 u3 ).
Therefore, there is also a simple expression for the determinant of (3.28)
k
Y sTl yl4∗
det(Bk+1 ) = det(Bkk−m+1
e
) .
sTl Bl sl
l=k−m+1
e
Proof. Assume that lim inf k kgk k > 0, i.e., there exists a constant c2 > 0 such
that
(4.10) kgk k ≥ c2 , k = 0, 1, 2, . . . .
From Assumption A(iii) and Taylor’s formula, we have
(4.11) mksk k2 ≤ sTk G(xk + θ1 sk )sk = sTk yk ≤ M ksk k2 ,
combining with (4.3), we get
m T M T
(4.12) s yk ≤ sTk yk4∗ ≤ s yk .
M k m k
Taking the trace operation in both sides of (3.28), we get
k
X k
X
kBl sl k2 kyl4∗ k2
(4.13) Tr(Bk+1 ) = Tr(Bkk−m+1
e
)− + ,
sTl Bl sl slT yl4∗
l=k−m+1
e l=k−m+1
e
where Tr(Bk ) denotes the trace of Bk . Repeating this trace operation, we have
k
X k
X
kBl sl k2 kyl4∗ k2
Tr(Bk+1 ) = Tr(Bkk−m+1
e
) − +
sTl Bl sl sTl yl4∗
l=k−m+1
e l=k−m+1
e
= ···
k
X k
X
kBl sl k2 ky 4∗ k2
l
(4.14) = Tr(B0 ) − + .
l=0
sTl Bl sl l=0
sTl yl4∗
and
(4.17) Tr(Bk+1 ) ≤ Tr(B0 ) + (k + 1)M1 .
MODIFIED LIMITED MEMORY BFGS METHOD 779
−gjT dj
cos θj = .
kgj kkdj k
780 GONGLIN YUAN, ZENGXIN WEI, AND YANLIN WU
we have
k+1
Y · ¸k+1
(1 − ε2 )c3 c22
(4.23) max{ksj k, ksj k1−p }kgj k cos θj ≥ .
j=0
Tr(B0 ) + M1
According to Lemma 4.4 and Assumption A we know that there exists a con-
stant M20 > 0 such that
(4.24) ksk k = kxk+1 − xk k ≤ kxk+1 k + kxk k ≤ 2M20 .
Combining (4.23) and (4.24), and noting that kgj k cos θj = ηj , we get for all
k≥0
Yk · ¸k+1
(1 − ε2 )c3 c22
ηj ≥ 0 , 1, (2M 0 )1−p } = ²k+1
0 .
j=0
(Tr(B 0 ) + M1 ) max{2M 2 2
∞
X 2−p
2 1−p
min min{η(n(k)+1)q+n(k)−j , η(n(k)+1)q+n(k)−j } < +∞.
0≤j≤n(k)
q=1
k
1 Y n(k)
≤ [ηp(q) b3 ]
η0 q=0
k
1 kn(k) Y
(4.30) = b3 ηp(q) .
η0 q=0
Then we have
k
" #k " #k
Y n(k)
n(k)+1
²0
n(k)+1
²0 n(k)
ηp(q) ≥ η0 ² 0 n(k)
≥ n(k)
min{1, η0 ²0 } .
q=0 b3 b3
Using Lemma 4.6 we have
(4.31) lim sup ηp(q) > 0,
q→∞
5. Numerical results
In this section, we report some numerical results on the problems [30] with
initial points. All codes were written in MATLAB 7.0 and run on PC with
2.60GHz CPU processor and 256MB memory and Windows XP operation sys-
tem. The parameters are chosen as: σ1 = 0.1, σ2 = 0.9, ε = 10−5 , ε1 =
0.1, ε2 = 0.01, p = 5, H = 8, m1 = 5, and the initial matrix B0 = I is the
unit matrix.
The following Himmeblau stop rule is used [47]:
If | f (xk ) |> e1 , let stop1 = |f (xk|f)−f (xk+1 )|
(xk )| ; Otherwise, let stop1 =| f (xk ) −
f (xk+1 ) | .
For each problem, if kg(x)k < ε or stop1 < e2 was satisfied, the program
will be stopped, where e1 = e2 = 10−5 .
Since the line search cannot always ensure the descent condition dTk gk < 0,
uphill search direction may occur in the numerical experiments. In this case,
the line search rule maybe fails. In order to avoid this case, the stepsize αk will
be accepted if the searching number is more than twenty five in line search.
We also stop the program if the iteration number is more than one thousand,
and the corresponding method is considered to be failed.
In Figure 1-3, “BFGS-WP-Ak1” and “BFGS-WP-Ak2” stand for the mod-
ified BFGS formula (2.8) with WWP rule and the modified BFGS formula
MODIFIED LIMITED MEMORY BFGS METHOD 783
Dolan and Moré [13] gave a new tool to analyze the efficiency of algorithms.
They introduced the notion of a performance profile as a means to evaluate and
compare the performance of the set of solvers S on a test set P. Assuming that
there exist ns solvers and np problems, for each problem p and solver s, they
defined tp,s = computing time (the number of function evaluations or others)
required to solve problem p by solver s.
Requiring a baseline for comparisons, they compared the performance on
problem p by solver s with the best performance by any solver on this problem;
that is, using the performance ratio
tp,s
rp,s = .
min{tp,s : s ∈ S}
Suppose that a parameter rM ≥ rp,s for all p, s is chosen, and rp,s = rM if and
only if solver s does not solve problem p.
The performance of solver s on any given problem might be of interest, but
we would like to obtain an overall assessment of the performance of the solver,
then they defined
1
ρs (t) = size{p ∈ P : rp,s ≤ t},
np
thus ρs (t) was the probability for solver s ∈ S that a performance ratio rp,s
was within a factor t ∈ R of the best possible ration. Then function ρs was the
(cumulative) distribution function for the performance ratio. The performance
profile ρs : R 7→ [0, 1] for a solver was a nondecreasing, piecewise constant
function, continuous from the right at each breakpoint. The value of ρs (1) was
the probability that the solver would win over the rest of the solvers.
According to the above rules, we know that one solver whose performance
profile plot is on top right will win over the rest of the solvers.
Figures 1, 2, and 3 show that the performance of these methods is relative
to N I, N F G, and T ime, respectively, where N I denotes the total number of
iterations, N F G denotes the total number of the function evaluations and the
gradient evaluations where N T = N F + 5N G (see [6, 18]), and T ime denotes
the cpu time that these methods spent. From these three figures it is clear
that the L-BFGS-A11 method has the most wins (has the highest probability
of being the optimal solver).
Figure 1 shows that L-BFGS-A11 and L-BFGS-A1 outperform BFGS-WP-
Ak1 and BFGS-WP-Ak2 about 5% and 8% test problems, respectively. The
L-BFGS-A11 method is predominant among the other three methods for t ≤ 5.
784 GONGLIN YUAN, ZENGXIN WEI, AND YANLIN WU
0.9
0.8
0.7
0.6
Pp:r(p,s)<=t
0.5
0.4
0.3
.. L−BFGS−A1
0.2 − − BFGS−WP−Ak2
−.L−BFGS−A11
0.1 − BFGS−WP−Ak1
0
5 10 15 20 25 30
t
Moreover, the L-BFGS-A11 and L-BFGS-A1 and solve 100%, and the BFGS-
WP-Ak1 and BFGS-WP-Ak2 method solve about 95% and 92% of the test
problems successfully, respectively.
Figure 2 shows that L-BFGS-A11 and L-BFGS-A1 are superior to BFGS-
WP-Ak1 and BFGS-WP-Ak2 about 15% test problems. The L-BFGS-A11
and L-BFGS-A1 method can solve 100% of the test problems successfully at
t ≈ 4.2 and t ≈ 7.2, respectively. The BFGS-WP-Ak1 and BFGS-WP-Ak2
method solve about 85% of the test problems successfully.
Figure 3 shows that L-BFGS-A11 outperforms the other three methods. The
L-BFGS-A1 method and the BFGS-WP-Ak1 method solve about 95% and 91%
of the test problems, respectively, and the BFGS-WP-Ak2 solves about 88% of
the test problems successfully.
In summary, the presented numerical results reveal that Algorithm 1 and
Algorithm 11, compared with other two methods with WWP line search and
BFGS update, have potential advantages for these problems.
6. Conclusion
This paper gives two modified L-BFGS method with one nonmonotone line
search technique for solving unconstrained optimization, which include the
function value at the current and next iterative point values. The global con-
vergence for the uniformly convex functions are established. The numerical
MODIFIED LIMITED MEMORY BFGS METHOD 785
0.9
0.8
0.7
0.6
Pp:r(p,s)<=t
0.5
0.4
0.3
.. L−BFGS−A1
0.2 − − BFGS−WP−Ak2
−.L−BFGS−A11
0.1 − BFGS−WP−Ak1
0
2 4 6 8 10 12 14 16
t
0.9
0.8
0.7
0.6
Pp:r(p,s)<=t
0.5
0.4
0.3
.. L−BFGS−A1
0.2 − − BFGS−WP−Ak2
−.L−BFGS−A11
0.1 − BFGS−WP−Ak1
0
2 4 6 8 10 12 14 16
t
786 GONGLIN YUAN, ZENGXIN WEI, AND YANLIN WU
results show that the given methods are competitive to the other standard
BFGS methods for the test problems.
For further research, we should study the performance of the new algorithm
at different stop rules and different testing environment (such as [15]). More-
over, more numerical experiments for large practical problems should be done
in the future.
References
[1] C. G. Broyden, J. E. Dennis Jr, and J. J. Moré, On the local and superlinear convergence
of quasi-Newton methods, J. Inst. Math. Appl. 12 (1973), 223–245.
[2] R. H. Byrd and J. Nocedal, A tool for the analysis of quasi-Newton methods with appli-
cation to unconstrained minimization, SIAM J. Numer. Anal. 26 (1989), no. 3, 727–739.
[3] R. H. Byrd, J. Nocedal, and R. B. Schnabel, Representations of quasi-Newton matrices
and their use in limited memory methods, Math. Programming 63 (1994), no. 2, Ser.
A, 129–156.
[4] R. Byrd, J. Nocedal, and Y. Yuan, Global convergence of a class of quasi-Newton meth-
ods on convex problems, SIAM J. Numer. Anal. 24 (1987), no. 5, 1171–1190.
[5] Y. Dai, Convergence properties of the BFGS algorithm, SIAM J. Optim. 13 (2002), no.
3, 693–701.
[6] Y. Dai and Q. Ni, Testing different conjugate gradient methods for large-scale uncon-
strained optimization, J. Comput. Math. 21 (2003), no. 3, 311–320.
[7] W. C. Davidon, Variable metric methods for minimization, A. E. C. Research and
Development Report ANL-599, 1959.
[8] R. Dembo and T. Steihaug, Truncated Newton algorithms for large-scale unconstrained
optimization, Math. Programming 26 (1983), no. 2, 190–212.
[9] J. E. Dennis Jr. and J. J. Moré, Quasi-Newton methods, motivation and theory, SIAM
Rev. 19 (1977), no. 1, 46–89.
[10] , A characterization of superlinear convergence and its application to quasi-
Newton methods, Math. Comp. 28 (1974), 549–560.
[11] J. E. Dennis Jr. and R. B. Schnabel, Numerical Methods for Unconstrained Optimization
and Nonlinear Equations, Prentice Hall Series in Computational Mathematics. Prentice
Hall, Inc., Englewood Cliffs, NJ, 1983.
[12] L. C. W. Dixon, Variable metric algorithms: necessary and sufficient conditions for
identical behavior of nonquadratic functions, J. Optimization Theory Appl. 10 (1972),
34–40.
[13] E. D. Dolan and J. J. Moré, Benchmarking optimization software with performance
profiles, Math. Program. 91 (2002), no. 2, Ser. A, 201–213.
[14] R. Fletcher, Practical Methods of Optimization, Second edition. A Wiley-Interscience
Publication. John Wiley & Sons, Ltd., Chichester, 1987.
[15] N. I. M. Gould, D. Orban, and Ph. L. Toint, CUTEr (and SifDec), a constrained and un-
constrained testing environment, revisite, ACM Transactions on Mathematical Software
29 (2003), 373–394.
[16] A. Griewank, The global convergence of partitioned BFGS on problems with convex
decompositions and Lipschitzian gradients, Math. Programming 50 (1991), no. 2, (Ser.
A), 141–175.
[17] A. Griewank and Ph. L. Toint, Local convergence analysis for partitioned quasi-Newton
updates, Numer. Math. 39 (1982), no. 3, 429–448.
[18] L. Grippo, F. Lamparillo, and S. Lucidi, A nonmonotone line search technique for
Newton’s method, SIAM J. Numer. Anal. 23 (1986), no. 4, 707–716.
MODIFIED LIMITED MEMORY BFGS METHOD 787
[19] , A truncated Newton method with nonmonotone line search for unconstrained
optimization, J. Optim. Theory Appl. 60 (1989), no. 3, 401–419.
[20] , A class of nonmonotone stabilization methods in unconstrained optimization,
Numer. Math. 59 (1991), no. 8, 779–805.
[21] J. Y. Han and G. H. Liu, Global convergence analysis of a new nonmonotone BFGS
algorithm on convex objective functions, Comput. Optim. Appl. 7 (1997), no. 3, 277–289.
[22] D. Li and M. Fukushima, A modified BFGS method and its global convergence in non-
convex minimization, J. Comput. Appl. Math. 129 (2001), no. 1-2, 15–35.
[23] , On the global convergence of the BFGS method for nonconvex unconstrained
optimization problems, SIAM J. Optim. 11 (2001), no. 4, 1054–1064.
[24] G. Li, C. Tang, and Z. Wei, New conjugacy condition and related new conjugate gradient
methods for unconstrained optimization, J. Comput. Appl. Math. 202 (2007), no. 2,
523–539.
[25] G. H. Liu and J. Y. Han, Notes on the general form of stepsize selection, OR and
Decision Making I (1992), 619–624.
[26] , Global convergence Analysis of the variable metric algorithm with a generalized
Wolf linesearch, Technical Report, Institute of Applied Mathematics, Academia Sinica,
Beijing, China, no. 029, 1993.
[27] G. H. Liu, J. Y. Han, and D. F. Sun, Global convergence of the BFGS algorithm with
nonmonotone linesearch, Optimization 34 (1995), no. 2, 147–159.
[28] G. H. Liu and J. M. Peng, The convergence properties of a nonmonotonic algorithm, J.
Comput. Math. 1 (1992), 65–71.
[29] W. F. Mascarenhas, The BFGS method with exact line searches fails for non-convex
objective functions, Math. Program. 99 (2004), no. 1, Ser. A, 49–61.
[30] J. J. Moré, B. S. Garbow, and K. E. Hillstrome, Testing unconstrained optimization
software, ACM Trans. Math. Software 7 (1981), no. 1, 17–41.
[31] S. G. Nash, A survey of truncated-Newton methods, Numerical analysis 2000, Vol. IV,
Optimization and nonlinear equations. J. Comput. Appl. Math. 124 (2000), no. 1-2,
45–59.
[32] M. J. D. Powell, On the convergence of the variable metric algorithm, J. Inst. Math.
Appl. 7 (1971), 21–36.
[33] , Some global convergence properties of a variable metric algorithm for min-
imization without exact line searches, Nonlinear programming (Proc. Sympos., New
York, 1975), pp. 53–72. SIAM-AMS Proc., Vol. IX, Amer. Math. Soc., Providence, R.
I., 1976.
[34] , A new algorithm for unconstrained optimization, 1970 Nonlinear Programming
(Proc. Sympos., Univ. of Wisconsin, Madison, Wis., 1970) pp. 31–65 Academic Press,
New York.
[35] M. Raydan, The Barzilai and Borwein gradient method for the large scale unconstrained
minimization problem, SIAM J. Optim. 7 (1997), no. 1, 26–33.
[36] J. Schropp, A note on minimization problems and multistep methods, Numer. Math. 78
(1997), no. 1, 87–101.
[37] , One-step and multistep procedures for constrained minimization problems, IMA
J. Numer. Anal. 20 (2000), no. 1, 135–152.
[38] Ph. L. Toint, Global convergence of the partitioned BFGS algorithm for convex partially
separable optimization, Math. Programming 36 (1986), no. 3, 290–306.
[39] D. J. van Wyk, Differential optimization techniques, Appl. Math. Modelling 8 (1984),
no. 6, 419–424.
[40] M. N. Vrahatis, G. S. Androulakis, J. N. Lambrinos, and G. D. Magolas, A class of gra-
dient unconstrained minimization algorithms with adaptive stepsize, J. Comput. Appl.
Math. 114 (2000), no. 2, 367–386.
View publication stats
[41] Z. Wei, G. Li, and L. Qi, New quasi-Newton methods for unconstrained optimization
problems, Appl. Math. Comput. 175 (2006), no. 2, 1156–1188.
[42] Z. Wei, G. Yu, G. Yuan, and Z. Lian, The superlinear convergence of a modified BFGS-
type method for unconstrained optimization, Comput. Optim. Appl. 29 (2004), no. 3,
315–332.
[43] G. L. Yuan, Modified nonlinear conjugate gradient methods with sufficient descent prop-
erty for large-scale optimization problems, Optim. Lett. 3 (2009), no. 1, 11–21.
[44] G. L. Yuan and X. W. Lu, A new line search method with trust region for unconstrained
optimization, Comm. Appl. Nonlinear Anal. 15 (2008), no. 1, 35–49.
[45] , A modified PRP conjugate gradient method, Ann. Oper. Res. 166 (2009), 73–90.
[46] G. L. Yuan, X. Lu, and Z. Wei, A conjugate gradient method with descent direction for
unconstrained optimization, J. Comput. Appl. Math. 233 (2009), no. 2, 519–530.
[47] Y. Yuan and W. Sun, Theory and Methods of Optimization, Science Press of China,
1999.
[48] G. L. Yuan and Z. X. Wei, New line search methods for unconstrained optimization, J.
Korean Statist. Soc. 38 (2009), no. 1, 29–39.
[49] , The superlinear convergence analysis of a nonmonotone BFGS algorithm on
convex objective functions, Acta Math. Sin. (Engl. Ser.) 24 (2008), no. 1, 35–42.
[50] , Convergence analysis of a modified BFGS method on convex minimizations,
Comput. Optim. Appl. doi: 10.1007/s10589-008-9219-0.
[51] J. Z. Zhang, N. Y. Deng, and L. H. Chen, New quasi-Newton equation and related
methods for unconstrained optimization, J. Optim. Theory Appl. 102 (1999), no. 1,
147–167.
Gonglin Yuan
College of Mathematics and Information Science
Guangxi University
Nanning, Guangxi, 530004, P. R. China
E-mail address: [email protected]
Zengxin Wei
College of Mathematics and Information Science
Guangxi University
Nanning, Guangxi, 530004, P. R. China
E-mail address: [email protected]
Yanlin Wu
College of Mathematics and Information Science
Guangxi University
Nanning, Guangxi, 530004, P. R. China
E-mail address: [email protected]