A direct Method for Parabolic PDE Constrained Optimization Problems (Andreas Potschka)
A direct Method for Parabolic PDE Constrained Optimization Problems (Andreas Potschka)
Edited by
Prof. Dr. Dres. h. c. Hans Georg Bock
Prof. Dr. Dr. h. c. Wolfgang Hackbusch
Prof. Mitchell Luskin
Prof. Dr. Dr. h. c. Rolf Rannacher
Andreas Potschka
ISSN 1616-2994
ISBN 978-3-658-04475-6 ISBN 978-3-658-04476-3 (eBook)
DOI 10.1007/978-3-658-04476-3
Springer Spektrum
© Springer Fachmedien Wiesbaden 2014
This work is subject to copyright. All rights are reserved by the Publisher, whether the whole
or part of the material is concerned, specifically the rights of translation, reprinting, reuse of
illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical
way, and transmission or information storage and retrieval, electronic adaptation, compu-
ter software, or by similar or dissimilar methodology now known or hereafter developed.
Exempted from this legal reservation are brief excerpts in connection with reviews or schol-
arly analysis or material supplied specifically for the purpose of being entered and executed
on a computer system, for exclusive use by the purchaser of the work. Duplication of this
publication or parts thereof is permitted only under the provisions of the Copyright Law of
the Publisher’s location, in its current version, and permission for use must always be obtained
from Springer. Permissions for use may be obtained through RightsLink at the Copyright
Clearance Center. Violations are liable to prosecution under the respective Copyright Law.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this
publication does not imply, even in the absence of a specific statement, that such names are
exempt from the relevant protective laws and regulations and therefore free for general use.
While the advice and information in this book are believed to be true and accurate at the date
of publication, neither the authors nor the editors nor the publisher can accept any legal re-
sponsibility for any errors or omissions that may be made. The publisher makes no warranty,
express or implied, with respect to the material contained herein.
Andreas Potschka
Abstract
In this thesis we develop a numerical method based on Direct Multiple Shooting
for Optimal Control Problems (OCPs) constrained by time-periodic Partial Differ-
ential Equations (PDEs). The proposed method features asymptotically optimal
scale-up of the numerical effort with the number of spatial discretization points.
It consists of a Linear Iterative Splitting Approach (LISA) within a Newton-type
iteration with globalization on the basis of natural level functions. We investigate
the LISA-Newton method in the framework of Bock’s κ-theory and develop reli-
able a-posteriori κ-estimators. Moreover we extend the inexact Newton method to
an inexact Sequential Quadratic Programming (SQP) method for inequality con-
strained problems and provide local convergence theory. In addition we develop
a classical and a two-grid Newton-Picard preconditioner for LISA and prove grid-
independent convergence of the classical variant for a model problem. Based on
numerical results we can claim that the two-grid version is even more efficient
than the classical version for typical application problems. Moreover we develop
a two-grid approximation for the Lagrangian Hessian which fits well in the two-
grid Newton-Picard framework and yields a reduction of 68 % in runtime for a
nonlinear benchmark problem compared to the use of the exact Lagrangian Hes-
sian. We show that the quality of the fine grid controls the accuracy of the solution
while the quality of the coarse grid determines the asymptotic linear convergence
rate, i.e., Bock’s κ. Based on reliable κ-estimators we facilitate automatic coarse
grid refinement to guarantee fast convergence. For the solution of the occurring
large-scale Quadratic Programming Problems (QPs) we develop a structure ex-
ploiting two-stage approach. In the first stage we exploit the Multiple Shooting
and Newton-Picard structure to reduce the large-scale QP to an equivalent QP
whose size is independent of the number of spatial discretization points. For the
second stage we develop extensions for a Parametric Active Set Method (PASM)
to achieve a reliable and efficient solver for the resulting, possibly nonconvex QP.
Furthermore we construct three illustrative, counter-intuitive toy examples which
show that convergence of a one-shot one-step optimization method is neither nec-
essary nor sufficient for the convergence of the forward problem method. For three
regularization approaches to recover convergence our analysis shows that de-facto
loss of convergence cannot be avoided with these approaches. We have further
implemented the proposed methods within a code called MUSCOP which features
VIII Abstract
automatic derivative generation for the model functions and dynamic system solu-
tions of first and second order, parallelization on the Multiple Shooting structure,
and a hybrid language programming paradigm to minimize setup and solution time
for new application problems. We demonstrate the applicability, reliability, and
efficiency of MUSCOP and thus the proposed numerical methods and techniques
on a sequence of PDE OCPs of growing difficulty ranging from linear academic
problems, over highly nonlinear academic problems of mathematical biology to
a highly nonlinear real-world chemical engineering problem in preparative chro-
matography: The Simulated Moving Bed (SMB) process.
Contents
Acknowledgments V
Abstract VII
1 Introduction 1
1.1 Results of this thesis . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Thesis overview . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
I Theoretical foundations 9
2 Problem formulation 11
2.1 Dynamical models described by Partial Differential Equations . . 11
2.2 Coupled ODEs and PDEs . . . . . . . . . . . . . . . . . . . . . . 15
2.3 The Optimal Control Problem . . . . . . . . . . . . . . . . . . . 15
II Numerical methods 31
5 Inexact Sequential Quadratic Programming 33
5.1 Newton-type methods . . . . . . . . . . . . . . . . . . . . . . . . 33
5.2 Local convergence . . . . . . . . . . . . . . . . . . . . . . . . . 36
X Contents
6 Newton-Picard preconditioners 75
6.1 The Newton-Picard method for finding periodic steady states . . . 76
6.2 Discretization of the model problem . . . . . . . . . . . . . . . . 77
6.3 Newton-Picard for optimal control problems . . . . . . . . . . . . 80
6.3.1 General considerations . . . . . . . . . . . . . . . . . . . 81
6.3.2 Simultaneous Newton-Picard iteration . . . . . . . . . . . 81
6.3.3 Convergence for classical Newton-Picard . . . . . . . . . 85
6.3.4 Numerical solution of the approximated linear system . . 90
6.3.5 Pseudocode . . . . . . . . . . . . . . . . . . . . . . . . . 91
6.3.6 Algorithmic complexity . . . . . . . . . . . . . . . . . . 95
6.4 Extension to nonlinear problems and Multiple Shooting . . . . . . 97
8 Condensing 111
Contents XI
Bibliography 201
List of acronyms
AD Algorithmic Differentiation
KKT Karush-Kuhn-Tucker
Mathematics today permeates an ever increasing part of the sciences far beyond
mathematical physics just as about 50 years ago Nobel Prize laureate Wigner has
hoped for. In particular mathematical methods for simulation and optimization of
quantitative mathematical models continue to face growing demand in disciplines
ranging from engineering, biology, economics, physics, etc. even to emerging ar-
eas of psychology or archeology (see, e.g., Sager et al. [139], Schäfer et al. [140]).
In this thesis we focus on mathematical and computational methods for the class
of Optimal Control Problems (OCPs) pioneered by Pontryagin and Bellman in the
middle of the 20th century. General mathematical optimization problems consist
of finding a solution candidate which satisfies a set of constraints and minimizes a
certain objective function. OCPs are optimization problems whose free variables
comprise states and controls from (usually infinite dimensional) function spaces
constrained to satisfy given differential equations. The differential equations de-
scribe the behavior of a dynamic system which can be controlled in a prescribed
way.
The treatment of constraints given by Partial Differential Equations (PDEs) is
one major challenge that we address in this thesis. PDEs appear when spatially
distributed phenomena need to be taken into account, e.g., when we describe the
diffusion of a substance in a liquid. Ordinary Differential Equations (ODEs),
which describe the evolution of a system in time, are not a satisfactory mathe-
matical tool for the description of spatial effects (although we shall use them to
approximate solutions of PDEs). A considerable amount of theory and practical
computational methods is available today for ODE OCPs. The presence of PDE
constraints causes additional difficulties both on the theoretical as well as on the
numerical side and is a much younger field of research especially in the aspect of
methods which heavily rely on high computing power.
OCPs are inherently infinite problems because we seek solutions in function
spaces. We can divide numerical methods for OCPs into two main classes: Direct
and indirect methods. The defining line between the two is somewhat blurry, espe-
cially when we cross borders of mathematical communities. We base our classifi-
cation here on the sequence of discretization and optimization. In indirect methods
we first derive optimality conditions in function space which we discretize after-
wards. In direct methods we discretize the problem first and then find an optimizer
of the resulting Nonlinear Programming Problem (NLP). Moreover we often end
up with an implicit characterization of the control via artificially introduced co-
state or adjoint variables in indirect methods. This is in contrast to direct methods
for which the discretized control usually occurs explicitly as one or the only re-
maining variable. Indirect methods for ODE OCPs are mostly based on Dynamic
Programming (see, e.g., Bellman [15]) or Pontryagin’s Maximum Principle (see,
e.g., Pontryagin et al. [126]). Tröltzsch [152] in his introductory textbook for PDE
OCPs also treats only indirect methods. A discussion of direct and indirect meth-
ods for PDE OCPs is given in Hinze et al. [86, Chapter 3]. In the 1980’s the
endeavor to apply numerical optimization quickly to new application areas and
new problems led to the development of direct methods for ODE OCPs, most no-
tably collocation methods (see, e.g., Bär [8], Bock [23], Biegler [18]) and Direct
Multiple Shooting (Bock [22], Plitt [125], Bock and Plitt [25]). One advantage of
direct methods is that the optimality conditions of an NLP are generic, whereas
optimality conditions of undiscretized OCPs need to be reestablished for each new
problem and often require partial a-priori knowledge of the mathematical structure
of the solution which in general is not available for many application problems. At
the crux of creating an efficient direct optimization method is structure exploita-
tion in the numerical solution of the NLP. Usually either Sequential Quadratic
Programming (SQP) or Interior Point methods are employed (see, e.g., the text-
book of Nocedal and Wright [121]). These iterative methods require the compu-
tation of derivatives of the objective function and the constraints. Derivative free
methods (see, e.g., the introductory textbook by Conn et al. [34]) are typically not
suited because of the high number of unknowns and because nonlinear constraints
can only be treated with excessive computational effort.
It is our goal in this thesis to extend Direct Multiple Shooting for ODE OCPs in
order to make it applicable and continue its success story for a class of PDE OCPs.
The first hurdle on this venture is the large problem size of the discretized OCPs.
Schäfer [141] describes in his dissertation approaches to address this difficulty by
exploitation of the special mathematical structure of the discretized OCPs. His
1.1 Results of this thesis 3
siderable importance for the reliability of the method on already badly conditioned
problems.
The analysis reveals that the quality of the fine grid controls the accuracy of
the solution while the quality of the coarse grid determines the asymptotic lin-
ear convergence rate, i.e., Bock’s κ, of the two-grid Newton-Picard LISA-Newton
method. Based on the newly established reliable a-posteriori κ-estimates we de-
velop a numerical strategy for automatic determination of when to refine the coarse
grid to guarantee fast convergence.
We further develop a structure exploiting two-stage strategy for the solution of
QP subproblems in the inexact SQP method. The first stage is an extension of the
traditional condensing step in SQP methods for Direct Multiple Shooting which
exploits the constraint for periodicity or alternatively given fixed initial values for
the PDE in addition to the Multiple Shooting matching conditions. This strat-
egy reduces the large-scale QP to an equivalent QP whose size is independent of
the spatial discretization. The reduction can be efficiently computed because it
additionally exploits the (two-grid) Newton-Picard structure in the QP constraint
and Hessian matrices. For the second stage we develop a Parametric Active Set
Method (PASM) which can also treat nonconvex QPs with indefinite Hessian ma-
trices. This capability is required because we want to treat nonconvex NLPs using
accurate approximations for Lagrangian Hessians. We propose numerical tech-
niques for improving the reliability of our PASM code which outperforms several
other popular QP codes in terms of reliability.
The Newton-Picard LISA method can also be interpreted as a one-shot one-step
approach for a linear PDE OCP. The almost optimal convergence theorem which
we prove for the considered model problem supports the conjecture that such one-
step approaches will in general yield optimization algorithms which converge as
fast as the algorithm for the forward problem, which consists of satisfying the
constraints for fixed controls. Contrary to common belief, however, we have con-
structed three small-scale, equality constrained QPs which illustrate that the con-
vergence for the forward problem method is neither sufficient nor necessary for
the convergence of the one-step optimization method. Furthermore we show that
existing one-step techniques to enforce convergence might lead to de-facto loss of
convergence with contraction factors of almost 1. These examples and results can
serve as a warning signal or guiding principle for the choice of assertions which
one might want to attempt to prove about one-step methods. It also justifies that
we prove convergence of the Newton-Picard LISA only for a model problem.
We have put substantial effort into the implementation of the proposed ideas in
a new software package called MUSCOP. Based on a hybrid programming design
principle we strive to keep the code both easy to use and easy to maintain/develop
6 1 Introduction
further at the same time. The code features parallelization on the Multiple Shoot-
ing structure and automatic generation of derivatives of first and second order of
the model functions and dynamic systems in order to reduce setup and solution
time for new application problems to a minimum.
Finally we use MUSCOP to demonstrate the applicability, reliability, and effi-
ciency of the proposed numerical methods and techniques on a sequence of PDE
OCPs of growing difficulty: Linear and nonlinear boundary control tracking prob-
lems subject to the time-periodic linear heat equation in 2D and 1D, a tracking
problem in bacterial chemotaxis which features a strong nonlinearity in the con-
vective term, and finally a real-world practical example: Optimal control of the
ModiCon variant of the SMB process.
of LISA and its connection with Bock’s κ-theory. On this basis we develop three
a-posteriori κ-estimators which are based on the LISA iterates. In addition we
propose an extension to SQP methods, prove that a first-order necessary optimal-
ity condition holds if the method converges, and further show that a second order
sufficiency condition transfers from the QP in the solution to the solution of the
NLP. Finally we present a novel extension to inexact SQP methods on the basis of
a generalized LISA for QPs.
In Chapter 6 we develop so-called Newton-Picard preconditioners for time-
periodic OCPs. We discuss a classical and a two-grid projective approach. For
the classical approach we show grid-independent convergence. We conclude the
chapter with a discussion of the application of Newton-Picard preconditioning in
a LISA-Newton method for nonlinear problems and for Multiple Shooting.
We present three counter-intuitive toy examples in Chapter 7 which show that
convergence of the forward problem method is neither sufficient nor necessary
for the convergence of a corresponding one-step one-shot optimization approach.
We furthermore analyze regularization approaches which are designed to enforce
one-step one-shot convergence and demonstrate that de-facto loss of convergence
cannot be avoided via these techniques.
In Chapter 8 we discuss condensing of the occurring large-scale QPs to equiva-
lent QPs whose size is independent of the number of spatial discretization points.
We further develop efficient numerical exploitation of the Multiple Shooting and
Newton-Picard structure. Moreover we propose a two-grid Hessian matrix ap-
proximation which fits well in the framework of the two-grid Newton-Picard pre-
conditioners. As a final remark we show scaling invariance of the Newton-Picard
LISA-Newton method for PDE OCPs.
The solution of the resulting medium-scale QPs via PASM is our subject in
Chapter 9. We identify numerical challenges in PASMs and develop strategies
to meet these challenges, in particular the techniques of drift correction and flip-
ping bounds. Furthermore we implement these strategies in a code called rpasm
and demonstrate that rpasm outperforms other popular QP solvers in terms of re-
liability on a well-known test set. We conclude the chapter with an extension to
nonconvex QPs which can arise when employing the exact Lagrangian Hessian or
the two-grid Newton-Picard Hessian approximation. The proposed PASM is also
considerably efficient because it can be efficiently hot-started.
In Chapter 10 we review numerical methods for automatic generation of deriva-
tives on the basis of Algorithmic Differentiation (AD) and Internal Numerical
Differentiation (IND). Furthermore we address issues with a monitor strategy in
implicit numerical integrators which can lead to violation of the IND principle for
8 1 Introduction
the example of a linear 1D heat equation. Then we conclude the chapter with a
short account on the numerical effort of IND.
We dedicate Chapter 11 to the design of the software package MUSCOP. Issues
we address include programming paradigms and description of the various soft-
ware components and their complex orchestration necessary for smart structure
exploitation. This concludes Part 2, numerical methods.
In Part 3 we present applications and numerical results which were generated
with MUSCOP. Linear boundary control for the periodic 2D heat equation is in the
focus of our presentation in Chapter 12. We give numerical evidence of the failure
of Euclidean instead of L2 projection in classical Newton-Picard preconditioners.
In accordance with the proof of mesh-independent convergence we give several
computational results for varying problem data and discuss why the two-grid vari-
ant is superior to the classical Newton-Picard preconditioner.
We extend the problem to nonlinear boundary control in 1D in Chapter 13 and
discuss numerical self-convergence. We can show that employing the two-grid
Hessian approximation leads to an overall reduction in computation time of 68 %.
We discuss parallelization issues and compare runtimes for different discretiza-
tions of the control in time. In all cases we give detailed information about the
runtime spent in different parts of the algorithm and show exemplarily that with
above 95 % most of the runtime is required for system simulation and IND.
In Chapter 14 we present a tracking type OCP for a (non-periodic) bacterial
chemotaxis model in 1D. The model is characterized by a highly nonlinear con-
vective term. We demonstrate the applicability of the proposed methods also to
this problem and discuss the self-convergence of the computation.
Chapter 15 is the last chapter of this thesis. In it we present the SMB process
and explain a mathematical model for chromatographic columns. We then present
numerical results for the ModiCon variant of the SMB process for real-world data.
We obtain optimal solutions with an accuracy which has not been achieved before.
This concludes Part 3 and this thesis.
Chapters 6, 7, 9, 12, and parts of Chapter 15 are based on own previously pub-
lished work. For completeness we reprint partial excerpts here with adaptions to
the unified nomenclature and structure of this thesis. We give the precise refer-
ences to the respective articles at the beginning of each of these chapters.
Part I
Theoretical foundations
2 Problem formulation
The goal of this chapter is to introduce the Optimal Control Problem (OCP) for-
mulation which serves as the point of origin for all further investigations in this
thesis. To this end we recapitulate elements of the theory of parabolic Partial Dif-
ferential Equations (PDEs) in Section 2.1 and present a system of PDEs coupled
with Ordinary Differential Equations (ODEs) in Section 2.2. The coupled system
is one of the constraints among additional boundary and path constraints for the
OCP which we describe in Section 2.3. We emphasize the particular aspects in
which our problem setting differs and extends the setting most often found in PDE
constrained optimization.
The spaces L p (Ω),W k,p (Ω) endowed with their usual norms are Banach spaces.
The spaces H k (Ω) := W k,2 (Ω) equipped with their usual scalar product are Hilbert
spaces. The construction of L p (Ω) and W k,p (Ω) can be generalized to functions
with values in Banach spaces:
Definition 2.1 (Bochner spaces). By L p (Ω; X), 1 ≤ p < ∞, we denote the space of
all measurable functions v : Ω → X satisfying
vXp dμ < ∞.
Ω
Thus, dT /dt ∈ D ((t1 ,t2 ); X). We assume now that X → Y , where → denotes
continuous embedding. Hence it holds that
Let u ∈ L2 ((t1 ,t2 ); X). We say that du/dt ∈ L2 ((t1 ,t2 );Y ) if there exists u ∈
L2 ((t1 ,t2 ),Y ) such that
t2 t2
dT
u (t)ϕ(t)dt = ϕ =− u(t)ϕ (1) (t)dt for all ϕ ∈ C∞ ([t1 ,t2 ]; R),
t1 dt t1
Lemma 2.5. The space W (t1 ,t2 ) is a Hilbert space when endowed with the scalar
product
t2 t2
(u, v)W (t1 ,t2 ) = (u(t), v(t))V dt + (∂t u(t), ∂t v(t))V ∗ dt.
t1 t1
Theorem 2.6. We can alter every u ∈ W (t1 ,t2 ) on a set of measure zero to obtain
a function in C0 ([t1 ,t2 ]; H). Furthermore, if we equip C0 ([t1 ,t2 ]; H) with the norm
of uniform convergence then
Proof. See Dautray and Lions [37, Chapter XVIII, Theorem 1].
Corollary 2.7. For u ∈ W (t1 ,t2 ) the traces u(t1 ), u(t2 ) have a well-defined mean-
ing in H (but not in V in general).
For the control variables we assume q ∈ L2 ((t1 ,t2 ); Q) where Q ⊆ L2 (Ω)nq or
Q ⊆ L2 (∂ Ω)nq for distributed or boundary control, respectively. We can then for-
mulate the parabolic differential equation
∂t u(t) + A(q(t), u(t)) = 0, (2.1)
with a nonlinear elliptic differential operator A : Q × V → V ∗.
In the numerical
approaches which we present in Chapters 5 and 6 we exploit that A is an elliptic
operator. We further assume that A is defined via a semilinear (i.e., linear in the
last argument) form a : (Q ×V ) ×V → R according to
A(q(t), u(t)), ϕ V ∗ ×V = a(q(t), u(t), ϕ) for all ϕ ∈ V. (2.2)
We consider Initial Value Problems (IVPs), i.e., PDE (2.1) subject to u(t1 ) = u0 ∈
H. The question of existence, uniqueness, and continuous dependence of solutions
on the problem data u0 and q cannot be answered satisfactorily in a general set-
ting. However, there are problem-dependent sufficient conditions (compare, e.g.,
Gajewski et al. [57] for the case A(q(t), u(t)) = Aq (q(t)) + Au (u(t))). A thorough
discussion of this question is beyond the focus of this thesis.
Example 1. For illustration we consider the linear heat equation with Robin bound-
ary control and initial values
∂t u = Δu in (0, 1) × Ω, (2.3a)
∂ν u + αu = β q on (0, 1) × ∂ Ω, (2.3b)
u = u0 ,
t=0
(2.3c)
where α, β ∈ L∞ (∂ Ω) and ∂ν
denotes the derivative in the direction of the outwards
pointing normal ν on ∂ Ω. We choose V = H 1 (Ω) and H = L2 (Ω). Multiplication
with a test function ϕ ∈ V and integration by parts transform equations (2.3a)
and (2.3b) into
0= ∂t u(t)ϕ − (Δu(t))ϕ (2.4a)
Ω Ω
= ∂t u(t)ϕ + ∇u(t)T ∇ϕ − ∇u(t)T ν ϕ (2.4b)
Ω Ω ∂Ω
= ∂t u(t)ϕ + ∇u(t)T ∇ϕ + αu(t)ϕ − β q(t)ϕ (2.4c)
Ω Ω ∂Ω ∂Ω
=: ∂t u(t)ϕ + a(q(t), u(t), ϕ), (2.4d)
Ω
2.2 Coupled ODEs and PDEs 15
which serves as the definition for the semilinear form a and the corresponding
operator A. We immediately observe that a is even bilinear on (Q ×V ) ×V in this
example.
Φ : H × Rnv → R, r b : H × R n v → H × R nv ,
c e
r c : Q × R nv → R nr , re : Rnv → Rnr .
with the choice Φ(u(1), v(1)) = vi (1)/2. The applications we are interested in,
however, can have economical objective functions which are not of tracking type.
Constraints (2.6b) and (2.6c) determine the dynamics of the considered system.
We have already described them in detail in Sections 2.1 and 2.2 of this chapter.
Initial or boundary value constraints are given by equation (2.6d). Typical ex-
amples are pure initial value conditions via constant
or periodicity conditions
in which the intermediate control and state iterates of the method may be infeasi-
ble for equations (2.6b) through (2.6d). Of course feasibility must be attained in
the optimal solution.
Inequality (2.6e) is supposed to hold for almost all t ∈ (0, 1) and can be used
to formulate constraints on the controls and ODE states. We deliberately do not
include PDE state constraints in the formulation which give rise to various theo-
retical difficulties and are currently a very active field of research. We allow for
additional inequalities on the ODE states at the end via inequality (2.6f). In the
context of chemical engineering applications, the constraints (2.6e) and (2.6f) can
comprise flow rate, purity, throughput constraints, etc.
Problems with free time-independent parameters can be formulated within prob-
lem class (2.6) via introduction of additional ODE states vi with vanishing time
derivative v̇i (t) = 0. Although the software package MUSCOP (see Chapter 11)
treats time-independent parameters explicitly, we refrain from elaborating on these
issues in this thesis in order to avoid notational clutter.
OCP (2.6) also includes the cases of free start and end time via a time transfor-
mation, e.g., τ(t) = (1 − t)τ1 + tτ2 ∈ [τ1 , τ2 ],t ∈ [0, 1]. This case plays an impor-
tant role in this thesis, e.g., in periodic applications with free period duration, see
Chapter 15.
Concerning regularity of the functions involved in OCP (2.6), we take a prag-
matic view point: We assume that the problem can be consistently discretized
(along the lines of Chapter 3) and that the resulting finite dimensional optimiza-
tion problem is sufficiently smooth on each discretization level to allow for em-
ployment of fast numerical methods (see Chapter 5).
3 Direct Optimization: Problem
discretization
The goal of this chapter is to obtain a discretized version of OCP (2.6). We discuss
a so-called direct approach and summarize its main advantages and disadvantages
in Section 3.1 in comparison with alternative approaches. In Sections 3.2 and 3.3
we discretize OCP (2.6) in two steps. First we discretize in space and obtain a
large-scale ODE constrained OCP which we then discretize in time to obtain a
large-scale Nonlinear Programming Problem (NLP) presented in Section 3.5. The
numerical solution of this NLP is the subject of Part II in this thesis.
Necessary infinite
Infinite dimensional optimize
−−−−→ dimensional
optimization problem
optimality conditions
⏐ ⏐
⏐ ⏐
discretize discretize
Necessary finite
Finite dimensional
−−−−→ dimensional
optimization problem optimize
optimality conditions
Figure 3.1: The two main routes to approximate an infinite dimensional optimization prob-
lem (upper left box) with necessary finite dimensional optimality conditions (lower right
box) are direct approaches (discretize-then-optimize, lower left path) versus indirect ap-
proaches (optimize-then-discretize, upper right path).
space and then in time is called Method Of Lines (MOL) and is often applied for
parabolic problems (see, e.g., Thomée [150]). We must exercise care that the spa-
tial discretization is appropriate for the concrete problem at hand. For instance, an
FEM for advection dominated problems must be stabilized, e.g., by a Streamline
Upwind Petrov Galerkin formulation (see, e.g., Brooks and Hughes [31]), and an
NDGM for diffusion dominated problems must be stabilized by a jump penalty
term (see, e.g., Warburton and Embree [161]).
We assume that after discretization we obtain a finite dimensional space Qh ⊆ Q
and a hierarchy of finite dimensional spaces Vhl , l ∈ N, satisfying
Vh1 ⊂ Vh2 ⊂ · · · ⊂ V.
We choose this setting for the following reasons: For many applications, especially
in chemical engineering, an infinite dimensional control is virtually impossible to
implement on a real process. In the case of room heating for instance, the temper-
ature field is distributed in the three-dimensional room, but the degrees of freedom
for the control will still be the scalar valued position of the radiator operation knob.
In this case, a one-dimensional discretization Qh of Q is fully sufficient. For the
applications we consider, we always assume Qh to be a low dimensional space.
It is of paramount importance, however, to accurately resolve the system state u.
For this reason, we assume that the spaces Vhl are high dimensional for for large
l. The numerical methods we describe in Part II will rely on and exploit these
assumptions on the dimensionality of Qh and Vhl .
We can then use the finite dimensional spaces Vhl and Qh to obtain a discretiza-
tion of the semilinear form a and thus the operator A from equation (2.2). On each
level l we are led to an ODE of the form
Mul u̇l (t) = f PDE(l) (q(t), ul (t), v(t)),
l
with symmetric positive-definite NVl -by-NVl matrix Mul , ul (t) ∈ RNV , q(t) ∈ RNQ ,
l l
and f PDE(l) : RNQ × RNV × Rnv → RNV . In this way we approximate PDE (2.1)
with ODEs which are of large scale on finer discretization levels l. Let us illustrate
this procedure in an example.
Example 2. We continue the example of the heat equation (Example 1) on the
unit square Ω = (0, 1)2 with boundary control. For the discretization in space we
employ FEM. Let us assume that we have a hierarchy of nested triangular grids for
the unit square (compare Figure 3.2) with vertices ξil ∈ Ω, i = 1, . . . , NVl , on level
l ∈ N. Let the set of triangular elements on level l be denoted by T l . We define
the basis functions ϕil by requiring
ϕil (ξ jl ) = δi j , ϕil is linear on each element T ∈ T l ,
22 3 Direct Optimization: Problem discretization
Figure 3.2: First four levels of an exemplary hierarchy of nested triangular meshes for the
unit square obtained by uniform refinement with NVl = (2l−1 + 1)2 vertices on level l =
1, . . . , 4.
with δi j denoting the Kronecker Delta. This construction yields the well-known hat
functions. We then define the spaces Vhl simply as the span of the basis functions
ϕil , i = 1, . . . , NVl .
For the discretization of Q we assume that a partition of ∂ Ω in segments Sk ,
k = 1, . . . , NQ , is given and choose Qh as the span of their characteristic functions
ψk = χSk , which yields a piecewise constant discretization of the control on the
boundary of the domain.
Let
Nl Nl
ul = ∑i=1 wl = ∑i=1 q = ∑i=1
N
V
uli ϕil ∈ Vhl , V
wli ϕil ∈ Vhl , Q
qi ψ i ∈ Q h
denote arbitrarily chosen discretized functions and their coordinates (in bold type-
face) within their finite dimensional spaces. Then we obtain the following expres-
sions for the terms occurring in equation (2.4) which allow for the evaluation of
the integrals via matrices:
Nl
ul wl = ∑i, Vj=1 uli ϕil ϕ lj wlj =: (ul )T MVl wl ,
Ω Ω
l
(∇u ) ∇w = ∑i, j=1 ui
N
l T l V l
(∇ϕi ) ∇ϕ j wlj =: (ul )T Sl wl ,
l T l
Ω Ω
NVl
αwl q = ∑i=1 ∑ j=1 i ∂ Ω i j wlj =: qT MQl wl ,
NQ
q αψ ϕ l
∂Ω
Nl
β ul wl = ∑i, Vj=1 uli β ϕil ϕ lj wlj =: (ul )T M∂l wl .
∂Ω ∂Ω
3.3 Direct Multiple Shooting: Discretization in time 23
The occurring matrices are all sparse because each basis function has by construc-
tion a support of only a few neighboring elements. We now substitute u(t), ϕ and
q(t) in equation (2.4) with their discretized counterparts to obtain
u̇l (t)T MVl eli = −ul (t)T Sl eli − ul (t)T M∂l eli + q(t)T MQl eli , for i = 1, . . . , NVl ,
where eli denotes the i-th column of the N l -by-N l identity matrix. Exploiting sym-
metry of MVl , Sl , and M∂l yields the equivalent linear ODE formulation
MVl u̇l (t) = (−Sl − M∂l )ul (t) + (MQl )T q(t) =: f PDE(l) (q(t), ul (t), v(t)). (3.1)
It is well-known that the state mass matrix on the left hand side of equation (3.1) is
symmetric positive definite. To multiply equation (3.1) from the left with the dense
matrix (MVl )−1 is often avoided in order to preserve sparsity of the right hand side
matrices.
We want to conclude Example 2 with the remark that in a FVM or an NDGM,
where the basis functions are discontinuous over element boundaries, the mass
matrix has block diagonal form and hence sparsity is preserved for (MVl )−1 . For
spectral methods, all occurring matrices are usually dense anyway. In both cases,
the inverse mass matrix is usually formulated explicitly in the right hand side of
equation (3.1).
I i = (t i−1 ,t i ), i = 1, . . . , nMS ,
with values
q(t) = ∑i=1
n
q χI i (t).
MS i−1
24 3 Direct Optimization: Problem discretization
Piecewise higher order discretizations in time are also possible, as long as the
shooting intervals stay decoupled. Otherwise we loose the possibility for structure
exploitation which is important for numerical efficiency reasons as we discuss
in Chapter 8. In this thesis we restrict ourselves to piecewise constant control
discretizations in time for reasons of simplicity.
We now introduce artificial initial values (sl,i , vi ), i = 0, . . . , nMS , for the semi-
discretized PDE states ul (t) and the ODE states v(t), respectively. We define
Nl
f ODE(l) (q(t), ul (t), v(t)) := f ODE (∑ j=1 q j (t)ψ j , ∑ j=1
N
Q V
ulj (t)ϕ j , v(t))
Mul u̇l (t) = f PDE(l) (qi−1 , ul (t), v(t)), t ∈ Ii, ul (t i−1 ) = sl,i−1 , (3.2a)
v̇(t) = f ODE(l) (qi−1 , ul (t), v(t)), t ∈ Ii, v(t i−1 ) = vi−1 . (3.2b)
(ul,i (t; qi−1 , sl,i−1 , vi−1 ), vl,i (t; qi−1 , sl,i−1 , vi−1 )).
Local existence and uniqueness of (ul,i , vl,i ) are guaranteed by the Picard-Lindelöf
theorem if the functions f PDE(l) and f ODE(l) are Lipschitz continuous in the second
and third argument. By means of (ul,i , vl,i ) we obtain a piecewise, finite dimen-
sional parametrization of the state trajectories. To recover continuity of the entire
trajectory across the shooting grid nodes we have to impose matching conditions
l,i i i−1 l,i−1 i−1 l,i
u (t ; q , s ,v ) s
− = 0, i = 1, . . . , nMS .
vl,i (t i ; qi−1 , sl,i−1 , vi−1 ) vi
Remark 3.1. We introduce an additional artificial control variable qnMS on the last
shooting grid node in order to have the same structure of degrees of freedom in
each ti , i = 0, . . . , nMS . We shall always require qnMS = qnMS −1 . This convention
simplifies the presentation and implementation of the structure exploitation that
we present in Chapter 8.
Remark 3.2. It is also possible and numerically advantageous to allow for differ-
ent spatial meshes on each shooting interval I i in combination with a-posteriori
mesh refinement, see Hesse [83]. In that case the matching conditions have to
be formulated differently. This topic, however, is beyond the scope of this thesis.
We restrict ourselves to uniformly refined meshes which are equal for all shooting
intervals.
3.4 Discretization of path constraints 25
minimize
n
Φl (sl,nMS , vnMS ) (3.3a)
(qi ,sl,i ,vi )i=0
MS
For the efficient solution of NLP (3.3) we have developed an inexact Sequential
Quadratic Programming (SQP) method which we describe in Part II. We conclude
this chapter with a summary of the numerical challenges:
Large scale. The NLPs have nNLP(l) = (nMS + 1) NVl + nv + nMS NQ variables
and are thus considered large-scale for finer levels l. The numerical methods
which we describe in Part II aim at the efficient treatment of NLP (3.3) for
large NVl ≈ 105 . The number of shooting intervals nMS will be between 101
and 102 which amounts to an overall problem size nNLP(l) ≈ 107 . We want
to remark that this calculation does not include the values of ul,i which have
to be computed in intermediate time steps between shooting nodes. There
can be between 101 and 102 time steps per interval.
Efficient derivative generation. It is inevitable for the solution of large-scale op-
timization problems to use derivative-based methods. Hence we need nu-
merical methods which deliver consistent derivatives of the functions occur-
ring in NLP (3.3), especially in the matching conditions (3.3c) and (3.3d). In
Chapter 10 we describe such a method which efficiently computes consistent
derivatives of first and second order in an automated way.
Structure exploitation. Efficient numerical methods must exploit the multiple
shooting structure of NLP (3.3). We present a condensing approach in Chap-
ter 8 which operates on linearizations of NLP (3.3). Furthermore we develop
preconditioners in Chapter 6 which exploit special spectral properties of the
shooting Wronksi matrices. These spectral properties arise due to ellipticity
of the operator A.
Mesh-independent local convergence. One of the main results of this thesis is
that these preconditioners lead to mesh-independent convergence of the in-
exact SQP method, i.e., the number of iterations is asymptotically bounded
by a reasonably small constant for l → ∞. We prove this assertion for a
model problem in Chapter 6 and the numerical results that we have obtained
on the application problems in Part III suggest that this claim also holds for
difficult real-world problems.
Global convergence. Often there is only few a-priori information available about
the solution of real-world problems. Hence it is paramount to enforce con-
vergence of the inexact SQP method also from starting points far away from
the solution. However, it must be ensured that an early transition to fast local
convergence is preserved. We describe such a globalization strategy based
on natural level functions in Chapter 5.
4 Elements of optimization theory
In this short chapter we consider the NLP
minimize
n
f (x) (4.1a)
x∈R
s. t. gi (x) = 0, i∈E, (4.1b)
gi (x) ≥ 0, i∈I, (4.1c)
where f : Rn → R and g : Rn → Rm
are twice continuously differentiable functions
and the sets E and I form a partition of {1, . . . , m} =: m = E ∪˙ I . In the case of
E = m, I = {}, NLP (4.1) is called Equality Constrained Optimization Problem
(ECOP).
∇x L (x∗ , y∗ ) = 0, (4.2a)
gi (x∗ ) = 0, i∈E, (4.2b)
gi (x∗ ) ≥ 0, i∈I, (4.2c)
y∗i ≥ 0, i∈I, (4.2d)
y∗i gi (x ) = 0,
∗
i ∈ m. (4.2e)
In the next definition we characterize a property which is favorable for the de-
termination of the active set in a numerical algorithm because small changes in the
problem data will not lead to changes in the active set at the solution.
C (x, y) = {d ∈ Fl (x) | d T ∇gi (x) = 0, for all i ∈ A (x) ∩ I with yi > 0}.
The cone of critical directions plays an important role in the following sufficient
optimality condition.
Theorem 4.14. Let (x∗ , y∗ ) satisfy the KKT conditions (4.2). If furthermore the
Second Order Sufficient Condition (SOSC)
Numerical methods
5 Inexact Sequential Quadratic
Programming
In this chapter we develop a novel approach for the solution of inequality con-
strained optimization problems. We first describe inexact Newton methods in Sec-
tion 5.1 and investigate their local convergence in Section 5.2. In Section 5.3 we
review strategies for the globalization of convergence and explain a different ap-
proach based on generalized level functions and monotonicity tests. An example
in Section 5.4 illustrates the shortcomings of globalization strategies which are not
based on the so called natural level function. We review the Restrictive Monotonic-
ity Test (RMT) in Section 5.5 and propose a Natural Monotonicity Test (NMT) for
Newton-type methods based on a Linear Iterative Splitting Approach (LISA). This
combined approach allows for estimation of the critical constants which character-
ize convergence. We finally present how these results can be extended to global
inexact SQP methods. We present efficient numerical solution techniques of the
resulting sequence of Quadratic Programming Problems (QPs) in Chapters 8 and 9.
with scalars αk ∈ (0, 1] and matrices M(z) ∈ RN×N . The scalar αk is a damping or
underrelaxation parameter. We shall see in Sections 5.2 and 5.3 that the choice of
αk = 1 is necessary for fast convergence close to a solution but choices αk < 1 are
necessary to achieve convergence from initial guesses z0 which are not sufficiently
close to a solution.
Different choices of M lead to different Newton-type methods. Typical and im-
portant choices for M include Quasi-Newton methods (based on secant updates,
see, e.g., Nocedal and Wright [121]), the Simplified Newton method (M(z) =
J −1 (z0 )), and the Newton method (M(z) = J −1 (z)), provided the inverses exist.
We have developed a method which uses Linear Iterative Splitting Approach (see
Section 5.6.2) with a Newton-Picard deflation preconditioner (described in Chap-
ter 6) to evaluate M.
Before we completely dive into the subject we want to clarify the naming of
methods. We use Newton-type method as a collective term to refer to methods
which can be cast in the form of equation (5.1). If the linearized subproblems
are solved by an iterative procedure we use the term inexact Newton method. Un-
fortunately the literature is not consistent here: The often cited paper by Dembo
et al. [41] uses inexact Newton method in the sense of our Newton-type method and
Newton-iterative method in the sense of our inexact Newton method to distinguish
between solving
˜ k )Δzk = −F(zk )
J(z or J(zk )Δzk = −F(zk ) + rk , (5.2)
˜ k )Δzk = −F(zk ) + rk
J(z
in each step. From the point of view that equations (5.2) are merely characterizing
the backward and the forward error of Δzk for J(zk )Δzk = −F(zk ), we believe that
equations (5.2) should not be the basis for categorizing algorithms but rather be
kept in mind for the analysis of all Newton-type methods. The following lemma
shows that one can move from one interpretation to the other:
Lemma 5.1. Let Δz∗ solve J(zk )Δz∗ = −F(zk ). If Δzk is given via
Conversely, if rk and Δzk are given and Δzk 2 > 0 then one possible J(z
˜ k ) is given
by
˜ k ) = J(zk ) − rk (Δz ) .
k T
J(z
(Δzk )T Δzk
Moreover, if J(zk ) is invertible and (Δzk )T Δz∗ = 0 then
k −1 (Δzk − Δz∗ )(Δzk )T
M(z ) = J(z ) = I +
k ˜ J(zk )−1 .
(Δzk )T Δz∗
Proof. The first assertion is immediate. The second assertion can be shown via
rk (Δz )
k T
J(zk )−1 (Δz k −1
k )T Δzk J(z )
k −1 k −1
M(z ) = J(z
k ˜ ) = J(z ) + .
1 − (Δz(Δz )
k T
k −1 r
k )T Δzk J(z ) k
The last assertion then follows from J(zk )−1 rk = Δzk − Δz∗ .
In this thesis we focus on computing Δzk iteratively via the iteration
with M̂(zk ) ∈ RN×N . We call iteration (5.3) Linear Iterative Splitting Approach
(LISA) to emphasize that the iteration (which we further discuss in Section 5.6.2)
is linear and based on a splitting
J(zk ) = J(z
ˆ k ) − ΔJ(zk ), ˆ k )−1 .
M̂(zk ) = J(z
A linear iteration like (5.3) can in principle be accelerated by the use of Krylov-
space methods at the cost of making the iteration nonlinear. We abstain from
nonlinear acceleration in this thesis because the Newton-Picard preconditioners
are already powerful enough when used without acceleration (see Chapter 6).
In the following sections we review the theory for local and global convergence
of Newton-type methods.
N = {(z, z ) ∈ D × D | z = z − M(z)F(z)}
Remark 5.4. If M(z) is invertible, then the κ-condition can also be written as
M(z )(M −1 (z) − J(z))(z − z ) ≤ κ z − z , ∀(z, z ) ∈ N .
The following theorem characterizes the local convergence of a full step (i.e.,
αk = 1) Newton-type method in a neighborhood of the solution. Because of its
importance we include the well-known proof.
5.2 Local convergence 37
Theorem 5.5 (Local Contraction Theorem). Let J and M satisfy the ω-κ-condi-
tions in D and let z0 ∈ D. If c0 < 1 and D0 ⊆ D, then zk ∈ D0 and the sequence
(zk ) converges to some z∗ ∈ D0 with convergence rate
2
k+1
Δz ≤ ck Δzk = κ Δzk + (ω/2) Δzk .
M(z∗ )F(z∗ ) = 0.
F(z∗ ) = 0.
Proof based on the Banach Fixed Point Theorem. The assumption c0 < 1 and the
Definition of D0 imply that z0 , z1 ∈ D0 . We assume that zk+1 ∈ D0 . Then
k+1
Δz = M(zk+1 )F(zk+1 )
follows that (zk ) is a Cauchy sequence and thus converges to a fixed point z∗ ∈ D0 .
For the a-priori estimate consider
∞ ∞ (ck ) j
k+ j k
z − z∗ ≤ ∑ Δzk+ j+i ≤ ∑ (ck )i Δzk+ j ≤ Δz .
i=0 i=0 1 − ck
In the limit
z∗ = z∗ − M(z∗ )F(z∗ ) ⇒ M(z∗ )F(z∗ ) = 0
holds which shows the remaining assertions.
Remark 5.6. If F is linear we obtain ω = 0 and if furthermore M(z) is constant the
convergence theory is completely described by Theorem 5.26 to be presented.
Remark 5.7. Assume J is nonsingular throughout D0 . Then the full step Newton
method with M(z) = J −1 (z) converges quadratically in D0 due to κ = 0.
Remark 5.8. In accordance with Deuflhard’s algorithmic paradigm (see Deuflhard
[44]) we assume the constants κ and ω to be the infimum of all possible candidates
which satisfy the inequalities in their respective definitions. These values are in
general computationally unavailable. Within the algorithms to be presented we
approximate the infima from below with computational estimates denoted by [κ]
and [ω] by sampling the inequalities over a finite subset N ⊂ N which comprises
various iterates of the algorithm.
solve the linear system exactly (see Battermann and Heinkenschloss [9], Batter-
mann and Sachs [10], Biros and Ghattas [19]).
We shall show in Section 5.4 that both line search and trust region methods can
lead to unnecessarily damped iterates in the vicinity of the solution where fast local
convergence in the sense of the Local Contraction Theorem 5.5 is already possible.
Our aim in this section is to present the theoretical tools to understand this un-
desirable effect and to introduce the idea of monotonicity tests.
We begin the discussion on the basis of the Newton method and extend it to
Newton-type methods in Section 5.6 and to inexact SQP methods for inequality
constrained optimization problems in Section 5.7.
Lemma 5.9. Let A, B ∈ GL(N). Then the iterates zk for the Newton method on
F(z) and the iterates z̃k for
F̃(z̃) := AF(Bz̃)
with z̃0 := B−1 z0 are connected via
z̃k = B−1 zk , k ∈ N.
˜ k )−1 F(z̃k ) = z̃k − B−1 J(Bz̃k )−1 A−1 AF(Bz̃k ) = B−1 zk+1 ,
z̃k+1 = z̃k − J(z̃
We conclude this section with a descent result for the Newton direction on gen-
eralized level functions
T (z|A) := 12 AF(z)22 , A ∈ GL(N).
Generalized level functions extend the concept of the classical level function T (z|I)
and play an important role in affine invariant globalization strategies. The follow-
ing simple but nonetheless remarkable lemma (see, e.g., Deuflhard [44]) shows
that the Newton direction is a direction of descent for all generalized level func-
tions.
Lemma 5.10. Let F(z) = 0. Then, for all A ∈ GL(N), the Newton increment
Δz = −J(z)−1 F(z) satisfies
ΔzT ∇T (z|A) = −2T (z|A) < 0.
T
Proof. ΔzT ∇T (z|A) = −F(z)T J(z)−T F(z)T AT AJ(z) = −2T (z|A) < 0.
However, decrease T (z + αΔz|A) < T (z|A) might only be valid for α 1. We
shall illustrate this problem with an example in Section 5.4. For the construction of
efficient globalization strategies, A must be chosen such that the decrease condition
is valid for a maximal range of α, as we shall discuss in Sections 5.5 and 5.6.
The geometric derivation of the Newton path due to Deuflhard [42, 43, 44] and
the connection to the continuous analog of the Newton method characterized by
Davidenko [38] is given by the following result.
5.3 Globalization of convergence 41
Theorem 5.11. Let J(z) be nonsingular for all z ∈ D. For some A ∈ GL(N) let the
in z0 be compact and contained in D. Then
path-connected component of G(z0 |A)
the path-connected component of G(z0 ) is a topological path z : [0, 2] → RN , the
so-called Newton path, which satisfies
which gives rise to the function z(α) upon invocation of the Implicit Function
Theorem. After solving equation (5.5) for F(z0 ) and using the reparametrization
α(t) = 1 − exp(−t) we can recover the so-called continuous Newton method or
Davidenko differential equation (Davidenko [38])
Theorem 5.11 justifies that the Newton increment Δzk is a distinguished direc-
tion not only locally but also far away from a solution. It might only be too large,
hence necessitating the need for damping through αk .
In other words, the Newton path is the set of points generated from infinitesimal
Newton increments (denoted by ż(t) instead of Δzk ). Performing nonzero steps
αk Δzk in iteration (5.1) for the Newton method gives rise to a different Newton
path emanating from each zk , k ∈ N. It seems inevitable that for a proof of global
convergence based on the Newton path the iterates zk must be related to a single
Newton path, which we discuss in Section 5.5.
which have several attractive properties. We can already observe that descent in
Tk∗ can be evaluated by testing for natural monotonicity
k+1 k
Δz < Δz ,
where one potential step of the Simplified Newton method can be used to evaluate
Δzk+1 according to
J(zk )Δzk+1 = −F(zk+1 ).
As already mentioned, the Lipschitz constant ω plays a fundamental role in the
global convergence theory based on generalized level functions. In contrast to
Bock [24], however, Deuflhard [44] uses a different definition for the Lipschitz
constant:
Definition 5.13 (ω-condition for the Newton method). The Jacobian J satisfies
the ω-condition < ∞ such that for all (z, z ) ∈ D × D
in D if there exists ω
2
J(z)−1 J(z ) − J(z) (z − z) ≤ ω
z − z .
which coincides with Definition 5.13 of ω except for the evaluation of the weight-
ing matrix in the Lipschitz condition at a different point. Because Nt is much
smaller than D × D, the constant ω must be expected to be much larger than ω.
5.3 Globalization of convergence 43
This will lead to smaller bounds on the step sizes αk . Furthermore, in practical
computations with a Newton-type method, only ω can be estimated efficiently
is not explicitly restricted to the set of Newton pairs
from the iterates because ω
N.
Remark 5.15. Most proofs which rely on ω can also be carried out in a similar
fashion with ω but some theoretical results cannot be stated as elegantly. As an
example, the occurrence of the condition number cond(AJ(zk )) in Theorem 5.16
We take up the pragmatic position that ω should be used for all
relies on using ω.
if we can gain theoretical insight
practical computations but we also recede to ω
about qualitative convergence behavior of Newton-type methods.
The first theorem which relies on ω characterizes step sizes αk which yield
optimal values for a local descent estimate of generalized level functions T (z|A)
in the Newton method.
Theorem 5.16. Let D be convex, J(z) nonsingular for all z ∈ D. Let furthermore J
satisfy the ω-condition in D, zk ∈ D, A ∈ GL(N), and G(zk |A) ⊂ D. Let Δzk denote
the Newton direction and define the Kantorovich quantities
hk := Δzk ω, hk := hk cond(AJ(zk )).
where
tk (α|A) := 1 − α + (1/2)α 2 hk (A).
The optimal choice of the damping factor in terms of this local estimate is
αk ∈ [ε, 2α k (A) − ε]
and thus the largest value for the step size α k . As already mentioned at the begin-
ning of this section, we recall that descent in Tk∗ can be evaluated by the NMT
k+1 k
Δz < Δz ,
where Δzk+1 is the increment for one potential Simplified Newton step.
Natural level functions have several outstanding properties as stated by Deufl-
hard [44, Section 3.3.2]:
Extremal properties For A ∈ GL(N) the reduction factors tk (α|A) and the theo-
retical optimal damping factors α k (A) satisfy
With A = Ak we obtain
Merging property Full steps and thus fast local convergence are guaranteed in
proximity of the solution
k ⇒ hk ≤ 1 ⇒ α k (Ak ) = 1.
Δz ≤ 1/ω
2
with starting guess z0 = (50, 1) and solution z∗ = (0, −12.5) (compare Figure 5.1).
We observe that the classical level set (contained within the dashed curve) is
shaped like a bent ellipse. The excentricity of the ellipse is due to the condi-
tion number of J(z0 ), which is cond2 (J(z0 )) = 50. The ellipse is bent because of
the mild nonlinearity in the second component of F. A short computation yields
ω ≤ 0.01.
We further observe that the direction of steepest descent for the classical level
function T (z|I) and for the natural level function T (z|J(z0 )−1 ) describe an angle of
87.7 degrees. Thus the Newton increment, which coincides with the direction of
steepest descent for the natural level function (see Section 5.3.3), is almost parallel
to the tangent on the classical level set. Consequently only heavily damped Newton
steps lead to a decrease in the classical level function. We obtain optimal descent
in T (z|I) with a stepsize of α0 ≈ 0.077, although the solution z∗ can be reached
46 5 Inexact Sequential Quadratic Programming
with two iterations of a full step Newton method. This behavior illustrates how
the requirement of descent in the classical level function impedes fast convergence
within the domain of local contraction.
In real-world problems the conditioning of J(z∗ ) is typically a few orders of
magnitude higher than 50, leading to even narrower valleys in the classical level
function. Additionally, nonlinear and highly nonlinear problems with larger ω
give rise to higher curvature of these valleys, rendering the requirement of descent
in the classical level function completely inappropriate.
Especially in the light of inexact Newton methods as we describe in Section 5.6,
the use of natural level functions T (z|J(zk )−1 ) is paramount: Even relatively small
perturbations of the exact Newton increment can result in the inexact Newton in-
crement being a direction of ascent in T (z|I).
60
40
20
z2 coordinate
-20
-40
-60
-80
-50 0 50
z1 coordinate
Figure 5.1: Rosenbrock-type example (adapted from Bock [24]): The Newton increment
(solid arrow) is almost perpendicular to the direction of steepest descent for the classi-
cal level function T (z|I) (dashed arrow) in the initial iterate (marked by ◦). Only heavily
damped Newton steps lead to descent in the classical level function due to the narrow and
bent classical level set (contained within the dashed curve). In contrast, the Newton step
is completely contained within the more circle-shaped natural level set (contained within
the solid curve) corresponding to the natural level function T (z|J(z0 )−1 ). Within two full
Newton steps, the solution (marked by ×) is reached.
Let Δzk denote the exact Newton step and δ zk the inexact Newton step obtained
from LISA. Furthermore, define the residual
in terms of the exact Newton steps Δzk and the exact Simplified Newton steps
Δzk+1 (α) defined via
J(zk )Δzk+1 (α) = −F(zk + αδ zk ).
We emphasize the occurrence of the inexact Newton step δ zk on the right hand
side.
Lemma 5.18. Let δk < 1 and define hδ := ω
2
δ zk . Then we obtain the estimate
k
δk 1 hδ
Θk (α) ≤ 1 − 1 − α + α2 k . (5.7)
1 − δk 2 1 − δk
The optimal damping factor is
α k = min(1, (1 − 2δk )/hδk ).
If we implicitly define ρ via
ρ δ
δk =
αh (5.8)
2 k
and assume that ρ ≤ 1 we obtain the optimal damping factor
Lemma 5.19. With the current notation and definitions we obtain the estimate
k+1
Δz − (1 − α)δ zk ≤ 12 α 2 hδk δ zk .
J(zk )δ zk+1
i = (−F(zk + αδ zk ) + rk ) +
rk+1
i .
which can be computed efficiently. The following lemma characterizes the depen-
k on α.
dence of Θ
Lemma 5.20. Assume that LISA for equation (5.10) with initial value crossing
(5.11) has been iterated until
k+1
Δz − δ zk+1
i
ρi =
k+1
< 1.
(5.12)
Δz − δ zk+1
0
Proof. The proof for the LISA case is the same as for the GBIT case, see Deuflhard
[44, Lemma 3.20].
The quantity ρi , however, cannot be evaluated directly because we must not
compute Δzk+1 exactly for efficiency reasons. Instead we define the computable
estimate
k+1
Δz − δ zk+1i εi
ρi = ≈ , (5.13)
k+1 k+1
δ zi − δ z0 δ zi − δ zk+1
k+1
0
where
ε i is an estimate for the error of LISA, see Section 5.6.2.
Lemma 5.21. With the notation and assumptions of Lemma 5.20 we have
k+1 k+1 k+1
Δz − δ zk+1 i ≤ ρ i (1 + ρ i ) Δz − δ z0 .
Proof. We follow Deuflhard [44, Lemma 3.20 for GBIT and below]: The applica-
tion of the triangle inequality and assumptions (5.12) and (5.11) yield
k+1
δ zi − (1 − α)δ zk ≤ Δzk+1 − (1 − α)δ zk + δ zk+1 i − Δzk+1
= (1 + ρi ) Δzk+1 − (1 − α)δ zk .
Using definition (5.13) on the left hand side then delivers the assertion.
An immediate consequence of Lemma 5.21 is the inequality
Deuflhard [44] proposes to base the estimation of ρi on equating the left and right
hand sides of inequality (5.14) to obtain
Then accuracy control for the inner iterations can be based on the termination
condition
ρi ≤ ρmax with ρmax ≤ 14 ,
or, following (5.15),
ρ i ≤ ρ max with ρ max ≤ 13 .
We feel urged to remark that this is heuristic insofar as from inequality (5.14) we
can only conclude
ρ i ≥ ρi /(1 + ρi ) (and not “≤”).
5.6 Natural Monotonicity for LISA-Newton methods 51
Thekoptimal
damping factor αk from Lemma 5.18 depends on the unknown hδk =
δ z which must be approximated. Using the [.] notation (see Remark 5.8) we
ω
approximate hδk with a so-called Kantorovich estimate
[hδk ] = [ω]
δ zk ≤ hδk ,
which leads via equation (5.8) to a computable estimate of the optimal step size
We now find a bound for αhδk with optimal realizable step size
α = [α k ] = min(1, 1/ (1 + ρ)[hδk ] ).
Remark 5.23. With the additional assumption that σ < 1/(1 + 2ρ) we obtain Θk <
1.
Based on Lemma 5.22 we can now formulate an inexact NMT. To this end we
substitute the bound (5.18) by the computable inexact NMT
k+1
δ z ρ − ρ
k =
Θ < 1− α, (5.20)
δ zk 1+ρ
by replacing σ with its upper bound σ = 1. In order to be a meaningful test we
additionally require ρ < ρ. The inexact NMT (5.20) becomes computable if we
select a ρ < 1 and further impose for the relative error of the inner LISA that
ρ ρ(1 + σ )
δk ≤ α[hδk ] ≤
2 2(1 + ρ)
which is substituted by the computable condition
ρ 1
δk ≤ =: δ ≤
2(1 + ρ) 4
for the case of α < 1.
If in the course of the computation the inexact NMT is not satisfied, we reduce
the step size on the basis of the a-posteriori Kantorovich estimate (5.16), denoted
by [hδk ]∗ , according to
αknew := max(min(1/ (1 + ρ)[hδk ]∗ , α/2), αmaxred α) .
α=αk
old
Taking the min and max is necessary to safeguard the stepsize adaption against
too cautious (reduction by at least a factor of two) and too aggressive changes
(reduction by at most a factor of αmaxred ≈ 0.1). Especially aggressive reduction
must be safe-guarded because the computation of the Kantorovich estimate [hδk ]
in equation (5.16) is inflicted with a cancellation error which is then amplified by
1/α 2 . Furthermore, the cancellation error gets worse for smaller α because then
δ zk+1 is closer and closer to (1 − α)δ zk as a consequence of Lemma 5.20.
For the initial choice for αk we recede to the a-priori Kantorovich estimate (5.17)
via
αk+1 = min(1, 1/ (1 + ρ)[hδk ] )
The initial step size α0 has to be supplied by the user. As a heuristic one can
choose α0 = 1, 0.01, 0.0001 for mildly nonlinear, nonlinear, and highly nonlinear
problems, respectively.
54 5 Inexact Sequential Quadratic Programming
If κlin < 1 then M̂ and Jˆ are invertible and LISA (5.21) converges for all F̂, ζ0 ∈ RN .
Conversely, if LISA converges for all F̂, ζ0 ∈ RN , then κlin < 1. The asymptotic
linear convergence factor is given by κlin .
We now investigate the connection of the preconditioner M̂(z) with the approxi-
mated inverse M(z) of the Local Contraction Theorem 5.5. The results have al-
ready been given by Ortega and Rheinboldt [122, Theorem 10.3.1]. We translate
them into the framework of the Local Contraction Theorem 5.5.
Lemma 5.27. Let ζ0 = 0 and l ≥ 1. Then the l-th iterate of LISA is given by the
truncated Neumann series for M̂ Jˆ according to
l−1
ζl = − ∑ (I − M̂J)ˆ i M̂ F̂.
i=0
Proof. Let l ∈ N and assume that the assertion holds for ζl . Then we obtain
l−1
ζl+1 = (I − M̂ J)ζ
ˆ l − M̂ F̂ = − ∑ (I − M̂J)ˆ i+1
+ I M̂ F̂
i=0
l
=− ∑ (I − M̂J)ˆ
i
M̂ F̂.
i=0
M(z)J(z) = I − (I − M̂(z)J(z))l ,
J(z)M(z) = I − (I − J(z)M̂(z))l .
Theorem 5.29. Let z∗ ∈ D satisfy M̂(z∗ )F(z∗ ) = 0. For the LISA-Newton method
with continuous preconditioner M̂(z) and l steps of LISA with starting guess ζ0 =
0, the following two assertions are equivalent:
i) The LISA at z∗ converges for all starting guesses and right hand sides, i.e.,
ii) The matrices M̂(z∗ ) and J(z∗ ) are invertible and for every ε > 0 there exists
a norm .∗ and a neighborhood U of z∗ such that the κ-condition 5.3 for
M(z) in U based on .∗ is satisfied with
κ ≤ κlin
l
+ε (where κlin < 1).
and that M̂(z∗ ) and J(z∗ ) are invertible. Recall that we have
by Lemma 5.28. Let ε > 0. Then the Hirsch Theorem delivers a norm .∗ such
that
I − M(z∗ )J(z∗ )∗ ≤ σr (I − M̂(z∗ )J(z∗ ))l + ε/2 ≤ κlin
l
+ ε/2.
5.6 Natural Monotonicity for LISA-Newton methods 57
By continuity of F, J, M̂, .∗ , and the inverse we obtain the existence of a neigh-
borhood U of z∗ such that all z ∈ U satisfy
M(z − M(z)F(z))(I − J(z)M(z))F(z)∗
≤ M(z − M(z)F(z))(M −1 (z) − J(z))∗ M(z)F(z)∗
≤ (κlin
l
+ ε) M(z)F(z)∗
because
M(z∗ − M(z∗ )F(z∗ ))(M −1 (z∗ ) − J(z∗ )) = I − M(z∗ )J(z∗ ).
Comparison with the κ-condition 5.3 yields
κ ≤ κlin
l
+ ε.
ii) ⇒ i): Let ẑ ∈ RN be a nonzero vector, small enough such that
z(t) := z∗ + t ẑ ∈ U, t ∈ (0, 1]. (5.22)
Because M(z∗ ) is invertible we obtain F(z∗ ) = 0 and write F(z(t)) in the form
1
F(z(t)) = F(z∗ + t ẑ) − F(z∗ ) = t J(z∗ + τt ẑ)ẑdτ,
0
which leads to
1
z (t) − z(t) := −M(z(t))F(z(t)) = −tM(z(t)) J(z(τt))dτ ẑ.
0
where
ai bi
Ci = .
−bi ai
We assume that the eigenvalues are sorted to satisfy |λi | ≥ |λi+1 |. Let V j denote
the j-th column of V . If b1 = 0, we choose α = 0 small enough such that with
ẑ = αV1 assumption (5.22) holds. We then obtain
M(z∗ )J(z∗ ) [I − M(z∗ )J(z∗ )] ẑ = |λ1 | M(z∗ )J(z∗ )ẑ . (5.24)
κ M(z∗ )J(z∗ )ẑ∗ ≥ M(z∗ )J(z∗ ) [I − M(z∗ )J(z∗ )] ẑ∗ ≥ |λ1 | M(z∗ )J(z∗ )ẑ∗ .
Thus we have
κlin
l
+ ε ≥ κ ≥ |λ1 | = σr (I − M(z∗ )J(z∗ )) = σr (I − M̂(z∗ )J(z∗ ))l ,
and the tangent on the Newton path according to Lemma 5.28. Theorem 5.29
guarantees that although a LISA-Newton method with larger number l of inner
iterations is numerically more expensive per outer iteration than using l = 1, the
numerical effort in the vicinity of the solution is asymptotically fully compensated
by less outer inexact Newton iterations.
A = I − M̂ J.ˆ
Lemma 5.30. Assume A ≤ κ̂ < 1. Then the following estimates hold:
ζl+1 − ζl ≤ κ̂ l ζ1 − ζ0 ,
κ̂ l
ζl − ζ∞ ≤ ζ1 − ζ0 .
1 − κ̂
Proof. Let l ≥ 1. The first assertion follows from
ζl+1 − ζl = A(ζl − ζl−1 ) ≤ Al ζ1 − ζ0 ≤ κ̂ l ζ1 − ζ0 .
Thus we obtain
∞ ∞
κ̂ l
ζl − ζ∞ ≤ ∑ ζk − ζk+1 ≤ ∑ κ̂ k ζ1 − ζ0 ≤ ζ1 − ζ0 ,
k=l k=l 1 − κ̂
5.6.2.4 Estimation of κ̂
In order to make use of Lemma 5.30 we need a computable estimate [κ̂] of κ̂. We
present three approaches which are all based on eigenvalue techniques. They differ
mainly in the assumptions on the iteration matrix A, in the required numerical
effort, and in memory consumption.
For l = 1, 2, . . . we define
δ ζl = ζl − ζl−1 .
δ ζl+1 = Aδ ζl = Al δ ζ1 .
60 5 Inexact Sequential Quadratic Programming
Thus LISA behaves like a Power Method (see, e.g., Golub and van Loan [61]). The
common idea behind all three κ̂ estimators is to obtain a good estimate for σr (A)
by approximation of a few eigenvalues during LISA. Based on Theorem 5.29 we
expect σr (A) to be a good asymptotic estimator for the norm-dependent κ̂.
with a gap in modulus between the first and second eigenvalue. If furthermore δ ζ1
has a component in the direction of the eigenvector corresponding to μ1 we obtain
δ ζlT δ ζl+1
[κ̂]l := → σr (A) for l → ∞.
δ ζlT δ ζl
Proof. The proof coincides with the convergence proof for the Power Method. For
a discussion of the convergence we refer the reader to Wilkinson [165] and Parlett
and Poole [124]. The quotient
δ ζlT δ ζl+1 δ ζ T Aδ ζl
= lT .
δ ζl δ ζl
T δ ζ l δ ζl
Lemma 5.32 (Root κ-estimator). Let σr (A) > 0 and let δ ζ1 have a component
in the dominant invariant subspace corresponding to the eigenvalues of A with
largest modulus. Then the quotient of roots
δ ζl+1 1/l
[κ̂]l+1 := , l ≥ 1,
δ ζ1 1/l
l l 1/l 1/l
δ ζl+1 1/l A δ ζ1 1/l A δ ζ1 1/l
= ≤ = Al ,
δ ζ1 1/l δ ζ1 1/l δ ζ1 1/l
A = XΛX −1 ,
Al = XΛl X −1
we see that the columns of X corresponding to each Jordan block span a cyclic
invariant subspace of Al . There exists a constant c > 0 such that
z ≥ c zX := c X −1 z2
1/l 1/l
l
A δ ζ1 ≥ c1/l Λl z . (5.26)
2
2 m 2 2
l
Λ z =
2
∑ Jp j (λ j )z̃ j 2 ≥ Jp1 (λ1 )z̃1 2 , (5.27)
j=1
which tends to zero for l → ∞. This shows that the (1, p) entry of Ã(l) dominates
for large l. Thus Ã(l)z̃2 converges to
σr (A)1−p z̃ p > 0.
(If z̃ p = 0 we can use the same argument with the last non-vanishing component
of z̃.) Consequently we can find l0 ∈ N such that
1
σr (A)−l J p (λ )l z̃ ≥ σr (A)1−p z̃ p > 0 for all l ≥ l0 . (5.28)
2 2
We now combine equations (5.26), (5.27), and (5.28) and obtain for l ≥ l0 that
l 1/l
A δ ζ1 1/l cσr (A)1−p z̃ p
≥ σr (A) ,
δ ζ1 1/l 2 δ ζ1
5.6 Natural Monotonicity for LISA-Newton methods 63
i.e., ζl circulates backwards through all basis vectors e j and then suddenly drops
to zero. Figure 5.2 depicts iterations with varying λ and N. We observe that the
Jordan structure leads to non-monotone transient convergence behavior in the first
iterations. Only in the diagonalizable case of N = 1 is the convergence monotone.
Figure 5.2 also suggests that the Root κ-estimator of Lemma 5.32 can grossly
overestimate σr (A) in a large preasymptotic range of iterations.
λ = 0.9 λ = 0.7
103 102
102
100
Error
101
10−2
100
10−1 10−4
0 5 10 15 20 0 5 10 15 20
λ = 0.5 λ = 0.3
2
10 105
100
100
10−2
Error
10−5
−4
10
10−10
10−6
10−8 10−15
0 5 10 15 20 0 5 10 15 20
Iteration Iteration
Figure 5.2: The errors of the iterates of LISA in the Euclidean norm .2 with a Jordan
iteration matrix given by Example 3 exhibit non-monotone transient convergence behavior.
The subfigures depict different values for λ . The iterations are performed with values for
N = 1, . . . , 5, marked by •, , ∗, ◦, ×, respectively.
The eigenvalues of P are given by the complex roots of the characteristic polyno-
mial λ N − 1. Thus they all satisfy |λi | = 1. Let X ∈ GL(N) and κ ∈ [0, 1). Then
the matrix
A := κXPX −1
5.6 Natural Monotonicity for LISA-Newton methods 65
has all the eigenvalues satisfying |κλi | = κ = σr (A). Again, Theorem 5.26 yields
convergence of LISA for all starting values ζ0 . By virtue of PN = I we obtain
A jN = (κ N ) j I,
which results in monotone N-step convergence. The behavior between the first and
N-th step can be non-monotone in an arbitrary norm as displayed in Figure 5.3. If
we take instead the X-norm
zX := X −1 z2
Lemma 5.33 (Ritz κ-estimator). Let δ ζ1 have a component in the dominant in-
variant subspace corresponding to the eigenvalues of A with largest modulus. De-
fine
Z(i, j) = (δ ζi , . . . , δ ζ j )
and let R ∈ R p×p be an invertible matrix such that Z(1, p) = QR with orthonormal
Q ∈ RN×p and maximal p ≤ l. Then
Kl (A, δ ζ1 ) := span(A0 δ ζ1 , . . . , Al δ ζl ).
The Ritz values solve the following variational eigenvalue problem: Find v ∈
Kl (A, δ ζ1 ) such that
102
Error 100
10−2
10−4
10−6
0 5 10 15 20
Iterations
Figure 5.3: The errors of the iterates of LISA with a 5-by-5 matrix A = 12 XPX −1 given by
Example 4 exhibit non-monotone cyclic convergence behavior in the Euclidean norm.2
(• marks). The convergence is monotone if measured in the X-norm zX := X −1 z2 (◦
marks). We chose the matrix X to be a random symmetric matrix with condition number
100.
H := QT AQ = R−T Z(1, p)T AZ(1, p)R−1 = R−T Z(1, p)T Z(2, p + 1)R−1 .
ligible if the matrix vector products with A dominate the overall effort, which is
certainly the case in MUSCOP especially on finer spatial grids.
Remark 5.35. We further propose that one should build the matrix Q iteratively,
e.g., via an Arnoldi process with upper Hessenberg matrix R (see, e.g., Golub and
van Loan [61]). This raises a couple of further questions which would have to
be addressed and exceed the scope of this thesis: If an orthonormal basis of the
Krylov space is anyway available, is a different solver for the linear systems more
appropriate? GMRES by Saad and Schultz [136], e.g., is explicitly built on an
Arnoldi process but lacks the property of affine invariance in the residual space
and an error criterion in the variable space. Furthermore, a connection between
the nonlinear κ and a descriptive constant for convergence of the linear solver like
in Theorem 5.29 should be investigated.
Based on the κ-estimators from Section 5.6.2.4 we can adaptively control the qual-
ity of the preconditioner M(z). The procedure is as follows: Let κmax < 1 and an
integer ipre be given. If in the i-th LISA iteration
then we need to improve the quality of M(x) to decrease κ. The integer ipre is a
safeguard to discard preasymptotic estimates of κ̂ which have not come close to
the actual spectral radius of the iteration matrix yet. In#our numerical experience
with the applications that we present in Part III, κmax = 1/2 and ipre = 8 produce
reasonable results.
Depending on the type of preconditioner M(z) the improvement can consist of
different strategies: In an adaptive Simplified Newton Method, e.g., we keep M(z)
constant until condition (5.30) is satisfied which triggers a new evaluation of M
at the current iterate zk . In Chapter 6 we describe a two-grid preconditioner M(z)
which can be improved by refinement of the coarse grid if condition (5.30) holds.
minimize
n
f (x) s. t. gi (x) = 0, i ∈ E , gi (x) ≥ 0, i ∈ I ,
x∈R
with E ∪˙ I = {1, . . . , m} =: m.
The quintessence of our approach is to formulate the stationarity and primal
feasibility condition of the KKT conditions (4.2) in a function F and ensure that
dual feasibility and complementarity hold in the solution via suitable choice of M.
We treat the case with exact derivatives in Section 5.7.1 and the extension to
inexact derivatives in a LISA-Newton method in Section 5.7.2.
with Jacobian
J (z) −J2 (z)T ∇2xx L (z) −∇g(x)
J(z) = 1 = .
J2 (z) 0 ∇g(x)T 0
Now we generalize the use of an approximated inverse matrix M(z) in the step
computation to a nonlinear function J ⊕ : RN+N → RN to compute
Δz = J ⊕ (z, −F̂) instead of Δz = −M(z)F̂,
where we have dropped the iteration index k for clarity. We define J ⊕ implicitly
in two steps. The fist step consists of computation of the primal-dual solution
x, y) ∈ Rn+m of the QP
z = (
1 T T
minimize x + F̂1 − J1 (z)x + J2 (z)T y x
x J1 (z) (5.32a)
x∈R n 2
s. t. J2 (z)
x + F̂2 − J2 (z)x i = 0, i∈E, (5.32b)
J2 (z)
x + F̂2 − J2 (z)x i ≥ 0, i∈I, (5.32c)
which is not formulated in the space of increments Δz = (Δx, Δy) ∈ Rn+m but rather
in the space of variables
z = z + Δz. In the second step we reverse this transforma-
tion and obtain Δz = z − z.
Lemma 5.36. Assume that QP (5.32) at z ∈ RN has a unique solution. If ẑ =
(x̂, ŷ) ∈ RN satisfies ŷi ≥ −yi for i ∈ I then
J ⊕ (z, J(z)ẑ) = ẑ.
Proof. Let ẑ = (x̂, ŷ) ∈ RN be given and define
−J1 (z)x̂ + J2 (z)ŷ
F̂ = −J(z)ẑ = .
−J2 (z)x̂
To prove the lemma we show that J ⊕ (z, −F̂) = Δz = ẑ. With aforementioned
choice of F̂ QP (5.32) becomes
1 T T
minimize x − J1 (z)(x̂ + x) − J2 (z)T (ŷ + y) x
x J1 (z)
x∈Rn 2
s. t. J2 (z)i (
x − x̂ − x) = 0, i∈E,
J2 (z)i (
x − x̂ − x) ≥ 0, i∈I.
Its stationarity condition reads
J1 (z) (
x − x̂ − x) − J2 (z) (
y − ŷ − y) = 0.
We thus observe that z = ẑ + z is stationary and primal feasible. Dual feasibility
holds due to yi = ŷi + yi ≥ 0 for i ∈ I by assumption. Complementarity is satisfied
by virtue of x− x̂ − x = 0. Thus Δz = z − z = ẑ.
Lemma 5.36 reveals that under the stated assumptions J ⊕ operates linear on the
second argument like a generalized inverse of J.
70 5 Inexact Sequential Quadratic Programming
Theorem 5.37. Assume that αk−1 = 1 and that the solutions of QP (5.32) at
(zk−1 , −F(zk−1 )) and (zk , −F(zk )) share the same active set A and satisfy the
SOSC and the SCC. Then there exists a matrix M k and a neighborhood U of F(zk )
such that
−M k F̂ = J ⊕ (zk , −F̂) for all F̂ ∈ U.
Proof. We first notice that the solution
zk−1 of QP (5.32) at (zk−1 , −F(zk−1 )) sat-
isfies
zk = zk−1 + Δzk−1 = zk−1 .
Thus we have for all inactive inequality constraints that
yki = yk−1
i =0 and yk+1
i = yki = 0 for i ∈ m \ A
whose solution depends linearly on F(zk ) and defines the submatrix of M k corre-
sponding to primal and active dual variables. We further notice that zk+1 does not
depend on F2 (zk )m\A . Thus we can set all remaining columns n + i, i ∈ m \ A of
M k to zero. This fully defines the matrix M k .
Because the SOSC and the SCC hold, the active set A is stable under perturba-
tions (see, e.g., Robinson [132]) which yields the existence of a neighborhood U
of F(zk ) such that
given by the QP active sets with jumps between these segments. The assumption
that the reduced Jacobian given in equation (5.33) is invertible on each segment is
now as unrestrictive as the assumption of invertibility of J(zk ) for the root finding
problem F(z) = 0.
Remark 5.39. Algorithmically, the evaluation of M k is performed in the follow-
ing way: If M k is evaluated for the first time, a full QP (5.32) is solved. For all
further evaluations the active (or working) set is kept fixed and a purely equality
constrained QP is solved.
Remark 5.40. We are not aware of results how the jumps due to J ⊕ can be analyzed
within the non-local theory developed in Section 5.6 for globalization based on an
NMT. We have not yet attempted an approach to fill this gap yet, either. However,
the numerical results that we present in Part III are encouraging to undertake such
a probably difficult endeavor.
The following theorem ensures that limit points of the SQP iteration with J ⊕ are
indeed KKT points or even local solutions if SOSC holds on the QP level.
Theorem 5.41. If the SQP method with J ⊕ converges to z∗ then z∗ is a KKT point
of NLP (4.1). Furthermore, the conditions SOSC and SCC transfer from QP (5.32)
at z∗ to NLP (4.1) (at z∗ ).
z = z∗ is a solution of
i.e.,
1 T T
minimize x J1 (z∗ )
x + F1 (z∗ ) − J1 (z∗ )x∗ + J2 (z∗ )T y∗ x
x∈Rn 2
s. t. (J2 (z∗ )
x + F2 (z∗ ) − J2 (z∗ )x∗ )i = 0, i∈E,
(J2 (z∗ )
x + F2 (z∗ ) − J2 (z∗ )x∗ )i ≥ 0, i∈I.
0 = J1 (z∗ )
x + F1 (z∗ ) − J1 (z∗ )x∗ + J2 (z∗ )T y∗ − J2 (z∗ )T y = F1 (z∗ ) = ∇x L (z∗ ).
Dual feasibility and complementarity for the NLP as well as SOSC and SCC follow
directly from the QP.
72 5 Inexact Sequential Quadratic Programming
We perform the construction of a preconditioner M̂(zk ) for LISA based on Jˆk now
analogously to the construction of J ⊕ (zk , .) from J(zk ). The key point is that the
transformation now requires the sum of the current Newton and the current LISA
iterate zk + δ zkl . Dropping the index k we solve the QP
1 T T
minimize x Bx + F̂1 −C(x + δ xl ) +CT (y + δ yl ) x (5.34a)
x∈Rn 2
s. t. C x + F̂2 −C(x + δ xl ) i = 0, i∈E, (5.34b)
C x + F̂2 −C(x + δ xl ) i ≥ 0, i∈I, (5.34c)
As a guiding principle, the iterations should not be forced to lie on the subset of
feasible (possibly non-optimal) points, which satisfy Jˆ2 xk = −F̂2 for all k, i.e., the
PDE constraints are allowed to be violated in iterates away from the optimal solu-
tion. Instead, feasibility and optimality are supposed to hold only at the solution.
The presence or absence of this property defines the terms sequential/segregated
and simultaneous/all-at-once/coupled method, whereby a method with only fea-
sible iterates is called sequential or segregated. The preconditioners we present
work on formulations of the problem which lead to simultaneous iterations. From
a computational point of view, simultaneous methods are more attractive because
it is not necessary to find an exact solution of Jˆ2 xk = −F̂2 in every iteration.
This chapter is organized as follows: In Section 6.1 we give a short review
of the Newton-Picard related literature. We discuss the discretization of prob-
lem (6.1) afterwards in Section 6.2. In Section 6.3 we present the Newton-Picard
preconditioners in the framework of LISA (see Chapter 5). For the discretized
problem we discuss the cases of classical Newton-Picard projective approximation
and of a coarse-grid approach for the constraint Jacobians. The importance of the
choice of the scalar product for the projection is highlighted. We establish a mesh-
independent convergence result for LISA based on classical Newton-Picard split-
ting. In this section we also outline the fast solution of the subproblems, present
pseudocode, and analyze the computational complexity. Moreover we discuss ex-
tensions to nonlinear problems and the Multiple Shooting case in Section 6.4.
In Chapter 12 of this thesis we present numerical results for different sets of
problem and discretization parameters for the Newton-Picard preconditioners. In
addition, we compare the indefinite Newton-Picard preconditioners with a sym-
metric positive definite Schur complement preconditioner in a Krylov method set-
ting.
on [90, 147]: Lust et al. [110] successfully applied the Newton-Picard method
for computation and bifurcation analysis of time-periodic solutions of PDEs and
Burrage et al. [32] develop the notion of deflation preconditioners. To our knowl-
edge the first paper on deflation techniques is by Nicolaides [119] who explicitly
introduces deflation as a modification to the conjugate gradient method and not
as a preconditioner in order to improve convergence. Fast two-grid and multigrid
approaches for the determination of time-periodic solutions for parabolic PDEs on
the basis of a shooting approach are due to Hackbusch [77].
Mi j = ϕi ϕ j , ûi = ûϕi .
Ω Ω
It is well known that the mass matrix M is symmetric positive definite. We can
now discretize equation (6.3) with MOL: The matrix of the discretized spatial dif-
ferential operator is L = −S − Q which leads to the Ordinary Differential Equation
(ODE)
M u̇(t) = Lu(t) +U(q1 (t) · · · qnq (t))T , (6.4)
78 6 Newton-Picard preconditioners
as a basis for the discrete control space. Here χ denotes the characteristic function
of the subscript interval. We can then define the symmetric positive definite control
mass matrix N ∈ Rnq m×nq m according to
Ni j = ψi ψ j .
Σ
where Inu denotes the nu -by-nu identity matrix. If we consider the special case
q̃ = 0 we immediately observe that matrix Gu is given by the matrix exponential
Because ODE (6.4) is autonomous the matrix Gq can be composed piece by piece
on the control time grid based on the matrices ∂ Gu := exp((1/m)M −1 L) and
∂ Gq := (∂ Gu − Inu )L−1U for a single interval. We obtain
Gq = ∂ Gm−1
u ∂ Gq · · · ∂ G1u ∂ Gq ∂ G0u ∂ Gq . (6.7)
RM (M −1 L)R−1 −T −1
M = RM LRM
R−T −1
M LRM Z̃ = Z̃ Ẽ (or, equivalently, R−1 −T −1 −1
M RM LRM Z̃ = RM Z̃ Ẽ).
We define Z := R−1
M Z̃ and immediately obtain the assertions.
Now we prove a negative upper bound on the eigenvalues of M −1 L.
Lemma 6.2. There exists a grid-independent scalar μ̃¯ < 0 such that all eigenval-
ues μ̃ of M −1 L satisfy μ̃ ≤ μ̃¯ .
M −1 Lv = μ̃v,
Then we apply the generalized Friedrichs inequality [152, Lemma 2.5] which
yields a Γ-dependent constant c(Γ) > 0 such that
2 2 D min(1, δ )
μ̃ vL2 (Ω) ≤ −D ∇v2 + δ v ≤2
− v2H 1 (Ω) .
Ω Γ c(Γ)
With − v2H 1 (Ω) ≤ − v2L2 (Ω) we finally obtain the assertion for the choice μ̃¯ :=
−D min(1, δ )/c(Γ) < 0.
80 6 Newton-Picard preconditioners
Proof. The matrix Gu has the same eigenvectors as the matrix M −1 L. Thus, the
assertion is a direct consequence of equation (6.6) and Lemma 6.2 with μ̄ =
exp(μ̃¯ ) ∈ (0, 1).
We now formulate the finite dimensional linear-quadratic optimization problem
1
minimize uT0 Mu0 − ûT u0 + γqT Nq (6.8a)
u0 ∈Rnu ,q∈Rnq m 2
Proof. Due to convexity of problem (6.8), necessary optimality conditions are also
sufficient, i.e., if there exists a multiplier vector λ ∈ Rnu such that for u0 ∈ Rnu , q ∈
Rnq m it holds that
⎛ ⎞⎛ ⎞ ⎛ ⎞
M 0 (GTu − Inu )M u0 û
⎜ ⎟⎝ ⎠ ⎝ ⎠
⎝ 0 γN GTq M ⎠ q = 0 (6.9)
M(Gu − Inu ) MGq 0 λ 0
then (u0 , q, λ ) is a primal-dual optimal solution and, conversely, all optimal solu-
tions must satisfy condition (6.9). The constraint Jacobian
M Gu − Inu Gq
has full rank due to Gu − Inu being invertible by virtue of Lemma 6.3. The Hes-
sian blocks M and γN are positive definite. Thus, the symmetric indefinite linear
system (6.9) is non-singular and has a unique solution.
Matrix vector products with Gq and GTq can then be evaluated based on equa-
tion (6.7).
The main difficulty here are the large and dense Gu blocks and thus approaches
based on the paper of Bramble and Pasciak [30] and also constraint preconditioners
(e.g., Gould et al. [67]), which do not approximate the blocks containing Gu but
only the M and γN blocks, do not attack the main difficulty of the problem and
will thus be not considered further in this thesis.
We investigate two choices for G̃u : The first is based on the classical Newton-
Picard projective approximation [110] for the forward problem, the second is based
on a two-grid idea.
82 6 Newton-Picard preconditioners
To compute the inverse of G̃u − Inu we have the following lemma which we invoke
with P = V and R = M p−1V T M:
6.3 Newton-Picard for optimal control problems 83
Lemma 6.5. Let G̃u ∈ Rnu ×nu , P ∈ Rnu ×p , R ∈ R p×nu , and E ∈ R p×p satisfy
Computation of the inverse thus only needs the inversion of the small p-by-p
matrices E − I p and M p . For inversion of G̃Tu − Inu we obtain similar to equa-
tion (6.10)
and consequently
−1
(G̃Tu − Inu )−1 = M(G̃u − Inu )M −1 = M(G̃u − Inu )−1 M −1 . (6.13)
on a coarse grid (cf. Hackbusch [77]), while the remaining computations are per-
formed on the fine grid. Let P and R denote the prolongation and restriction matri-
ces between the two grids and let superscripts c and f denote coarse and fine grid,
respectively. Then, Gfu is approximated by
i.e., we first project from the fine grid to the coarse grid, evaluate the exact Gcu on
the coarse grid, and prolongate the result back to the fine grid. Note that in contrast
to classical Newton-Picard, E is now not a diagonal matrix.
We use conforming grids, i.e., the Finite Element basis on the coarse grid can
be represented exactly in the basis on the fine grid. Thus, the prolongation P can
f c
be obtained by interpolation. Let uf ∈ Rnu , uc ∈ Rnu and define
nf nc
uf = ∑i=1
u
ufi ϕif ∈ H 1 (Ω), uc = ∑i=1
u
uci ϕic ∈ H 1 (Ω).
We define the restriction R in an L2 sense, such that given uf on the fine grid we
look for the projector R : uf → uc such that
or, equivalently,
M c uc = PT M f uf .
We then obtain
R = (M c )−1 PT M f .
which can be computed by only an inversion of a ncu -by-ncu matrix from the coarse
grid and the inversion of the coarse grid mass matrix in the restriction operator.
We obtain an expression for the inverse of the transpose similar to equation (6.13)
via
−1 −1
(G̃fu )T − Infu = M G̃fu − Infu M −1 .
6.3 Newton-Picard for optimal control problems 85
In this section we show that for problem (6.1), LISA (6.2) with classical Newton-
Picard preconditioning converges with a grid-independent contraction rate.
For the proof of Theorem 6.7 we need the following lemma. The lemma asserts
the existence of a variable transformation which transforms the Hessian blocks to
identity, and furthermore reveals the structure of the matrices on the subspaces of
fast and slow modes.
⎛ ⎞
Inu −p 0 0 0 −Inu −p
⎜ 0 Ip EV − I p 0 ⎟
⎜ 0 ⎟
˜ =⎜
T T JT ⎜ 0 0 γN GTq MV T ⎟
Gq MW ⎟ ,
⎜ ⎟
⎝ 0 EV − I p V T MGq 0 0 ⎠
−Inu −p 0 W T MGq 0 0
⎛ ⎞
0 0 0 0 −EW
⎜ 0 0 ⎟
⎜ 0 0 0 ⎟
⎜ ⎟
T T ΔJT = ⎜ 0 0 0 0 0 ⎟.
⎜ ⎟
⎝ 0 0 0 0 0 ⎠
−EW 0 0 0 0
Proof. The existence of the matrices V and W , as well as conditions (i) and (ii)
follow from Lemma 6.1. To show (iii), we choose
⎛ ⎞
W V 0 0 0
⎜ ⎟
T =⎝0 0 Inq m 0 0 ⎠.
0 0 0 V W
86 6 Newton-Picard preconditioners
T T ΔJT
⎛ ⎞⎛ ⎞
0 0 ΠT − Inu GTu M W V 0 0 0
= TT⎝ 0 0 0 ⎠⎜ ⎟
⎝ 0 0 I nq m 0 0 ⎠
MGu (Π − Inu ) 0 0 0 0 0 V W
⎛ ⎞
W T 0 0
⎜V T 0 ⎛ ⎞
⎜ 0 ⎟⎟ 0 0 0 0 −MGuW
⎜ ⎟⎝ ⎠
= ⎜ 0 Inq m 0 ⎟ 0 0 0 0 0
⎜ ⎟
⎝ 0 0 V ⎠ −MGuW 0 0 0
T 0
0 0 WT
⎛ ⎞
0 0 0 0 −EW
⎜ 0 0 0 0 0 ⎟
⎜ ⎟
⎜ ⎟
= ⎜ 0 0 0 0 0 ⎟.
⎜ ⎟
⎝ 0 0 0 0 0 ⎠
−EW 0 0 0 0
T T JT
˜
⎛ ⎞⎛ ⎞
M 0 (MVV T GTu − Inu )M WV 0 0 0
⎜ ⎟⎜ ⎟
= TT⎝ 0 γN GTq M ⎠ ⎝ 0 0 I nq m 0 0 ⎠
M(GuVV M − Inu ) MGq
T 0 0 0 0 VW
⎛ ⎞
WT 0 0 ⎛ ⎞
⎜ V T 0 0 ⎟ MW 0 MV (EV − I p ) −MW
⎜ ⎟ MV
⎜ ⎟⎜ ⎟
= ⎜ 0 Inq m 0 ⎟ ⎝ 0 0 γN GTq MV GTq MW ⎠
⎜ ⎟
⎝ 0 0 V ⎠ −MW MV (EV − I p ) MGq
T
0 0
0 0 WT
⎛ ⎞
Inu −p 0 0 0 −Inu −p
⎜ 0 Ip EV − I p 0 ⎟
⎜ 0 ⎟
⎜ ⎟
=⎜ 0 0 γN Gq MV Gq MW ⎟ .
T T
⎜ ⎟
⎝ 0 EV − I p V MGq
T 0 0 ⎠
−Inu −p 0 W T MGq 0 0
6.3 Newton-Picard for optimal control problems 87
then LISA (6.2) with Newton-Picard preconditioning applied to problem (6.1) con-
verges with a contraction rate of at most μ p+1 /μ1 .
Proof. Due to Theorem 5.26 the contraction rate is given by the spectral radius
σr (J˜−1 ΔJ) = σr (T −1 J˜−1 T −T T T ΔJT ). We obtain the eigenvalue problem
−1
T T JT
˜ T T ΔJT v − σ v = 0,
˜ v = 0,
−T T ΔJT v + σ T T JT
with the matrices given by Lemma 6.6. We prove the theorem by contradiction.
Assume that there is a complex eigenpair (v, σ ) such that |σ | ≥ μ p+1 /μ1 > μ p+1 .
Division by σ yields the system
where v was divided into five parts v1 , . . . , v5 corresponding to the blocks of the
system. Because |σ | > μ p+1 we obtain invertibility of Inu −p − (1/σ )EW and thus
we can eliminate
v5 = (Inu −p − (1/σ )EW )−1 v1 , v4 = (I p − EV )−1 v2 , (6.17a)
v2 = (I p − EV )−1 V T MGq v3 , v1 = (Inu −p − (1/σ )EW )−1 V T MGq v3 . (6.17b)
Substituting these back in equation (6.16c) yields
We denote the complex valued matrix on the left hand side with A(σ ). The final
step of the proof consists of showing that A(σ ) is invertible if γ is large enough.
Since M and N are positive definite matrices they have Cholesky decompositions
M = RTM RM , N = RTN RN
with invertible RM ∈ Rnu ×nu , RN ∈ Rnq m×nq m . If we define
−1
I p − EV 0 V
B(σ ) := RTM RM Gq R−1
N ∈C
nq m×nq m
0 Inu −p − σ −1 EW W
we obtain
γ −1 R−T −1 −1
N A(σ )RN = Inq m + γ B(σ ) B(σ ).
T
≤ C2 + sup Ḡq q
nq m L2 (Ω)
q=∑i=1 qi ψi
qL (Σ) =1
2
≤ C2 + sup Ḡq q
L2 (Ω)
q∈L2 (Σ)
qL (Σ) =1
2
≤ C1 +C2 .
If now γ > (C1 + C2 )2 /(1 − μ1 )2 then γ −1 B(σ )T B(σ )2 < 1 and thus A(σ ) is
invertible. It follows that v3 = 0, which implies v = 0 via equations (6.17). Thus,
(v, σ ) cannot be an eigenpair.
The main result of this section is now at hand:
Corollary 6.8. The asymptotic convergence rate of LISA with classical Newton-
Picard preconditioning on the model problem is mesh-independent, provided γ is
large enough.
Proof. For finer and finer discretizations the largest p + 1 eigenvalues of M −1 L
converge. Thus, also the eigenvalues μ1 and μ p+1 of Gu converge to some μ̄1 < μ̃
and μ̄ p+1 ≤ μ̄1 , with μ̄ given by Lemma 6.3. We construct Ḡq as the infinite
dimensional counterpart to Gq , i.e., Ḡq maps controls in L2 (Σ) to the end value of
the heat equation (6.1b)–(6.1c) with zero initial values for the state. This operator
is continuous (see, e.g., Tröltzsch [152]). Let ε > 0. We can assume that Gq
satisfies the discretization error condition (6.15) with C2 = ε for a fine enough and
also for all finer discretizations. Define
γ̄ = (C1 + ε)2 /(1 − μ̄)2 .
Theorem 6.7 yields that if γ > γ̄ then the asymptotic convergence rate of LISA is
below the mesh-independent bound μ̄ p+1 /μ̄1 .
90 6 Newton-Picard preconditioners
We remark here that our numerical experience suggests the conjecture that the
contraction rate is actually μ p+1 instead of μ p+1 /μ1 .
Solving the last block-row for u0 and the first for λ , the second block-row becomes
−1 −1 −1
Hq = r2 − GTq G̃Tu − Inu r1 − M G̃u − Inu M r3 =: r̃, (6.19)
Lemma 6.9. Assume there exists a linear operator Ḡu : L2 (Σ) → L2 (Ω) which
satisfies assumptions (6.14) and (6.15). Then the spectral condition number of
matrix N −1 H is bounded by
(C1 +C2 )2
cond2 (N −1 H) ≤ 1 + ,
γ(1 − μ̄)2
n m
and define q = ∑i=1
q
qi ψi ∈ L2 (Σ). We obtain
2
σ qT Nq = qT Hq = γqT Nq + RM (G̃u − Inu )−1 R−1
M RM Gq q 2 (6.20a)
"
≥ γqT Nq ⇒ σ ≥ γ,
2 2 (6.20b)
≤ γqT Nq + RM (G̃u − In )−1 R−1 RM Gq q ,
u M 2 2
By virtue of Lemma 6.3 the largest singular value of G̃u − Inu is bounded by 1 − μ̄
and thus we obtain
RM (G̃u − In )−1 R−1 2 ≤ 1/(1 − μ̄)2 . (6.21)
u M 2
We now combine inequalities (6.20), (6.21), and (6.22) to obtain the assertion.
As a consequence of Lemma 6.9 we obtain that the number of required PCG
iterations bounded by a grid-independent number. In our numerical experience
10–20 PCG iterations usually suffice for a reduction of the relative residual to
10−6 .
Solving for u0 and λ is then simple:
−1 −1 −1
u0 = G̃u − Inu M r3 − Gq q , λ = M −1 G̃Tu − Inu (r1 − Mu0 ) .
Note that once Gq and G̃u (in a suitable representation) have been calculated no
further numerical integration of the system dynamics is required.
6.3.5 Pseudocode
In this section we provide pseudocode to sketch the implementation of the pro-
posed Newton-Picard preconditioners. We focus on the case nq m nu in which
it is economical to solve equation (6.19) by forming H and using Cholesky de-
composition, but we also discuss implementation alternatives for the case of large
nq m. We use a Matlab® oriented syntax here and assume that the reader is familiar
with linear algebra routines in Matlab® . For readability purposes we further as-
sume that matrices L, M, Lc , M c , P, R, E, X, N and dimensions nu , nq , m are globally
92 6 Newton-Picard preconditioners
For the two-grid version, we assume that the prolongation P and restriction R
are given. Because the occurring matrices are small, we can employ the LAPACK
methods in eig instead of IRAM in eigs.
Now P, R, E, X are known and we can formulate matrix vector and matrix trans-
pose vector products with G̃u .
In the same way we can evaluate matrix vector and matrix transpose vector
products with the inverse of G̃u − Inu according to Lemma 6.5.
6.3 Newton-Picard for optimal control problems 93
We want to remark that we take the liberty to call the four previous functions
also with matrix arguments. In this case the respective function is understood to
return a matrix of the same size and to be evaluated on each column of the input
matrix. For the computation of matrix vector products with Gu and Gq we define
an auxiliary function which integrates ODE (6.4) for given initial state and control
variables. The control coefficients are constant in time.
Based on the previous function we can now assemble matrix Gq . There are
alternative ways for the assembly. We have chosen an approach for the case that
large intervals for dG can be efficiently and accurately computed through adaptive
step size control as in, e.g., ode15s.
Function Gq = computeGq
output: Matrix Gq ∈ Rnu ×nq m
for j = 1 : nq do
Gq (:, j + nq * (m − 1)) = dG(1 / m, 0, Inq (:, j));
for i = 1 : m − 1 do
for j = 1 : nq do
Gq (:, j + nq * (i − 1)) = dG(1 − i / m, Gq (:, j + nq * (m − 1)), 0);
We can alternatively compute matrix vector and matrix transpose vector prod-
ucts with Gq via the following functions. For the transpose we exploit the expres-
sion
i T i −1
∂ Gu ∂ Gq = U T L−1 ∂ GTu − Inu ∂ GTu = U T L−1 M ∂ Gi+1 u − ∂ Gu M .
i
94 6 Newton-Picard preconditioners
Function u1 = Gq(q)
input : q ∈ Rnq m
output: u1 = Gq q ∈ Rnu
u1 = zeros(nu , 1);
for i = 0 : m − 1 do
u1 = dG(1 / m, u1 , q(i * nq + (1 : nq )));
Function q = GqT(λ )
input : λ ∈ Rnu
output: q = GTq λ ∈ Rnq m
q = zeros (nq * m,1); λ̃ + = M \ λ ;
for i = m − 1 : −1 : 0 do
λ̃ = λ̃ + ; λ̃ + = dG(1 / m, λ̃ , 0);
q(i * nq + (1 : nq )) = U T * (L \ (M * (λ̃ + − λ̃ )));
We can also formulate functions for matrix vector and matrix transpose vector
products with Gu .
Function u1 = Gu(u0 )
input : u0 ∈ Rnu
output: u1 = Gu u0 ∈ Rnu
u1 = dG(1, u0 , 0);
Function u0 = GuT(λ )
input : λ ∈ Rnu
output: u0 = GTu λ ∈ Rnu
u0 = M * dG(1, M \ λ , 0);
We can finally state the function for a matrix vector product with the symmetric
indefinite matrix J.ˆ For readability we split up the argument and result into three
subvectors.
6.3 Newton-Picard for optimal control problems 95
At last we present pseudocode for matrix vector products with the precondi-
tioner J˜−1 . Again, we split up the argument and result into three subvectors.
Function [u, q, λ ] = iJp(r1 , r2 , r3 )
input : r1 ∈ Rnu , r2 ∈ Rnq m , r3 ∈ Rnu
output: [u; q; λ ] = J˜−1 [r1 ; r2 ; r3 ] ∈ Rnu +nq m+nu
q = RH \ (RTH \ (r2 − iGupTmI(GTq * (r1 − M * iGupmI(r3 )))));
u = iGupmI(r3 − Gq * q);
λ = iGupTmI(r1 − M * u);
We can also substitute the functions Gq and GqT for the occurrences of Gq and
GTq in J and iJp.
the contraction rate of the outer iteration is bounded by κ̃. Thus, we cansolve
the optimization problem (4.1) up to a tolerance εO > 0 within O(Jz ˆ 0 + F̂ /εO )
iterations.
We now count the number of system integrations per iteration under the as-
sumption that we perform nH inner PCG iterations per outer iteration. For the
evaluation of matrix vector products with Jˆ we need two system integrations for
maxtrix vector products with (Gu Gq ) and its transpose. Concerning the matrix
vector product with the preconditioner J˜−1 we observe that multiplications with
G̃u and (G̃u − Inu )−1 do not need any system integrations. However, the setup
of r̃ in equation (6.19) requires one integration for a matrix vector product with
Gq . Furthermore, each inner PCG iteration requires two additional simulations for
matrix vector products with Gq and its transpose. Thus we need 3 + 2nH system
simulations per outer iteration which yields an optimal complexity of O(1/εO )
system integrations for the solution of the optimization problem (4.1) up to toler-
ance εO .
When performed this way the additional linear algebra consists of matrix vector
multiplications with sparse matrices, lower order vector computations, and dense
linear algebra for system sizes bounded by a factor of the grid-independent num-
bers p or ncu , respectively.
In the case of classical Newton-Picard approximation we need to add the com-
plexity of IRAM for the one-time determination of the dominant subspace spanned
by V . A detailed analysis of the numerical complexity for this step is beyond the
scope of this thesis. Suffice it that based on Saad [135, Theorem 6.3 and Cheby-
shev polynomial approximation (4.49)] together with the approximation μnu ≈ 0
we assume that the tangent of the angle between the p-th eigenvector of Gu and
the l-th Krylov subspace decreases linearly in l with a factor depending on the
ultimately grid-independent ratio μ p /μ p+1 which we assume to be greater than
one. We need one matrix vector product with Gu per Arnoldi iteration. Because
this computation is only needed once independently of εO the asymptotic com-
plexity O(1/εO ) does not deteriorate. It does, however, have an effect for practical
computations (see Section 12.3) and can easily dominate the overall cost of the
algorithm already for modest values of p.
We have also found the approach with explicit solution of equation (6.19) via
Cholesky decomposition of H beneficial for the practical computations presented
in Chapter 12. Although we obtain a one-time cubic complexity in nq m and a
square complexity in nq m per outer iteration, the runtime can be much faster than
iterative solution of equation (6.19) because per outer iteration only two system
integrations are required instead of 3 + 2nH .
6.4 Extension to nonlinear problems and Multiple Shooting 97
We want to close this section with the remark that the optimal choice of p, κ̃,
εM , and εH is a complex optimization problem which exceeds the scope of this
thesis.
with fixed xc can be successfully solved with Newton-type methods (see Chapter 5)
In most cases a cheap approximation Gk ≈ ∂∂xgs (xsk , xc ) with linear contraction rate
of, say, κ = 0.8 is already good enough to produce an efficient numerical method.
In general, cheaper computation of the action of G−1 k on the residual compared
to the action of ( ∂∂xgs )−1 must compensate for the loss of locally quadratic con-
vergence of a Newton method to obtain an overall performance gain within the
desired accuracy. It is a tempting idea to use the same Jacobian approximations
Gk from the Newton-type method in an inexact SQP method for the optimization
problem with the same constraint
From this point of view we call problem (7.1) the forward problem of optimization
problem (7.3) and we will refer to the variables xc as control or design variables
and to xs as state variables.
Using (inexact) SQP methods which do not satisfy g = 0 in every iteration
for (7.3) is usually called simultaneous, or all-at-once approach and has proved
to be successful for several applications, e.g., in aerodynamic shape optimiza-
tion Bock et al. [27], Hazra et al. [80], chemical engineering Potschka et al. [130],
or for the model problem (6.1) in Chapter 6. Any inexact SQP method for equality
constrained problems of the form (7.3) is equivalent to a Newton-type method for
the necessary optimality conditions
∇x L(x, y) = 0, g(x) = 0,
we obtain preconditioners which are contractive, i.e., the eigenvalues of the pre-
conditioned system lie in a ball around 1 with radius less than 1. The contraction
property suggests the use of a simultaneous one-step approach. However, we can
show that contraction for the forward problem is neither sufficient nor necessary
for convergence of the simultaneous one-step method.
The structure of this chapter is the following: Based on the Local Contraction
Theorem 5.5 we present in Section 7.1 illustrative, counter-intuitive examples of
convergence and divergence for the forward and optimization problem which form
the basis for the later investigations on recovery of convergence. We continue
with presenting a third example and three prototypical subproblem regularization
strategies in Section 7.2 and perform an asymptotic analysis for large regulariza-
tion parameters in Section 7.3. We also show de facto loss of convergence for one
of the examples and compare the regularization approaches to Griewank’s One-
Step One-Shot projected Hessian preconditioners.
min 1 T
2 x Hx, s.t. (A1 A2 )x = 0 (7.5)
x=(xs ,xc )∈Rn
H AT H T
A
A = A1 A2 , = A
A 1 A2 , K= , =
K .
A 0
A 0
According to the Local Contraction Theorem 5.5 and Remark 5.6 the choice of
1 = I leads to a fast linear contraction rate for the forward problem of
A
−1 A1 ) = σr (I − A1 ) ≈ 0.077 < 1.
κF = σr (I − A1
However, for the contraction rate of the inexact SQP method with exact Hessian
and exact constraint derivative with respect to xc , we get
Thus the full-step inexact SQP method does not have the property of linear local
convergence. In fact it diverges if the starting point z0 has a non-vanishing compo-
nent in the direction of any generalized eigenvector of I − K −1 K corresponding to
a Jordan block with diagonal entries greater than 1.
We obtain
−1 A1 ) ≈ 1.20 > 1,
κF = σr (I − A −1 K) ≈ 0.014 < 1,
κO = σr (I − K
1
i.e., fast convergence of the method for the optimization problem but divergence
of the method for the forward problem. From these two examples we see that in
general only little can be said about the connection between contraction for the
forward and the optimization problem.
7.2 Subproblem regularization 103
where μ > 0 is some chosen weighting factor for relative scaling of primal and
dual variables. Based on
Z(λ ) = (λ I − Gy )−1 Gu , H(λ ) = (Z(λ )T , I)Nxx (Z(λ )T , I)T ,
we can numerically verify that the projected Hessian preconditioners H(λ ), λ ∈
[−1, 1], do not restore contraction. The lowest spectral radius of the iteration ma-
trix is 1.17 for λ = −0.57 and larger for all other values (compare Figure 7.1).
We now investigate three different modifications of the subproblems which do
not alter the Jacobian blocks of the KKT systems. These modifications are based
on
κO = σr (I − K−1 K) = σr (K
−1 (K − K)),
which suggests that small eigenvalues of K might lead to large κO . Thus we regu-
−1
larize K such that the inverse K does not have large eigenvalues in the directions
of inexactness of ΔK = K − K.
We consider three prototypical regularization methods here which all add a pos-
The regularizing matrices are
itive multiple α of a matrix Λ to K.
⎛ ⎞ ⎛ ⎞ ⎛ ⎞
I 0 0 I 0 0 0 0 0
Λp = ⎝0 I 0⎠ , Λpd = ⎝0 I 0 ⎠ , Λhp = ⎝0 I 0⎠ ,
0 0 0 0 0 −I 0 0 0
104 7 One-shot one-step methods and their limitations
Contraction κ
20
10
0
-1 -0.5 0 0.5 1
λ
Figure 7.1: Contraction rates for One-Step One-Shot preconditioning with the projected
Hessians H(λ ). The two gaps are due to the two eigenvalues of Gy rendering λ I − Gy
singular.
where the subscripts stand for primal, primal-dual, and hemi-primal (i.e., only in
the space of design variables), respectively.
where is the current Lagrange gradient and r is the current residual of the equality
constraint. We use a nullspace method to solve the α-dependent system (7.6). Let
matrices Y ∈ Rn×m and Z ∈ Rn×(n−m) have the properties
= 0, ZTZ 0
AZ (Z Y )T (Z Y ) = , det(Y Z) = 0.
0 Y TY
Lemma 7.1. Under the general assumptions of Section 7.3 the solution of equa-
tion (7.6) for the primal regularization for large α is asymptotically given by
+ r + (1/α)ZZ + H A
Δx(α) = −A + r − + o(1/α),
A
Δy(α) = α(A + )T (H A
T )−1 r + (A + r − ) + o(1).
is invert-
Proof. From the second block-row of equation (7.6) and the fact that AY
−1
ible due to A having full rank we obtain p = −(AY ) r. Premultiplying the first
block-row of equation (7.6) with Z T from the left yields the α-dependent equation
Z T HY p + Z T HZq + αZ T Zq + Z T = 0. (7.7)
It holds that Φ(0, 0) = 0 and ∂∂Φq (0, 0) = Z T Z is invertible, as Z has full rank.
Therefore the Implicit Function Theorem yields the existence of a neighborhood
U ⊂ R of 0 and a continuously differentiable function q̄ : U → Rm such that q̄(0) =
0 and
Ψ(β ) := Φ(q̄(β ), β ) = 0 ∀β ∈ U.
∂ Φ dq̄
Using 0 = dΨ
dβ = ∂ q dβ + ∂∂ Φ
β and Taylor’s Theorem we have
)−1 = A
and obtain Y (AY T QQ−1 (A AT )−1 = A
+ , which yields the first assertion of
the Lemma.
For the corresponding dual solution we multiply the first block-row of equa-
tion (7.6) with Y T from the left to obtain
)T Δλ (α) +Y T = 0.
Y T (H + αI)Δx(α) + (AY
Lemma 7.2. Under the general assumptions of Section 7.3 the solution of equa-
tion (7.6) for the primal-dual regularization with large α is asymptotically given
by
Δx(α) 1 1 −
= − Λpd + o(1/α) = + o(1/α).
Δy(α) α r α r
7.3 Analysis of the regularized subproblems 107
It holds that
∂Φ + Λpd , ∂Φ
Φ(0, 0) = 0, = βK (0, 0) = Λpd .
∂z ∂z
The Implicit Function Theorem and Taylor’s Theorem yield the assertion.
Consider the limit case
Δx(α) 1
= − Λpd
Δy(α) α r
If all the
and the corresponding local contraction rate κpd = σr (I − (1/α)Λpd K).
are larger
real parts of the (potentially complex) eigenvalues of the matrix Λpd K
than 0, contraction for large α can be recovered although contraction may be ex-
tremely slow, leading to de facto loss of convergence.
Lemma 7.3. Under the general assumptions of Section 7.3 the solution of equa-
tion (7.6) for the hemi-primal regularization is for large α asymptotically given
by
−1
−A−1 r A r
Δx(α) = 1 + (1/α)ZZ T H 1 − + o(1/α), (7.9a)
0 0
−1 T −1
A r
A
Δy(α) = 1 H 1 − + o(1), (7.9b)
0 0
T
−1 A2 )T I
with the choice Z = (−A 1 A2 )T = A
and Y = (A T .
1
108 7 One-shot one-step methods and their limitations
Hence xck = xc∗ stays constant and xk converges to a feasible point x∗ with the con-
traction rate κF of the Newton-type method for problem (7.1). For the asymptotic
step in the dual variables we then obtain
T
−T
Δyk = −A −T −T
1 (∇x f (xk ) + ∇x g(xk )yk ) + A1 0 H A1 0 rk .
For the convergence of the coupled system with xk and yk let us consider the Jaco-
bian of the iteration (xk+1 , yk+1 ) = T (xk , yk ) (with suitably defined T )
dT I−A −1 ∇x g(x)T 0
= 1 .
d(x, y) ∗ I−A −T ∇x g(x)
1
2.5
Contraction κO 2
1.5
1
0.5
10−4 10−3 10−2 10−1 100 101 102 103 104 105 106
Regularization α
1.005
Contraction κO
0.995
10−4 10−3 10−2 10−1 100 101 102 103 104 105 106
Regularization α
Figure 7.2: Divergence of the primal regularization (− · ) and de facto loss of convergence
for primal-dual (−−) and hemi-primal (−) regularization for example Ex3 depending on
the regularization value α. The lower diagram is a vertical close-up around κO = 1 of the
upper diagram.
Thus the primal-dual iterates converge to a point which is feasible and stationary
with respect to xs but not necessarily to xc . Taking α large but finite we see that the
hemi-primal regularization damps design updates while correcting state and dual
variables with a contraction of almost κF .
In Section 8.1 of this chapter we describe the typical multiple shooting struc-
ture. We highlight the additional Newton-Picard structures in Section 8.2. Then
we present the exploitation of these structures for the elimination of the discretized
PDE states in a rather general way in Section 8.3 and develop a particular Newton-
Picard Hessian approximation which fits well in the condensing framework in Sec-
tion 8.4. Based on the introduced notation we end this chapter with a result of
scaling invariance of the Newton-Picard LISA-Newton method in Section 8.5.
In their seminal paper, Bock and Plitt [25] have described a condensing technique
for the quadratic subproblems arising in an SQP method for Direct Multiple Shoot-
ing. We specialize this approach for the case that either fixed initial values or
boundary value equality constraints are posed on the PDE states. In Section 8.2
we extend the approach with the exploitation of the Newton-Picard structure in the
approximated Hessian and Jacobian matrices (see Chapter 6).
Recall the discretized NLP (3.3) on level l. To avoid notational clutter we drop
the discretization level index l. The NLP (3.3) then reads
minimize
n
Φ(snMS , vnMS ) (8.1a)
(qi ,si ,vi )i=0
MS
where we have split up the boundary condition rb into two parts rbs and rbv . We
abbreviate derivatives that occur in the remainder of this chapter according to
∂ ui ∂ ui ∂ ui
Giq := , Gis := , Giv := ,
∂ qi−1 ∂ si−1 ∂ vi−1
∂ vi ∂ vi ∂ vi
Hqi := i−1 , Hsi := i−1 , Hvi := i−1 ,
∂q ∂s ∂v
∂ ri ∂ ri re
Ri,i
q := , Ri,i
v := , Re := n .
∂ qi−1 ∂ vi−1 ∂ v MS
The elimination is based on the s0 -dependent part of boundary condition (8.1b) and
on the matching conditions (8.1d). Thus we also shuffle the constraint order such
8.1 Multiple shooting structure 113
that they are the first ns (nMS + 1) constraints. We are lead to consider NLP (8.1)
in the ordering
minimize f (x1 , x2 ) (8.2a)
(x1 ,x2 )∈Rn1 +n2
s. t. gi (x1 , x2 ) = 0, i ∈ E1 , (8.2b)
gi (x1 , x2 ) = 0, i ∈ E2 , (8.2c)
gi (x1 , x2 ) ≥ 0, i∈I, (8.2d)
where |E1 | = n1 . In Section 8.2 we describe how to exploit gi , i ∈ E1 , for partial re-
duction on the QP level. Now x1 contains only the discretized PDE state variables
and gi , i ∈ E1 , comprises the boundary and matching conditions (8.1b) and (8.1d).
Then the compound derivative of the constraints has the form
⎛ ⎞
C11 C12
C = ⎝C21 C22 ⎠
C31 C32
⎛ ⎞
−I Rbss Rbsv
⎜ 1 ⎟
⎜ Gs −I G1v G1q ⎟
⎜ ⎟
⎜ .. .. .. .. ⎟
⎜ . . . . ⎟
⎜ ⎟
⎜ G
nMS
−I G
nMS
G
nMS ⎟
⎜ s v q ⎟
⎜ R b −I R b ⎟
⎜ vs vv ⎟
⎜ H1 Hv1 −I Hq1 ⎟
⎜ s ⎟
⎜ ⎟
=⎜ ⎜
..
.
..
.
..
.
..
. ⎟.
⎟
⎜ ⎟
⎜ Hs
nMS
Hv
nMS
−I Hq
nMS
⎟
⎜ ⎟
⎜ I −I ⎟
⎜ ⎟
⎜ i,1 i,1 ⎟
⎜ Rv Rq ⎟
⎜ ⎟
⎜ . .. . .. ⎟
⎜ ⎟
⎜ i,nMS i,nMS ⎟
⎝ Rv Rq ⎠
e,nMS
Rv
We want to stress that contrary to the appearance the block C11 is several orders of
magnitude larger than the blocks C22 and C32 on fine spatial discretization levels.
The next lemma shows that under suitable assumptions C11 is invertible. We use
products of non-commuting matrices where the order is defined via
nMS 0
∏ Gis := Gns MS · · · G1s and ∏ Gis = I by convention.
i=1 i=1
114 8 Condensing
n
Lemma 8.1. Let MB = I − (∏i=1 MS
Gis )Rbss . If MB is invertible so is C11 and the
−1
inverse is given by C11 =
⎛ ⎞ ⎛ ⎞
I
0
i Rb ⎛ ⎞ I
⎜ ∏ G s ss ⎟ I ⎜ 1 ⎟
⎜ i=1 ⎟ ⎜ ⎟
⎟ ⎜ ∏ Gs I
i
⎜ .. .. ⎟⎜ .. ⎟
⎜ . . ⎟⎜ . ⎟ ⎜ i=1 ⎟
−⎜ n −1 ⎟ ⎜ ⎟⎜ . .. .. ⎟.
⎜ ⎟⎝ I ⎠ ⎜ .. . . ⎟
⎜ ⎟ ⎜ ⎟
MS
⎝ I ∏ Gs Rss ⎠
i b
−1 ⎝nMS nMS ⎠
i=1 MB ∏ sG i ···
∏ s G i I
I i=1 i=nMS
Proof. We premultiply C11 with the matrices in the assertion one after the other to
obtain
⎛ ⎞
⎛ ⎞ 0
⎛ ⎞
⎜−I ∏ Gs Rbss ⎟
i
I
⎜ G1 I ⎟ −I1 Rbss ⎜ i=1 ⎟
⎜ s ⎟ ⎜G −I ⎟ ⎜ 1 ⎟
⎜ . ⎟⎜ s ⎟ ⎜ −I ∏ Gis Rbss ⎟
⎜ .. .. ..
. . ⎟⎜ . . ⎟=⎜ ⎟,
⎜ ⎟⎝ .. .. ⎠ ⎜ i=1 ⎟
⎝nMS ⎠ ⎜ . . ⎟
nMS
∏ Gs · · · G s I
i nMS
Gs −I ⎝ .. .
. ⎠
i=1 −MB
⎛ ⎞ ⎛ ⎞
0 0
⎛ ⎞ −I ∏ G i Rb −I ∏ G i Rb
I ⎜ s ss ⎟ ⎜ s ss ⎟
⎜ i=1 ⎟ ⎜ i=1 ⎟
⎜ .. ⎟⎜ 1 ⎟ ⎜ 1 ⎟
⎜ . ⎟⎜ −I ∏ Gis Rbss ⎟ ⎜ −I ∏ Gis Rbss ⎟
⎜ ⎟⎜ ⎟ = ⎜ ⎟,
⎝ I ⎠⎜ i=1 ⎟ ⎜ i=1 ⎟
⎜ .. .. ⎟ ⎜ .. .. ⎟
−1 ⎝ . . ⎠ ⎝ . . ⎠
MB
−MB −I
⎛ ⎞⎛ ⎞
0 0
⎜−I − ∏ Gis Rbss ⎟ ⎜−I ∏ Gis Rbss ⎟
⎜ i=1 ⎟⎜ i=1 ⎟
⎜ .. .. ⎟⎜ 1 ⎟
⎜ . . ⎟⎜ −I ∏ G i Rb ⎟
ss ⎟ = I.
⎜ n −1 ⎟ ⎜ s
⎜ ⎟⎜ i=1 ⎟
⎜ ∏ Gis Rbss ⎟ ⎜ ⎟
MS
−I − .. ..
⎝ i=1
⎠ ⎝ . . ⎠
−I −I
This proves the assertion.
The assumption of invertibility of MB is merely that one is not an eigenvalue of
the matrix nMS i b
GB := ∏i=1 Gs Rss
which coincides with the monodromy matrix of the periodicity condition (8.1b) in
the solution of NLP (8.1).
8.2 Newton-Picard structure 115
RP = I. (8.3)
We now approximate the blocks in C. Let hatted matrices (ˆ) denote the evaluation
of a matrix on either a coarse grid (two-grid variant) or on the dominant subspace
(classical variant). Then we assemble the approximations () from the hatted ma-
trices preceded and/or succeeded by appropriate prolongation and/or restriction
matrices according to
The following lemma shows that the approximation of MB can be cheaply evalu-
ated and inverted because it only involves operations on the coarse grid or on the
low-dimensional dominant subspace.
Lemma 8.2. Let
nMS nMS
ĜB := ∏ Ĝis R̂bss , B :=
G ∏ is
G Rbss ,
i=1 i=1
M̂B := I − ĜB , M B .
B := I − G
and thus
B = I − PĜB R.
M
116 8 Condensing
The calculation
% &
−1 = I − PĜB R I − P I − I − ĜB −1 R
BM
M B
% −1 &
= I − PĜB R − P I − ĜB I − I − ĜB R
= I − P ĜB + I − ĜB − I R = I
yields the assertion.
The structure of C and the assertion of Lemma 8.1 is also preserved if we use the
proposed approximations. Thus Lemma 8.2 suggests that it is possible to compute
the inverse of the approximation of the large block C11 in a cheap way. We prove
this supposition in Theorem 8.3 but before we need to introduce another notational
convention, the Kronecker product of two matrices. We only use the special case
where the left-hand factor is the identity and thus we have for an arbitrary matrix
A that
⎛ ⎞
A
⎜ .. ⎟
I p×p ⊗ A := ⎝ . ⎠ (p instances of A blocks on the diagonal).
A
Theorem 8.3. Define the projectors
Πslow = InMS ×nMS ⊗ (PR), Πfast = I − Πslow .
Then
C−1
11 Π
slow −1
= (I ⊗ P)Ĉ11 (I ⊗ R), C−1
11 Π
fast
= −Πfast .
−1
Proof. Lemma 8.1 yields a decomposition of C11 into a product of the three ma-
trices A1 A2 A3 . The same type of decomposition holds when the blocks in C are
substituted by their tilde or hat counterparts. We now show in three steps that
k (I ⊗ P) = (I ⊗ P)Âk ,
A k = 1, 2, 3,
from which we can immediately infer the assertion
C−1
11 Π
slow 1 A
=A 2 A
3 (I ⊗ (PR)) = (I ⊗ P)Â1 Â2 Â3 (I ⊗ R) = (I ⊗ P)Ĉ−1 (I ⊗ R).
11
The proof of the assertion for Πfast is based on equation (8.3) which yields
We obtain
1 Πfast = −Πfast ,
A 2 Πfast = Πfast ,
A 3 Πfast = Πfast ,
A
because all off-diagonal blocks are eliminated due to equation (8.5) and
M −1 (I − PR) = I − P I − M̂ −1 R (I − PR) = I − PR.
B B
C−1
11 Π
fast
=A 2 A
1 A 3 Πfast = −Πfast
Proof. Consider
Imagine the variable vector x1 comprising all discretized PDE states and the small
variable vector x2 containing the remaining degrees of freedom. The proof of the
following theorem can be carried out via KKT transformation rules as in Leinewe-
ber [106]. We want to give a slightly shorter proof here which can be interpreted
as a partial null-space approach.
Theorem 8.6. Assume that C11 in QP (8.6) is invertible and define
−1
−C11 C12
Z= , B = Z T BZ,
I
c1 = C11
−1
c1 , b = B21 c1 + b2 −C12
T −T
C11 (B11 c1 + b1 ),
c2 = c2 −C21 c1 , C2 = C22 −C21C11
−1
C12 ,
c3 = c3 −C31 c1 , C3 = C32 −C31C11
−1
C12 .
Let furthermore (x2∗ , y∗E2 , y∗I ) ∈ Rn2 +m2 +m3 be a primal-dual solution of the QP
1 T
minimize x B x2 + bT x2 s. t. C2 x2 = c2 , C3 x2 ≥ c3 . (8.7)
x2 ∈R
n 2 2 2
If we choose
x1∗ = C11
−1
(c1 −C12 x2∗ ), (8.8a)
y∗E1 = C11
−T
(B12 − B11C11 −1
C12 )x2∗ + B11 c1 + b1 −C21
T ∗ T ∗
yE2 −C31 yI (8.8b)
then (x∗ , y∗ ) := (x1∗ , x2∗ , y∗E1 , y∗E2 , y∗I ) is a primal-dual solution of QP (8.6).
Proof. We first observe that constraint (8.6b) is equivalent to equation (8.8a) and
that thus ∗
x1 ∗ c
∗ = Zx2 + 1
x2 0
8.3 Elimination of discretized PDE states 119
Feasibility is guaranteed by
Finally complementarity holds because the multipliers y∗I for the inequalities are
the same in the condensed QP (8.7) and in the structured QP (8.6).
The condensed QP (8.7) is of much smaller size than QP (8.6) and its size does
not depend on the spatial discretization level. It still exhibits the typical multiple
shooting structure in the ODE states and could thus be condensed one more time.
In the examples which we present in Part III, however, the computational savings
are only marginal between skipping the second condensing and solving QP (8.7)
directly with the method we describe in Chapter 9.
120 8 Condensing
Consequently we have
ZT Bfast Z = 0
and thus we can also compute the condensed Newton-Picard Hessian matrix purely
on the slow subspace according to
−1
T −Ĉ11 Ĉ12
B = Z BZ = Ẑ B̂Ẑ with Ẑ =
T
.
I
We now consider the family of scaled NLPs of the form of NLP (8.2)
minimize f (β x1 , D2 x2 ) (8.10a)
x=(x1 ,x2 )∈Rn1 +n2
s. t. αgi (β x1 , D2 x2 ) = 0, i ∈ E1 , (8.10b)
ai gi (β x1 , D2 x2 ) = 0, i ∈ E2 , (8.10c)
ai gi (β x1 , D2 x2 ) ≥ 0, i∈I. (8.10d)
After introducing the scaled variables xsc = D−1 x and ysc = A−1 y we can establish
for the function F of Section 5.7.1 in Chapter 5 and its scaled counterpart that
D∇x L (x, y)
F sc (xsc , ysc ) = = diag(D, A)F(x, y),
Ag(Dx)
dF sc (xsc , ysc ) D 0 dF(x, y) D 0
=
d(xsc , ysc ) 0 A d(x, y) 0 A
122 8 Condensing
1 T
minimize x Bx + bT x s. t. cl ≤ Cx ≤ cu , (9.1)
x∈Rn 2
economic dispatch, etc. Several benchmark and application problems are col-
lected in a repository [111] which is accessible through the CUTEr testing en-
vironment [68].
extension of the Simplex Algorithm [36]. Dual Active Set methods apply the
primal Active Set method to the dual of QP (9.1) (which exists if B is semidefinite).
A relatively new variant of Active Set methods are Parametric Active Set Methods
(PASM), e.g., the PQP method due to Best [17], which are the methods of interest
in this thesis. PASMs are based on an affine-linear homotopy between a QP with
known solution and the QP to be solved. It turns out that the optimal solutions
depend piecewise affine-linear on the homotopy parameter and that along each
affine-linear segment the active set is constant. The iterates of the method are
simply the start points of each segment.
The numerical behavior of Active Set and Interior Point methods is usually quite
different: While Active Set methods need on average substantially more iterations
than Interior Point methods, the numerical effort for one iteration is substantially
less for Active Set methods. Often one or the other method will perform favorably
on a certain problem instance, underlining that both approaches are important.
We want to concisely compare the main advantages of the different Active Set
versus Interior Point methods. One advantage of Interior Point methods is the
regularizing effect of the central path which leads to well-defined behavior on
problems with nonunique solutions due to, e.g., degeneracy or zero curvature in
a feasible direction at the solution. An advantage of all Active Set methods is
the possibility of hot starts which can give a substantial speed-up when solving
a sequence of related QPs because the active set between the solutions usually
changes only slightly. A unique advantage of PASM is that the so-called Phase 1
is not needed. The term Phase 1 describes the solution of an auxiliary problem to
find a feasible starting point for primal and dual Active Set methods or a strictly
feasible starting point for Interior Point methods. The generation of an appropriate
starting point with Phase 1 can be as expensive as the subsequent solution of the
actual problem.
Table 9.1: Software for convex Quadratic Programming (in alphabetical order).
offers the option of a concurrent optimizer which starts a Barrier and an Active Set
method in parallel and returns the solution which was found in the shortest amount
of CPU time.
For Parametric Active Set methods we are only aware of the code qpOASES (see
Ferreau [52], Ferreau et al. [53]). We have developed a prototype Matlab® code
called rpasm to demonstrate the efficacy of the proposed techniques and strategies
to fortify reliability of PASM.
9.2 Parametric Active Set methods 127
1
minimize x(τ)T Bx(τ) + b(τ)T x(τ) s. t. cl (τ) ≤ Cx(τ) ≤ cu (τ), (9.3)
x(τ)∈R n 2
with b, cl , and cu now being continuous functions b(τ), cl (τ), cu (τ). For fixed τ,
let the optimal primal-dual solution be denoted by z(τ) = (x(τ), y(τ)) which nec-
essarily satisfies (9.2) (with b = b(τ), cl = cl (τ), cu = cu (τ)). If we furthermore
restrict the homotopy to affine-linear functions b ∈ H n , cl , cu ∈ H m , where
it turns out that the optimal solutions z(τ) depend piecewise linearly but not nec-
essarily continuously on τ (see Best [17]). On each linear segment the active set
is constant. Parametric Active Set algorithms follow z(τ) by jumping from one
beginning of a segment to the next. We can immediately observe that this ap-
proach allows hot-starts in a natural way. As mentioned already in Section 9.1.2,
no Phase 1 is needed to begin the method: We can always recede to the homotopy
start b(0) = 0, cl (0) = 0, cu (0) = 0, x(0) = 0, y(0) = 0, although this is certainly
not the best choice as we discuss in Section 9.3 and Section 9.4.
on which (if any) bound is marked active in Wi . We can then determine the step
direction (Δx, Δy) by solving
Δx B T
CW Δx −(b(1) − b(τ))
KW (τ) := = . (9.4)
−ΔyW CW 0 −ΔyW cW (1) − cW (τ)
The dual step Δy must be assembled from ΔyW by filling in zeros at the entries of
constraints i which are not in the working set (i.e., Wi = 0). For the initial working
9.2 Parametric Active Set methods 129
set W we assume matrix CW to have full rank and matrix B to be positive definite on
the null space of CW . Thus matrix KW (0) is invertible. As we shall see in Steps 3
and 6, the PQP algorithm ensures the full rank and positive definiteness properties
and thus invertibility of KW (τ) for all further steps through exchange rules for the
working set W . We shall discuss a null space approach for the factorization of
KW (τ) in Step 9.
Step 2: Determination of step length. We can follow z(τ) in direction Δz along
the current segment until either an inactive constraint becomes active (blocking
constraint) or until the dual variable of a constraint in the working set becomes
zero (blocking dual variable). Following the straight line with direction Δz beyond
this point would lead to violation of conditions (9.2). The step length Δτ can be
determined by ratio tests
If the set of ratios is empty the minimum yields ∞ by convention. With the help of
RT, the maximum step towards the first blocking constraint is given by
Cl = ∑ ξiCi . (9.9)
i:Wi =0
Multiplying equation (9.9) by λWl+ with λ ≥ 0 and adding this as a special form
of zero to the stationarity condition in equations (9.2) yields
If λ = ∞ then the parametric QP does not possess a feasible point beyond τ + and
thus the QP to be solved (at τ = 1) is infeasible. Otherwise, let k be a minimizing
index of the ratio set.
Step 5: Jump in dual variables. Now let
"
−λWi+ for i = l,
ỹi = +
+
yi (τ ) + λWi ξi for i : Wi = 0,
and set y(τ + ) := ỹ. It follows from construction of λ that ỹk = 0 and thus, con-
straint k can leave the working set. As a consequence, matrix CW + \{k} preserves
the full rank property and has the same null space as CW , thus securing regularity
of matrix KW + (τ + ).
Step 6: Curvature test. The removal of a constraint from the working set can
lead to exposure of directions of zero curvature on the null space of CW + (which
9.2 Parametric Active Set methods 131
where ek is the k-th column of the m-by-m identity matrix. Only if ξ = 0 then B
is singular on the null space of CW + (see Best [17]). As for the linear indepen-
dence test of Step 3, the curvature test can be evaluated cheaply by reusing the
factorization needed to solve the step equation (9.4).
Step 7: Determination of exchange index l. It holds that ξ = 0 and s solves
Bs = 0, Ck s = −1, CW + s = 0. (9.13)
1 T
minimize x Bx + bT x s. t. c̃l ≤ Cx ≤ c̃u , (9.15)
x∈R n 2
where c̃li = max(cli , −M), c̃ui = min(cui , M), i = 1, . . . , m. We call a constraint bound
which attains the value ±M a far bound. Algorithmically we solve a sequence of
QPs with growing far bounds value M, see Algorithm 3. The total solution time
will mostly be dominated by the solution of the first QP as consecutive QPs of the
form (9.15) can be efficiently hot-started.
9.3 Fundamental numerical challenges 133
KW z = d
z0 = 0, rk = KW zk − d, KW δ zk = rk , zk+1 = zk − δ zk
recovers a fixed number of extra valid digits in each iteration. In the previous
example, the iterate z2 has at least twelve valid decimal digits after only one extra
step of Iterative Refinement. It is worth noticing that compared to the original
9.4 Strategies to meet numerical challenges 135
for i ∈ m. This annihilates the effects of drift after every iteration, at the cost of
splitting up the single homotopy into a sequence of homotopies which are, how-
ever, very close to the remaining part of the original homotopy. In exact arithmetic
the proposed modification does not alter any value.
Termination Criterion. It is tempting to use the homotopy parameter τ in the
termination criterion as proposed in Algorithm 2. However, this choice renders the
termination criterion dependent on the choice of the homotopy start, an undesirable
property. Instead we propose to use the relative distance δ in the data space
This choice also renders the termination criterion independent of the condition
number of KW (1). We observe that the termination criterion can give no guarantee
for the distance to the exact solution. Instead a backwards analysis result holds:
The computed solution is the exact solution to a problem which is as close as δ to
the one to be solved. The numerical results presented in Section 9.6 were obtained
with the tolerance δ ≤ δterm = 1e7 eps.
136 9 A Parametric Active Set method for QP solution
The entries j = l of c̃l and c̃u are set to the corresponding entries of cl and cu .
Consequently the Cholesky decomposition from the previous step stays valid for
the current projected Hessian.
The numerical results presented in Section 9.6 were obtained with the curvature
tolerance δcurv = 1e4 eps.
i = max(ui , εcut ),
ucut i ∈ m,
RTr (u, v, εcut , εden , εnum ) = min{ucut
i /vi | i ∈ m, vi ≥ εden , ucut
i ≥ εnum }.
We now explain the purpose of the three tolerances: The denominator tolerance
εden > 0 describes which small but positive values of vi should already be consid-
ered less than or equal to zero. They are consequently discarded as candidates for
the minimum.
The cutting tolerance εcut and the numerator tolerance εnum offer the freedom of
two different treatments for numerators close to zero. If εcut > εnum then negative
9.4 Strategies to meet numerical challenges 137
numerators are simply cut off at εcut before the quotients are taken, yielding that
the minimum is greater or equal to εcut /εden . For instance, we set εcut = 0 in
the ratio tests for determination of the step length (9.6) and (9.7). This choice is
motivated by the fact that in exact arithmetic ui ≥ 0 for all i ∈ m with vi > 0. Thus
only values ui which are negative due to round-off are manipulated and the step
length satisfies Δτ ≥ 0 also in finite precision arithmetic.
If εcut ≤ εnum then cutting does not have any effect. We have found it beneficial
for the reliability of PASM to set εnum = εden in the ratio tests (9.11) and (9.14) for
finding exchange indices.
The numerical results presented in Section 9.6 were obtained with the ratio test
tolerances εden = −εnum = 1e3 eps and εcut = 0 for step length determination and
εden = εnum = −εcut = 1e3 eps for the remaining ratio tests.
Linear independence and zero curvature test. After solution of systems (9.8)
and (9.12) for s and ξW we must compare the norm of s or ξ with zero. Let
ζ = (s, ξ ). We propose to use the relative conditions
s∞ ≤ εtest ζ ∞ for the linear dependence test and (9.16)
ξ ∞ ≤ εtest ζ ∞ for the zero curvature test. (9.17)
We remark that ζ ∞ = ξ ∞ if s = 0 and ζ ∞ = s∞ if ξ = 0. Thus we can
replace ζ ∞ in the code by ξ ∞ in test (9.16) and by s∞ in test (9.17). The
numerical results presented in Section 9.6 were obtained with εtest = 1e5 eps.
is an optimal solution to the homotopy start b̃(0), c̃l (0), c̃u (0), where for i ∈ m
"
cl (0), if Wi = −1,
c̃i (0) = i
l
ci (0) − ri , otherwise,
l 1
"
cu (0), if Wi = +1,
c̃ui (0) = i
ci (0) + ri , otherwise,
u 2
In other words, if we move the inactive constraint bounds further away from Cx(0)
and the dual variables of the active constraints further away from zero, x(0) stays
feasible and b(0) can be adapted to restore optimality of (x(0), ỹ(0)) with the same
working set W . Recall that the ratio tests depend exactly on the residuals of the in-
active constraints and the dual variables of the active constraints. In our numerical
tests, the simple choice of
has proved to work reliably. Because of the shape of r j , we call this strategy
ramping. It is important to avoid two entries of r j to have the same value because
many QP problems exhibit special structures, e.g., variable bounds of the same
value for several variables which lead to primal ties if the homotopy starts with
the same value for each of these variables. Of course, the choice of linear ramping
is somewhat arbitrary and if a problem happens to have variable bounds in the
form of a ramp, ties are again possible. However, this kind of structure is far less
common than equal variable bounds.
We employ ramping in the starting point of the homotopy and also after an
iteration which resulted in a zero step Δτ = 0. Of course, this can lead to large
jumps in the problem homotopy and practically catapult the current b(0) := b̃(0)
further away from b(1). However, a PASM is capable of reducing even a large
distance in the data space to zero in one step, provided the active set is correct.
Thus the distance of the working set W to the active set of the solution is a more
appropriate measure of the progress of a PASM. By construction, the active set is
preserved by the ramping strategy.
We further want to remark that ties can never be completely avoided. For in-
stance in case of a QP whose solution lies in a degenerate corner, a tie must occur
in (at least) one iteration of a PASM. In the numerical examples we have treated
so far, the ramping strategy effectively deferred these ties to the final step, where a
tie is not a problem any more because the solution at the end of the last homotopy
9.5 The code rpasm: A PASM in Matlab® 139
and 8 MB cache in Matlab® 7.6 under Linux 2.6 (64 bit). The quality of solutions
(x∗ , y∗ ) was measured using a residual ρ of conditions (9.2) defined via
ρstat = Bx∗ + b −CT y∗ ∞ ,
ρfeas = max(0, cl −Cx∗ ,Cx∗ − cu ),
ρcmpl
l
= max{(Cx∗ − cl )i y∗i | y∗i ≥ +10 eps},
ρcmpl
u
= max{|(Cx∗ − cu )i y∗i | | y∗i ≤ −10 eps},
ρ = max(ρstat , ρfeas , ρcmpl
l
, ρcmpl
u
).
We visualize the results for problems from the Maros-Mészáros test set [111] with
at most n = 1000 variables and m = 1001 two-sided inequality constraints (not
counting variable bound constraints) in the performance graphs of Figures 9.1
and 9.2. The graphs display a time factor on the abscissa versus the percentage
of problems that each code was able to solve within the time factor times the run-
time of the fastest method for each problem. Roughly speaking, the graph of a
fast method is close to the left hand side of the diagram, the graph of a reliable
method is close to the top of the diagram. We remark that the results for rpasm
were obtained using only dense linear algebra routines.
There is a certain arbitrariness in the notion of a “solved problem” between the
different codes. We choose to consider a problem as solved if ρ is less than or
equal to a certain threshold. This approach is not unproblematic either: A not
tight enough termination threshold of a code can lead to premature termination
and the problem would be considered “not solved” by our criterion, although the
method might have been able to recover a better solution with more iterations.
This is especially an issue for Interior Point/Barrier methods. Thus the graphs in
Figures 9.1 and 9.2 show reliability of the methods only in connection with their
default settings. However, we are not aware of any simple procedure which would
lead to a fairer comparison. Figure 9.1 shows the results with a relatively loose
threshold of ρ ≤ 1e-2 and Figure 9.2 with a tighter threshold of ρ ≤ 1e-8.
100
rpasm0
90 rpasm1
quadprog
Percentage of problems solved
OOQP
80 qpOASES
CPLEXP
70 CPLEXD
CPLEXB
60
50
40
30
20
10
0
100 101 102 103 104 105
Time factor
runtime and the number of solved problems (62 %). The primal and dual versions
of CPLEX are the second most reliable with 96 % and 97 %. CPLEX solves no
problem in less than 1.3 s, not even the small examples which are solved in a few
milliseconds by rpasm. We suspect that this is due to a calling overhead in CPLEX,
e.g., for license checking. This is also one reason why OOQP is much faster than
the Barrier version of CPLEX, albeit they both solve roughly the same number of
problems (70 % and 73 %, respectively). Even though the code qpOASES is only
appropriate for QPs with positive definite Hessian, which make up only 27 % of
the considered problems, it still solves 44 % of the test problems. Additionally,
we want to stress that those problems solved by qpOASES were indeed solved
quickly.
Now we discuss the differences between Figure 9.2 and Figure 9.1, i.e., when
switching to a tighter residual tolerance of ρ ≤ 1e-8: The ratio of solved prob-
lems drops dramatically for the Interior Point/Barrier methods (CPLEXB: 29 %,
OOQP: 37 %). This is a known fact and the reason for the existence of crossover
methods which refine the results of Interior Point/Barrier methods with an Active
Set method. The code qpOASES still solves 44 % of the problems, which indi-
cates that the solutions that qpOASES yields are of high quality. Furthermore,
142 9 A Parametric Active Set method for QP solution
100
rpasm0
90 rpasm1
quadprog
Percentage of problems solved
OOQP
80 qpOASES
CPLEXP
70 CPLEXD
CPLEXB
60
50
40
30
20
10
0
100 101 102 103 104 105
Time factor
qpOASES is fast: It solves 36 % of the problems within 110 % of the time of the
fastest method for each of these problems. The number of problems solved by
quadprog decreases to 53 %. The primal and dual Active Set versions of CPLEX
solve 78 % of the problems. Only the code rpasm is able to solve more than 80 %
of the problems to a residual of ρ ≤ 1e-8 (rpasm0: 82 %, rpasm1: 84 %).
We can conclude that the strategies proposed in Section 9.4 indeed lead to a
reliable method for the solution of convex QPs.
1 k−2 1 k=1
minimize ∑
2 i=1
(xk+i+1 − xk+i )2 − ∑ (xk−i + xk+i + αk−i+1 )2
2 i=1
(9.18a)
bound (white). Thus direct transitions from black to white or vice versa along a
vertical line indicate flipping bounds. We can observe that the chosen initial work-
ing set is completely different to the final working set in the solution. Still the
number of iterations is less than two times the number of constraints which indi-
cates that the proposed method works efficiently on this instance of the nonconvex
problem (9.18).
In the solution which corresponds to Figure 9.3, n = 199 out of m constraints
are active and strict complementarity is satisfied. Thus we indeed have obtained a
local optimum.
50
100
150
Iteration
200
250
300
350
Figure 9.3: Active set changes for nonconvex problem (9.18), k = 100. Each line of the
image corresponds to the working set in one iteration. The colors indicate constraints which
are inactive (gray), active at the lower bound (black), or active at the upper bound (white).
Direct transitions from black to white or vice versa along a vertical line indicate flipping
bounds.
10 Automatic derivative generation
The inexact SQP method which we describe in Chapter 5 requires first and second
order derivatives of the problem functions. There are several ways how deriva-
tives can be provided. The first is to have them provided along with the problem-
dependent model functions by the user. This can be cumbersome for the user and
it is impossible for the program to check whether the derivatives are free of errors,
even though consistency tests evaluated in a few points can somewhat mitigate the
problem. These are, however, severe drawbacks.
A second way is the use of symbolic calculation of derivatives. Although in
principle possible by the use of symbolic computer algebra systems, the resulting
expressions for the derivatives can become too large to be evaluated efficiently.
A third way is the use of numerical schemes. Finite differences can be computed
efficiently but they inevitably involve cancellation and truncation errors. While the
cancellation errors can be circumvented by using a complex step derivative (see
Squire and Trapp [148]) two further drawbacks still remain: First, complex step
derivatives without cancellation are limited to first order derivatives. Second, the
evaluation of the gradient of a scalar-valued function of many variables cannot be
carried out efficiently.
The aim of this chapter is to recapitulate an efficient and automated way to
compute derivatives from a given computer code. This fourth way does not suffer
from the drawbacks of the previous three. The main principles are Algorithmic
Differentiation (AD) and Internal Numerical Differentiation (IND). We refer the
reader to Griewank [70] and Bock [22, 23], respectively.
The chapter is structured in four sections. In Section 10.1 we give a concise
survey about the idea behind AD and in Section 10.2 about the principle of IND.
We continue with the discussion of a subtle difficulty in the application of the IND
principle to implicit time-stepping methods with monitor strategy in Section 10.3
and conclude the chapter in Section 10.4 with a short note on the numerical effort
needed for the first and second order derivative generation needed in the inexact
SQP method for NLP (3.3) described in Chapter 5.
Backward mode. We traverse the evaluation graph backwards from the depen-
dent variables to the independent variables while accumulating the deriva-
tive information. The numerical effort of the backward mode to compute an
adjoint directional derivative at x ∈ Rnind in the direction of s ∈ Rndep
∇F(x)s
is also only a small multiple of the evaluation of F(x). For the backward
mode, however, the function F(x) has to be evaluated in advance and all in-
termediate results must be accessible when traversing backwards through the
evaluation graph. This is usually accomplished by storing all intermediate
results to a contiguous memory block, the so called tape, or by a checkpoint-
ing strategy which stores only a few intermediate results and recomputes the
remaining ones on the fly. Both approaches have their value depending on
the ratio of computation speed and access time to the memory hierarchy on
a particular computer architecture.
The elemental operations can be formally generalized to operate on truncated
Taylor series. This approach makes the evaluation of arbitrary-order derivatives
possible in a unified framework. To circumvent the drawback of high memory
10.2 The principle of IND 147
requirements and irregular memory access patterns, Griewank et al. [74] suggest
to use only univariate truncated Taylor series from which mixed derivatives can be
obtained by an interpolation procedure. Univariate Taylor coefficient propagation
can also be used in forward and backward mode.
∂t u = Δu in (0, 1) × Ω,
u=0 on (0, 1) × ∂ Ω,
u t=0
=u . 0
We discretize the problem with the FDM in space on the equidistant grid
x j = jh, j = 0, . . . , N, h = π/N,
To satisfy the boundary conditions the values in the nodes x0 and xN are implicitly
set to zero and are not a part of the discretized vector u(t).
and that the eigenvalue of largest modulus tends towards minus infinity
2 2 4N 2
λN−1 = N (cos(π(N − 1)/N) − 1) ≈ − 2 .
π 2 π
Thus ODE (10.1) is stiff and becomes stiffer for finer spatial discretizations.
Because the vectors vk are N − 1 eigenvectors to pairwise different eigenvalues
λ they form a basis of RN−1 . If we rewrite ODE (10.1) in the basis spanned by
k
Consider now a Backward Euler method for ODE (10.2). Starting from u0 at
t0 u } such that the value
= 0 we compute a sequence {n u at t = t
n n n−1 + Δt solves
n
approximately
Although we could solve equation (10.3) exactly without much effort, this is not
the case for nonlinear ODEs. Efficient numerical integrators employ a Newton-
type method for the efficient solution of the nonlinear counterpart of equation
(10.3). The so called monitor strategy is a Simplified Newton method for equa-
tion (10.3) where the decomposition of the iteration matrix M k of a previous step
is utilized and the contraction of the iteration is monitored. Usually we update
the iteration matrix if the error has not been reduced satisfactorily within three
Newton-type iterations. For simplification of presentation we assume that we per-
form enough iterations to reduce the error to machine precision. In the case of
ODE (10.2) the iteration boils down to
un,i =
un,i−1 − (M k )−1 (M n
un,i−1 −
un−1 ).
150 10 Automatic derivative generation
un,0
This iteration is unstable if and only if there exists an index j with j = 0 and
such that
1 − (1 − Δt n λ j )/(1 − Δt k λ j ) > 1,
or, equivalently,
1 − Δt n λ j 1
>2 ⇔ 1 − Δt n λ j > 2 − 2Δt k λ j ⇔ Δt n > + 2Δt k . (10.4)
1 − Δt k λ j λj
A numerically optimal step size controller for stiff systems must maximize the step
length such that the method is always on the verge of becoming unstable. We see
from condition (10.4) that the step size can be more aggressively increased if the
initial value does not contain modes which belong to eigenvalues of large modulus.
The following problem now becomes apparent: Assume that the initial value is
a linear combination of only the first p N (low-frequency) modes, i.e.,
p
u0 = ∑ u0j v j .
j=1
Assume further that the monitor strategy keeps the iteration matrix at M 0 for the
first steps and that the step size controller chooses
1 1
Δt j = + 2Δt 1 < + 2Δt 1 , j = 2, . . . , n,
λ p+1 λ p
which yields a stable scheme. Say we want to compute the derivative of u(t n )
with respect to u0 in direction d by using the chosen scheme on the VDE for
ODE (10.1), which is again ODE (10.1) with initial value d because ODE (10.1)
is linear. If d has nonzero components in the directions of vk , k = p + 2, . . . , N − 1,
then this computation is not stable and thus not convergent. Hence this approach
does not satisfy the IND principle.
One possible solution to this problem is to perform adaptive error control on the
nominal values simultaneously with the forward derivative values (see, e.g., Al-
bersmeyer [2, Section 6.7.5]). However, computing the gradient of the Lagrangian
with forward sweeps is prohibitively expensive for large nind . Thus this approach
is computationally not feasible for the problems considered in this thesis.
10.4 Numerical effort of IND 151
Our pragmatic approach, which has worked reliably for the examples in Part III,
is to tighten the required tolerance for the Simplified Newton method that approxi-
mates the solution of the implicit system (equation (10.3) in our previous example)
by a factor of 1000. This leads to more frequent reevaluation of the iteration matrix
within the monitor strategy and enlarges thus the numerical stability regions. We
have not yet performed a thorough comparison of the extra numerical effort of this
approach which seems to amount to roughly 20%.
occurring in Chapter 5 in only a small multiple of the numerical effort spent for
evaluation of F(zk ). Even though the upper left block contains second derivatives
of ui and vi they need only be evaluated in one adjoint direction given by the current
Lagrange multipliers in yk and the forward direction given by v1 .
11 The software package MUSCOP
We have implemented the inexact SQP method based on the GINKO algorithm
with two-grid Newton-Picard generalized LISA as described in Chapter 5. It is our
goal in this chapter to highlight the most important software design decisions and
features. We have named the software package MUSCOP, which is an acronym
for Multiple Shooting Code for PDEs. The name alludes to the successful software
package MUSCOD-II (see Leineweber [105], Leineweber et al. [107] with exten-
sions by Leineweber [106], Diehl [47], Schäfer [141], Sager [138]) because we
had originally intended to design it as a MUSCOD-II extension. In Section 11.1
we discuss why we have decided to develop the method in a stand-alone parallel
form and outline which programming paradigms have proved to be useful in the
development of MUSCOP. Afterwards we explain the orchestration of the different
software components in Section 11.2.
by only writing assembler machine code for a specific computer architecture, but
such a code might soon become too complex and surpass a developer’s capacity to
maintain or extend it in a reasonable amount of time. On the other hand, the sole
use of a high-level numerical programming language like Matlab® or GNU Oc-
tave might result in a considerable loss of performance, especially if algorithms
are not easily vectorizable, while the development time of the code and its ex-
tensions might be dramatically reduced, mainly because debugging problems on
a numerical level is possible in a more accessible way by the use of well-tested
built-in methods like cond, eig, svd, etc., and data visualization.
We use hybrid language programming in the following way: All time-critical
algorithmic components should be implemented in lower-level programming lan-
guages and all other components in higher-level programming languages. This
concept is not new, GNU Octave being one example because it is written in C++
while most dense linear algebra components are based on an optimized BLAS and
LAPACK implementation called ATLAS (see Whaley et al. [163]).
In the case of MUSCOP the most time-critical component is the evaluation of
the shooting trajectories and their derivatives of first and second order (see Part III).
This task is performed by the software package SolvIND (see Albersmeyer and
Kirches [5]) which is entirely written in C++ and uses ATLAS, UMFPACK (see
Davis [39]), and ADOL-C (see Griewank et al. [72, 73], Walther et al. [159]). Most
of the remaining code is written in Matlab® /GNU Octave except for the interface
to SolvIND which is at the time of writing only available for GNU Octave and
not for Matlab® . This approach has proven to be beneficial for the development
speed of MUSCOP while only minor performance penalties have to be accepted
(see Part III).
The GNU Octave and C++ components of MUSCOP are separated on the left
hand and right hand side of Figure 11.1, respectively. We shall give a more detailed
explanation of Figure 11.1 in Section 11.2.
namely the computation of Ẑ, has been performed. Only then can we evaluate the
partially projected (coarse grid) Hessian B = Ẑ T B̂Ẑ.
We do not want to suggest that a functional encapsulation in logical blocks like
function evaluation, condensing, QP solution, etc. is impedimental. We believe,
however, that the encapsulation of the data of these blocks is. In our opinion the
structure of the code should indeed follow the logical structure of a mathemati-
cal exposition of the method but the exchange of data should not be artificially
obstructed by interfaces which follow the functional blocks.
Object Oriented Programming (OOP) raises the rigid coupling of function and
data interfaces (methods and private members in the language of OOP) to a design
principle. We believe that OOP is a valid approach for software which is supposed
to be used in a black-box fashion but we believe it to be more obstructive than help-
ful for the structure exploiting numerical methods we develop in this thesis. This is
the main reason why MUSCOP is not OOP and not an extension of MUSCOD-II.
In MUSCOP we use a global, hierarchical data structure which is accessible
in all software components, at least on the Matlab® /GNU Octave level. Hierar-
chical here means that the data structure consists of substructures which map the
functional blocks to data blocks without hiding them. The biggest disadvantage
of global variables is if course that users and developers have to know which vari-
ables they are allowed to write access and in what states the variables are when
performing a read access. This task is without doubt a difficult one to accomplish.
But the difficulty really stems from the complexity of the numerical method and
not from the choice of computer language or programming paradigm. No pro-
gramming paradigm can turn a complex and difficult-to-understand method into a
simple and easy-to-understand code.
)
#
&
( '(
!$
&#$
(see, e.g., Dongarra et al. [49], Lehoucq et al. [104]). Its most compelling features
are simplicity and flexibility, especially when multiple programming languages are
in use.
A program with Reverse Communication interface is called with an incomplete
set of input data first. If more input data is needed the program returns to the caller
indicating which input data is needed next. After computation of this data the user
calls the program again passing the new input data. This procedure is iterated until
the program signals termination to the user.
A typical example is an iterative linear algebra solver which returns to ask the
user for a matrix vector multiplication or a preconditioner vector multiplication.
Obviously the solver does not need to know the matrix or the preconditioner, nor
does it need to pose restrictions on how they are represented.
GINKO also uses Reverse Communication. When called without input param-
eters GINKO initializes a data structure which is then modified by the user and
passed back to GINKO. In this data structure there are two state flags, the flow
control and the action flag. The flow control flag tells GINKO which part of the
code is the next to be evaluated and the action flag tells the user on return which
variables in the data structure must be freshly computed before GINKO can be
11.2 Orchestration of software components 157
We observe that the right hand side of equation (11.1) is afflicted with a cancella-
tion error in the norm of the numerator. This error is then amplified by αk−2 causing
that if the step size αk drops below, say, 10−4 then [hδk ]∗ might be overestimated.
This in turn leads to even smaller step sizes
1
αk = .
(1 + ρ)[hδk ]∗
Thus GINKO gradually reduces αk to zero and the method stalls. We have imple-
mented an estimator for the cancellation error. The cancellation error is displayed
for each iteration of MUSCOP but does not influence the iterations of MUSCOP.
It rather serves as an indicator for failure analysis.
1
Eigenvalue μi
0.5
0
0 50 100 150 200
Index i
Figure 12.1: The eigenvalues μi of the spectrum of the monodromy matrix Gu decay expo-
nentially fast. Only few eigenvalues are greater than 0.5. Shown are the first 200 eigenvalues
calculated with D = 0.01 and β = 100χ(Γ) on a grid of 8-by-8 elements of order 5.
2
1
1.5
Controls q
1 0.5
0.5
0 0
1
1
-0.5 0 0
0 0.5 1 -1
x2 -1 x1
Time t
Figure 12.2: Optimal controls q (left) and optimal states u0 (right) for target function û,
calculated for D = 0.01 on a grid of 32-by-32 elements of order 5. The displayed mesh is
not the finite element mesh but an evaluation of the Finite Element function on a coarser
equidistant mesh.
1 100
0.9
10−2
0.8
0.7 10−4
Contraction
0.6
10−6
0.5
0.4 10−8
0.3 10−10
0.2
10−12
0.1
0
0 100 200 300 400 0 100 200 300 400
Subspace dimension p Subspace dimension p
Figure 12.3: LISA contraction with Newton-Picard preconditioning versus the subspace di-
mension p for the Euclidean projector (left) and the L2 projector. Note that the plot on the
right hand side is in logarithmic scale.
166 12 Linear boundary control for the periodic 2D heat equation
1 1
0.5 0.5
0 0
-0.5 -0.5
-1 -1
-1 -0.5 0 0.5 1 -1 -0.5 0 0.5 1
1 1
0.5 0.5
0 0
-0.5 -0.5
-1 -1
-1 -0.5 0 0.5 1 -1 -0.5 0 0.5 1
Figure 12.4: Top row: Unit circle and spectrum of iteration matrix for the classical Newton-
Picard with p = 20 using Euclidean projector (left column) and L2 projector (right column).
Bottom row: Like top row with p = 45.
Figure 12.5: Asymptotic contraction rate for classical Newton-Picard preconditioning ver-
sus subspace dimension p for varying diffusion coefficient D and spatial degrees of freedom
nu .
100
10−1
Contraction σr (J˜−1 ΔJ)
10−2
D = 0.001, nfu = 6561
D = 0.001, nfu = 25921
10−3
D = 0.01, nfu = 6561
D = 0.01, nfu = 25921
10−4 D = 0.1, nfu = 6561
D = 0.1, nfu = 25921
10−5
101 102
Coarse grid degrees of freedom ncu
Figure 12.6: Asymptotic contraction rate for two-grid Newton-Picard preconditioning ver-
sus coarse grid degrees of freedom ncu for varying diffusion coefficient D and fine grid
degrees of freedom nfu .
within 11, 34, 98 iterations for D = 0.1, 0.01, 0.001, respectively, with a termina-
tion tolerance εO = 10−4 . We remark that for inexact inner solutions with εM , εH
much larger than machine precision, Flexible GMRES (see Saad [134]) should be
employed.
As we have seen in Section 6.3.6, the effort on the coarse grid for the two-
grid Newton-Picard preconditioner is negligible compared to the effort on the fine
grid. Thus, even medium scale coarse grid degrees of freedom ncu are possible in
practical computations and lead to fast contraction rates. In this case, acceleration
of LISA (6.2) by nonlinear Krylov subspace methods does not lead to considerable
savings in the number of iterations.
12.4 Comparison with Schur complement preconditioning 169
−1
√
leads to JMGW Jˆ having exactly three different eigenvalues 1 and (1 ± 5)/2. As
a consequence, any Krylov subspace method with an optimality or Galerkin prop-
erty converges within 3 iterations for the preconditioned system. Inversion of the
lower right block of JMGW is computationally prohibitively expensive but we can
approximate this block by the Newton-Picard approach presented in Section 6.3
which leads with X̃ = (G̃u − Inu )M −1 (G̃Tu − Inu ) ∈ Rnu ×nu to the preconditioner
⎛ ⎞
M 0 0
⎜ ⎟
J˜MGW = ⎝ 0 N 0 ⎠.
−1
0 0 X̃ + γ Gq N Gq −1 T
with X̃ −1 = (G̃u − Inu )−1 M(G̃Tu − Inu )−1 . We observe that the occurring matrices
coincide with the matrices which need to be inverted for the indefinite Newton-
Picard preconditioner J˜ we have developed in Section 6.3. Thus, one iteration of
an iterative method with J˜MGW can be considered computationally as expensive as
one iteration with J.˜
Because the preconditioner J˜MGW is positive definite we can employ it within a
symmetry exploiting Krylov subspace method like MINRES (see Paige and Saun-
ders [123]), which is not possible with the indefinite preconditioner J.˜ On the
downside, it is not possible to use J˜MGW in the basic linear splitting approach (6.2)
−1
because the√real eigenvalues of the iteration matrix In1 +n2 − J˜MGW J cluster around
√
0 and (1 ± 5)/2. Since (1 + 5)/2 > 1 LISA does not converge.
In Figure 12.7 we compare the number of iterations for symmetry exploiting
MINRES preconditioned by J˜MGW with the number of iterations for GMRES pre-
conditioned by J˜ for varying fine and coarse grid degrees of freedom nfu and ncu .
170 12 Linear boundary control for the periodic 2D heat equation
100
MINRES with J˜MGW , nfu = 6561
MINRES with J˜MGW , nfu = 25921
Iterations
0
0 50 100 150 200 250 300
Coarse grid degrees of freedom ncu
Figure 12.7: Comparison of the iterations of MINRES with Newton-Picard Schur comple-
ment preconditioner J˜MGW and GMRES with the symmetric indefinite Newton-Picard pre-
conditioner J˜ for varying fine and coarse grid degrees of freedom nfu and ncu .
on Ω = (0, 1). We see that problem (13.1) is very similar to the model prob-
lem (6.1) except for the polynomial terms in the boundary control condition (13.1c)
of Stefan–Boltzmann type.
where for any symmetric positive definite matrix A we define x2A = xT Ax and
where xPDE denotes the composite vector of PDE states and yPDE denotes the com-
posite vector of dual variables for the PDE state dependent part of the time bound-
ary constraint (3.3b) and the PDE continuity condition (3.3c). The variables xrem
and yrem are placeholders for the remaining primal and dual variables. The occur-
rences of the mass matrices MV in the Kronecker products (see Chapter 8) make
sure that the PDE variables are measured in an L2 sense. For the correct weight-
ing of the dual PDE variables we have to consider that a dual PDE variable ỹ of
NLP (4.1) is from the canonical dual space of RNV . To obtain a discretized Riesz
representation ŷ ∈ RNV in an L2 sense we need to require that
NVl = 4 · 8l−1 + 1
equidistant grid points on levels l = 1, . . . , 5 and controls which are piecewise con-
stant on grids of nMS = 12, 24, 48 equally sized intervals.
Figure 13.1 depicts the solution on the finest grid for the case of nMS = 24.
1.1 1.8
1.6
1.05 1.4
1.2
1
1
0.95 0.8
0.6
0.9 0.4
0.2
0.85 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
space x time t
Figure 13.1: Solution of problem (13.1) with nMS = 24. In the left panel we depict the state
u(1; .) at the period end (solid line) and the desired state û (dashed line). In the right panel
we depict the optimal control q over one period.
that in contrast to the step norm δ zk the error zk − z∗ forms clear plateaus
in iterations 10–12, 18–19, 21–22, and 23–24. These plateaus occur because the
error in these iterations is dominated by the interpolation error of the spatial grid
on the coarser levels. We thus suggest for efficiency reasons to couple the fine grid
refinements to the contraction κ: If we can reduce the error by a factor of κ in
one step then we should refine the grid such that the interpolation error is reduced
with a similar magnitude. Thus we perform an aggressive refinement leading to
eight times more grid points after each refinement. We observe a reduction in
the (interpolation) error of about 1/8 in Figure 13.2 between iterations 19–22 and
22–24. We can thus infer by extrapolation that the final error is dominated by the
spatial interpolation error and lies around 7 · 10−4 . The observed error reduction
of O(h) in the grid size h is optimal from a theoretical point of view because it
coincides with the error of the spatial discretization.
We also perform refinement of the coarse grid aggressively. The rationale be-
hind an efficient choice of the coarse grid is to choose the coarse grid fine enough
to get fast convergence and thus fewer iterations on the numerically more expen-
sive finer grids while maintaining moderate or negligible cost of the full derivatives
on the coarse grid compared to the few directional derivatives on the fine grid.
174 13 Nonlinear boundary control of the periodic 1D heat equation
102
100
10−2
Damping αk
k δ∗ z
Step norm k
Error z − z
10−4
Coarse grid refinement
Fine grid refinement
10−6
0 5 10 15 20 25
iteration k
penalties in cache efficiency when running on four cores. Second, the adaptive
timestepping results in different integration times on each shooting interval espe-
cially when on some intervals fast transients have to be resolved and on others not
as we can observe in Figure 13.3. Such transients can for instance be caused by
large jumps in the control variables (compare Figure 13.1). We have only imple-
mented equal distribution of the shooting intervals to the different processes. This
leads to many processes being idle until the last one has finished all its work. Op-
timal distribution of the processes is known as the job shop scheduling problem.
One simple solution heuristic is, e.g., greedy work balancing which adaptively dis-
tributes the IVPs over the available processes by assigning the currently largest job
to the first free process. The relative sizes of the jobs can be assumed to be known
from the previous SQP step. In Figure 13.4 we can see that a greedy distribution
leads to improved parallelism. A rigorous investigation of efficient parallelization
is beyond the scope of this thesis.
300 8
0 0
0 10 20 0 10 20
Shooting interval Shooting interval
Figure 13.3: Steps and integration times per shooting interval on the finest level in the solu-
tion. The solid black lines indicate the average.
core number
4
3
2
1
0 5 10 15 20 25
core number
4
3
2
1
0 5 10 15 20 25
Integration time [s]
Figure 13.4: Comparison of regular distribution of IVPs to four processes/cores (upper) and
greedy scheduling (lower). The termination time (makespan) can be significantly reduced.
The exact Hessian and the two-grid version (see Chapter 8). The quality of the
two-grid Hessian approximation is so good that we obtain the solution after 25
major iterations in both cases. Usage of a two-grid approximation yields more
evaluations of matrix vector products with the Hessian on the coarser grids but
less on the finer grids. We observe that the two-grid Hessian approximation yields
13.2 Discussion of numerical convergence 177
Table 13.2: Cumulative time [s] for simulation and IND on different mesh levels for exact
Hessian approximation with nMS = 24.
Table 13.3: Cumulative time [s] for simulation and IND on different mesh levels for two-
grid Hessian approximation with nMS = 24.
a performance increase of 84 % for the Hessian evaluation on the finest grid. The
overall wall-time savings on four cores amount to 68 % in this example.
Table 13.4: SQP iterations on each spatial discretization level and runtimes of selected parts
for varying time discretizations of the control computed on four cores.
Table 13.5: Cumulative time [s] for simulation and IND on different mesh levels for two-
grid Hessian approximation with nMS = 12.
Second, the effort for the solution of the QP subproblems increases because the
amount of linear algebra operations for the condensing step (see Chapter 8) in-
creases quadratically with nMS and because the condensed QP grows linearly in
size with nMS . Third, we can see in Tables 13.5 and 13.6 that the effort for simula-
tion and IND increases. The reason lies in the adaptivity of the IVP solver because
every jump in the controls leads to transients in the dynamic system which require
finer time steps to be resolved to the requested accuracy.
13.2 Discussion of numerical convergence 179
Table 13.6: Cumulative time [s] for simulation and IND on different mesh levels for two-
grid Hessian approximation with nMS = 48.
14 Optimal control for a bacterial
chemotaxis system
Figure 14.1: Simplified schematic of E.coli with rotating flagella for directed movement.
four cores. The IVP integrator performed between 19 and 64 integration steps per
shooting interval with an average of 27.1 steps in the solution on the finest grid.
Figure 14.2 shows a self-convergence plot. We observe that after refinement of the
fine and the coarse grid, the globalization strategy needs to recede to damped steps
for iterations 15 and 16. Afterwards only full steps are taken. Only four iterations
are performed on the finest grid level with derivatives generated on the second
coarsest level. The error plateaus are even more prominent than for the example
in Chapter 13. From extrapolation of the error at iterations
15, 23, and 27, we can
expect the final solution to be accurate to about zk − z∗ ≈ 0.1. This accuracy
is not satisfactory but on our current system and with our implementation, finer
spatial discretizations are not possible. We do not believe that this is a generic
problem of our approach because the main memory problem is that DAESOL-
II currently keeps all IND tapes (including right hand side Jacobians and their
decompositions) in main memory. This takes up the largest amount of memory
within GINKO. We are positive that this memory bottleneck can be circumvented
184 14 Optimal control for a bacterial chemotaxis system
Damping αk
Step norm δ zk
102 Error zk − z∗
Coarse grid refinement
Fine grid refinement
100
10−2
10−4
0 5 10 15 20 25 30
iteration k
Figure 14.2: Self convergence plot for the chemotaxis problem (14.1).
0.2 0.2
0.15 0.15
control q
0.1 0.1
0.05 0.05
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
time t time t
Figure 14.3: Optimal control profiles for the chemotaxis problem (14.1). The left hand
panel shows the control at the boundary x = 0 and the right hand panel at x = 1.
186 14 Optimal control for a bacterial chemotaxis system
t = 0.00 t = 0.17
2 2
dens. z, conc. c
1.5 1.5
1 1
0.5 0.5
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
t = 0.33 t = 0.50
2 2
dens. z, conc. c
1.5 1.5
1 1
0.5 0.5
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
t = 0.67 t = 0.92
2 2
dens. z, conc. c
1.5 1.5
1 1
0.5 0.5
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
t = 0.94 t = 1.00
2 2
dens. z, conc. c
1.5 1.5
1 1
0.5 0.5
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
x coordinate x coordinate
Figure 14.4: Optimal states for the chemotaxis problem (14.1). For different time points t
we plot the bacteria density z (solid line) the chemoattractant concentration c (dashed line)
and the bacteria target distribution (dash-dotted line).
15 Optimal control of a Simulated Moving
Bed process
In this chapter we describe a variant of the Simulated Moving Bed (SMB) process.
For completeness we quote in large parts from the article Potschka et al. [130].
In a chromatographic column, different components that are dissolved in a liquid
are separated due to different affinities to the adsorbent. As a result the different
components move with different velocities through the column and hence can be
separated into nearly pure fractions at the outlet. The SMB process consists of
several chromatographic columns which are interconnected in series to constitute
a closed loop (see Figure 15.1). An effective counter-current movement of the sta-
tionary phase relative to the liquid phase is realized by periodic and simultaneous
switching of the inlet and outlet ports by one column in the direction of the liq-
uid flow. Compared to batch operation of a single chromatographic column, the
SMB process offers great improvements of process performance in terms of des-
orbent consumption and utilization of the solid bed. In the basic SMB process all
Figure 15.1: SMB configuration with six columns and four zones.
flow rates are constant and the switching of the columns is simultaneous with a
fixed switching period. By introducing more degrees of freedom the efficiency of
the separation can be increased further. The flow rates for instance can be varied
during the switching periods (PowerFeed), the feed concentration can be varied
during the switching periods (ModiCon) or asynchronous switching of the ports
can be introduced (VariCol) (see Schramm et al. [144, 145]).
phases are coupled by an algebraic condition, e.g., the nonlinear extended Lang-
muir isotherm equation
Hi2 cp,i
qp,i = Hi1 cp,i + , (15.3)
1 + (k1 cp,1 + k2 cp,2 )cref
ϕ(r) = 4σi (ci (t, z) − bi (t, z))r2 + bi (t, z) − σi (ci (t, z) − bi (t, z))
For completeness we assemble here the derivatives and surface values required for
the substitution of the cp,i terms by bi :
∂ϕ ∂ bi ∂ϕ
(0.5) = , (r) = 8σi (ci − bi )r,
∂t ∂t ∂r
1 ∂ ∂ϕ
r2 = 24σi (ci − bi ), ϕ(1) = 3σi (ci − bi ) + bi .
r2 ∂ r ∂r
∂t ci = Pe−1
i ∂zz ci − ∂z ci − Sti (ci − (3σi (ci − bi )) + bi ), (15.4a)
∂t b1 −1 η1 24σ1 (c1 − b1 )
= G(b1 , b2 ) , (15.4b)
∂t b2 η2 24σ2 (c2 − b2 )
15.1 Mathematical modeling of adsorption processes 191
In the case of several connected columns we use one reference flow velocity uref
for the non-dimensionalization. For a flow velocity u j = uref in zone j = I, . . . , IV
we have to multiply the right hand sides of equations (15.1) or (15.4), respectively,
with the quotient u j /uref .
The inflow concentrations of each column are the outflow concentrations of the
preceding column, except for the column after the feed and after the desorbent
ports which can be calculated from the feed concentration cFe,i and from the out-
flow concentrations cout
.,i of the previous column according to
for i = 1, 2. With the port concentrations and the flow rates the feed, extract, and
raffinate masses, and the product purities can be calculated via
t t
mFe,i (t) = cFe,i (τ)QFe dτ, mEx,i (t) = I,i (τ, 1)QEx dτ,
cout
0 0
t
mRa,i (t) = III,i (τ, 1)QRa dτ,
cout
0
mEx,1 (t) mRa,2 (t)
PurEx (t) = , PurRa (t) = .
mEx,1 (t) + mEx,2 (t) mRa,1 (t) + mRa,2 (t)
192 15 Optimal control of a Simulated Moving Bed process
at a constant feed flow QFe but varying feed concentration cFe (t). Over one period
T the average feed concentration must be equal to the given feed concentration
cSMB
Fe of a reference SMB process.
At the end of each period the switching of ports leads to a generalized periodicity
constraint of the form
succ( j) +
cij (0, .) − ci (T, .) = 0,
i = 1, 2, j = 1, . . . , Ncol ,
succ( j)
bij (0, .) − bi (T, .) = 0,
where succ( j) denotes the index of the column which is the successor of column
j in the investigated SMB configuration.
Furthermore we require the total feed mass of one period to satisfy
mFe (T ) = cSMB
Fe QFe T, (15.5)
where cSMB
Fe is a given feed concentration of a (non-ModiCon) SMB reference
process.
The remaining constraints bound the maximum and minimum feed concentra-
tion
cFe,max ≥ cFe (t) ≥ 0
and the flow rates
Qmax ≥ QDe , QEx , QFe , QRa , QRe , QI , QII , QIII , QIV ≥ Qmin .
reagent (see, e.g., Jupke [91] as cited by Küpper [100]). Only the R-enantiomer
has pharmaceutical activity and needs to be separated from the S-enantiomer after
chemical synthesis. We list the model parameters (taken from Küpper [100]) in
Table 15.1. Further model quantities are derived from these parameters which we
display in Table 15.2.
We computed the solution with nMS = 24 shooting intervals on a two level hier-
archy of spatial FDM grids with 21 and 81 equidistant grid points for each of the
Ncol = 6 columns and each species. The relative accuracy for the time-stepping
scheme was set to 10−5 on the coarse and 10−6 on the fine level and the GINKO
termination tolerance was 5 · 10−3 .
The optimization problem becomes more and more difficult for higher values of
product purity Purmin . We had to successively generate primal starting values via a
naive homotopy approach using ascending values for Purmin = 0.8, 0.9, 0.93, 0.95
on the coarse level. For Purmin = 0.95, GINKO needed 9 iterations on the coarse
level and then 11 iterations on the fine level with coarse grid derivatives. The com-
puted κ-estimates suggest [κ̂] ≤ 0.66. For the last four iterations only two LISA
are needed for each inexact solution of the linearized systems. Table 15.3 shows
the optimal values for the ModiCon SMB separation of EMB–53986 enantiomers.
We display the optimal feed concentration profile in Figure 15.4 and the optimal
moving concentration fronts of the moving phase for one period in Figure 15.3.
We can observe that the two concentration profiles travel to the right with different
velocities and thus there is almost only slow substance present at the extract port
after column 1 and almost only fast substance present at the raffinate port after
column 5.
The solution was computed on four cores within a total wall time of 98 min. Due
to memory restrictions for the IND tape sizes, finer grid levels were not possible.
In the solution there are between 24 and 177 integration steps per shooting interval
with an average of 42.1 steps per interval.
194 15 Optimal control of a Simulated Moving Bed process
Table 15.1: Model and optimization parameters for the ModiCon SMB process to separate
EMD–53986 enantiomers.
15.2 Numerical results 195
Table 15.2: Derived model quantities for the ModiCon SMB process. Index i = 1, 2 denotes
the species, index j = I, . . . , IV the zone.
Table 15.3: Optimal values for the ModiCon SMB separation of EMB–53986 enantiomers.
196 15 Optimal control of a Simulated Moving Bed process
3 3
2 2
1 1
0 0
0 1 2 3 4 5 6 0 1 2 3 4 5 6
t = 3.56 min t = 5.34 min
4 4
Concentration [g/l]
3 3
2 2
1 1
0 0
0 1 2 3 4 5 6 0 1 2 3 4 5 6
t = 7.12 min t = 8.90 min
4 4
Concentration [g/l]
3 3
2 2
1 1
0 0
0 1 2 3 4 5 6 0 1 2 3 4 5 6
Column index Column index
Figure 15.3: Traveling concentration profiles over one period t ∈ [0, T ]. The six columns
are arranged from left to right in each panel. The feed port is located after column 3, extract
after column 1, raffinate after colum 5 and desorbent after column 6.
15.2 Numerical results 197
15
10
0
0 2 4 6 8 10
Time t
Figure 15.4: Optimal feed concentration profile cFe for ModiCon SMB separation of EMB–
53986 enantiomers. All feed mass is injected at the end of the period with maximum con-
centration cFe,max , subject to satisfaction of total feed mass constraint (15.5).
16 Conclusions and future work
In this thesis we have developed a numerical method based on Direct Multiple
Shooting for OCPs with time-periodic PDE constraints. We have achieved an
asymptotically optimal scale-up of the numerical effort with the number of spa-
tial discretization points based on inexact SQP with an inner generalized Newton-
Picard preconditioned LISA which features extensive structure exploitation in a
two-stage solution process for the possibly nonconvex QPs for which we have
developed a condensing approach and a PASM. We have implemented a numeri-
cal code called MUSCOP and have demonstrated the applicability, efficiency, and
reliability of the proposed methods on PDE OCPs from illustrating academic to
challenging real-world applications.
Our research inspires a number of now exposed questions for future research
directions and projects. We want to conclude this thesis with a list of the most
obvious ones:
Convergence analysis for PASMs for nonconvex QPs. We have developed nu-
merical techniques for the solution of nonconvex QPs with PASMs. In our
experience the resulting QP solver works unexpectedly well on these diffi-
cult problems. We believe it worthwhile to construct examples for which
it does not work or investigate proofs if none can be found. These exam-
ples, if found, might also serve as the basis for further improvement of the
numerical method.
NMT for inexact SQP. We have presented a NMT globalization strategy for the
class of LISA-Newton based inexact SQP methods. However, no proof of
convergence exists for this approach. We conjecture such a proof is possible
even for the inexact SQP case on the basis of RMT techniques.
A-posteriori mesh error estimation and mesh refinement. We have computed
all numerical examples in this thesis on uniformly refined spatial meshes.
Obviously locally refined meshes promise a great improvement of the ratio
of numerical effort vs. accuracy (see, e.g., Becker and Rannacher [14], Mei-
dner and Vexler [113], Hesse [83]). Furthermore, we should refine the fine
and coarse grid according to two different goals: The fine grid for highest ac-
curacy (in terms of the question which inspires the problem) and the coarse
grid for best contraction, i.e., smallest κ. Moreover the required global er-
ror estimators should be exploited to trigger fine grid refinement as soon
as the inexact Simplified Newton increment becomes smaller than the grid
interpolation error.
[3] J. Albersmeyer and H.G. Bock. Efficient sensitivity generation for large
scale dynamic systems. Technical report, SPP 1253 Preprints, University of
Erlangen, 2009.
[4] J. Albersmeyer and M. Diehl. The Lifted Newton method and its application
in optimization. SIAM Journal on Optimization, 20(3):1655–1684, 2010.
[5] J. Albersmeyer and C. Kirches. The SolvIND webpage, 2007. URL http:
//www.solvind.org.
[6] M. Arioli, I.S. Duff, N.I.M. Gould, J.D. Hogg, J.A. Scott, and H.S. Thorne.
The HSL mathematical software library, 2007.
[7] U. Ascher and M.R. Osborne. A note on solving nonlinear equations and the
natural criterion function. Journal of Optimization Theory and Applications,
55(1):147–152, 1987.
[10] A. Battermann and E.W. Sachs. Block preconditioners for KKT systems
in PDE-governed optimal control problems. In Fast solution of discretized
optimization problems (Berlin, 2000), volume 138 of Internat. Ser. Numer.
Math., pages 1–18. Birkhäuser, Basel, 2001.
[12] I. Bauer, H.G. Bock, S. Körkel, and J.P. Schlöder. Numerical methods for
initial value problems and derivative generation for DAE models with appli-
cation to optimum experimental design of chemical processes. In Scientific
Computing in Chemical Engineering II, pages 282–289. Springer, 1999.
[13] I. Bauer, H.G. Bock, and J.P. Schlöder. DAESOL – a BDF-code for the
numerical solution of differential algebraic equations, 1999.
[17] M.J. Best. An algorithm for the solution of the parametric quadratic pro-
gramming problem. In H. Fischer, B. Riedmüller, and S. Schäffler, editors,
Applied Mathematics and Parallel Computing – Festschrift for Klaus Ritter,
chapter 3, pages 57–76. Physica-Verlag, Heidelberg, 1996.
[25] H.G. Bock and K.J. Plitt. A Multiple Shooting algorithm for di-
rect solution of optimal control problems. In Proceedings of the
9th IFAC World Congress, pages 242–247, Budapest, 1984. Pergamon
Press. URL http://www.iwr.uni-heidelberg.de/groups/agbock/
FILES/Bock1984.pdf.
[26] H.G. Bock, E.A. Kostina, and J.P. Schlöder. On the role of natural level
functions to achieve global convergence for damped Newton methods. In
M.J.D. Powell and S. Scholtes, editors, System Modelling and Optimization.
Methods, Theory and Applications, pages 51–74. Kluwer, 2000.
204 Bibliography
[27] H.G. Bock, W. Egartner, W. Kappis, and V. Schulz. Practical shape opti-
mization for turbine and compressor blades by the use of PRSQP methods.
Optimization and Engineering, 3(4):395–414, 2002.
[28] H.G. Bock, A. Potschka, S. Sager, and J.P. Schlöder. On the connection be-
tween forward and optimization problem in one-shot one-step methods. In
G. Leugering, S. Engell, A. Griewank, M. Hinze, R. Rannacher, V. Schulz,
M. Ulbrich, and S. Ulbrich, editors, Constrained Optimization and Opti-
mal Control for Partial Differential Equations, volume 160 of International
Series of Numerical Mathematics, pages 37–49. Springer Basel, 2011.
[30] J.H. Bramble and J.E. Pasciak. A preconditioning technique for indefinite
systems resulting from mixed approximations of elliptic problems. Mathe-
matics of Computation, 50(181):1–17, 1988.
[33] R.H. Byrd, F.E. Curtis, and J. Nocedal. An inexact SQP method for equality
constrained optimization. SIAM Journal on Optimization, 19(1):351–369,
2008.
[35] J. Dahl and L. Vandenberghe. CVXOPT user’s guide, release 1.1.3, 2010.
URL http://abel.ee.ucla.edu/cvxopt/userguide/index.html.
[41] R.S. Dembo, S.C. Eisenstat, and T. Steihaug. Inexact Newton methods.
SIAM Journal on Numerical Analysis, 19(2):400–408, 1982.
[45] P. Deuflhard, R. Freund, and A. Walter. Fast secant methods for the iterative
solution of large nonsymmetric linear systems. IMPACT of Computing in
Science and Engineering, 2(3):244–276, 1990.
[50] N. Dunford and J.T. Schwartz. Linear operators part I: General theory. In
R. Courant, L. Bers, and J.J. Stoker, editors, Pure and applied mathematics,
volume VII. Wiley, New York, 1958.
[51] J. Fernández, M. Anguita, E. Ros, and J.L. Bernier. SCE toolboxes for the
development of high–level parallel applications. In Proceedings of the 6th
International Conference Computational Science – ICCS 2006, Reading,
United Kingdom, Part II, volume 3992, pages 518–525, 2006.
[52] H.J. Ferreau. An online active set strategy for fast solution of para-
metric quadratic programs with applications to predictive engine control.
Diploma thesis, Ruprecht–Karls–Universität Heidelberg, 2006. URL http:
//homes.esat.kuleuven.be/~jferreau/pdf/thesisONLINE.pdf.
[53] H.J. Ferreau, H.G. Bock, and M. Diehl. An online active set strategy to
overcome the limitations of explicit MPC. International Journal of Robust
and Nonlinear Control, 18(8):816–830, 2008.
[56] C.A. Floudas and P.M. Pardalos, editors. State of the art in global opti-
mization: computational methods and applications. Springer, Dordrecht,
1995.
[58] E.M. Gertz and S.J. Wright. Object-oriented software for quadratic pro-
gramming. ACM Transactions on Mathematical Software, 29:58–81, 2003.
Bibliography 207
[59] P.E. Gill, W. Murray, M.A. Saunders, and M.H. Wright. Procedures for op-
timization problems with a mixture of bounds and general linear constraints.
ACM Transactions on Mathematical Software, 10(3):282–298, 1984.
[60] P.E. Gill, W. Murray, and M.A. Saunders. User’s Guide For QPOPT 1.0: A
Fortran Package For Quadratic Programming, 1995. URL http://www.
sbsi-sol-optimize.com/manuals/QPOPT%20Manual.pdf.
[61] G.H. Golub and C.F. van Loan. Matrix Computations. Johns Hopkins Uni-
versity Press, Baltimore, 3rd edition, 1996.
[63] N.I.M. Gould. On practical conditions for the existence and uniqueness of
solutions to the general equality quadratic programming problem. Mathe-
matical Programming, 32(1):90–99, 1985.
[65] N.I.M. Gould and Ph.L. Toint. Nonlinear programming without a penalty
function or a filter. Mathematical Programming, Series A, 122:155–196,
2010.
[66] N.I.M. Gould and P.L. Toint. A quadratic programming bibliography. Tech-
nical Report 2000-1, Rutherford Appleton Laboratory, Computational Sci-
ence and Engineering Department, 2010.
[67] N.I.M. Gould, M.E. Hribar, and J. Nocedal. On the solution of equality
constrained quadratic programming problems arising in optimization. SIAM
Journal on Scientific Computing, 23(4):1376–1395, 2001.
[68] N.I.M. Gould, D. Orban, and P.L. Toint. CUTEr testing environment for
optimization and linear algebra solvers, 2002. URL http://cuter.rl.
ac.uk/cuter-www/.
[69] N.I.M. Gould, D. Orban, and Ph.L. Toint. GALAHAD, a library of thread-
safe Fortran 90 packages for large-scale nonlinear optimization. ACM
Transactions on Mathematical Software, 29(4):353–372, 2004.
208 Bibliography
[75] W. Gropp, E. Lusk, and A. Skjellum. Using MPI: Portable Parallel Pro-
gramming with the Message Passing Interface. Scientific and Engineering
Computation Series. MIT Press, Cambridge, MA, USA, 2nd edition, 1999.
[80] S.B. Hazra, V. Schulz, J. Brezillon, and N.R. Gauger. Aerodynamic shape
optimization using simultaneous pseudo-timestepping. Journal of Compu-
tational Physics, 204(1):46–64, 2005.
[83] H.K. Hesse. Multiple Shooting and Mesh Adaptation for PDE Constrained
Optimization Problems. PhD thesis, University of Heidelberg, 2008.
[85] J.S. Hesthaven, S. Gottlieb, and D. Gottlieb. Spectral methods for time-
dependent problems, volume 21 of Cambridge Monographs on Applied
and Computational Mathematics. Cambridge University Press, Cambridge,
2007.
[87] D. Horstmann. From 1970 until present: The Keller-Segel model in chemo-
taxis and its consequences I. Jahresbericht der DMV, 105(3):103–165,
2003.
[88] D. Horstmann. From 1970 until present: The Keller-Segel model in chemo-
taxis and its consequences II. Jahresbericht der DMV, 106(2):51–69, 2004.
[89] IBM ILOG. IBM ILOG CPLEX V12.1, User’s Manual for CPLEX. IBM
Corp., New York, USA, 2009.
[93] Y. Kawajiri and L.T. Biegler. Optimization strategies for Simulated Moving
Bed and PowerFeed processes. AIChE Journal, 52(4):1343–1350, 2006.
[94] Y. Kawajiri and L.T. Biegler. Large scale nonlinear optimization for asym-
metric operation and design of simulated moving beds. Journal of Chro-
matography A, 1133:226–240, 2006.
[95] E.F. Keller and L.A. Segel. Model for chemotaxis. Journal of Theoretical
Biology, 30(2):225–234, 1971.
[97] C. Kirches, H.G. Bock, J.P. Schlöder, and S. Sager. Block structured
quadratic programming for the direct multiple shooting method for optimal
control. Optimization Methods and Software, 26(2):239–257, 2011.
[98] E. Kostina. The long step rule in the bounded-variable dual simplex method:
Numerical experiments. Mathematical Methods of Operations Research, 55
(3):413–429, 2002.
[101] S. Lauer. SQP–Methoden zur Behandlung von Problemen mit indefiniter re-
duzierter Hesse–Matrix. Diploma thesis, Ruprecht–Karls–Universität Hei-
delberg, 2010.
[102] D. Lebiedz and U. Brandt-Pollmann. Manipulation of Self-Aggregation
Patterns and Waves in a Reaction-Diffusion System by Optimal Boundary
Control Strategies. Physical Review Letters, 91(20):208301, 2003.
[103] R.B. Lehoucq and D.C. Sorensen. Deflation techniques for an implicitly
restarted Arnoldi iteration. SIAM Journal on Matrix Analysis and Applica-
tions, 17(4):789–821, 1996.
[104] R.B. Lehoucq, D.C. Sorensen, and C. Yang. ARPACK Users’ Guide: Solu-
tion of Large–Scale Eigenvalue Problems with Implicitly Restarted Arnoldi
Methods. Society for Industrial and Applied Mathematics (SIAM), 1998.
[105] D.B. Leineweber. The theory of MUSCOD in a nutshell. IWR-Preprint
96-19, Universität Heidelberg, 1996.
[106] D.B. Leineweber. Efficient reduced SQP methods for the optimization of
chemical processes described by large sparse DAE models, volume 613 of
Fortschritt-Berichte VDI Reihe 3, Verfahrenstechnik. VDI Verlag, Düssel-
dorf, 1999.
[107] D.B. Leineweber, I. Bauer, A.A.S. Schäfer, H.G. Bock, and J.P. Schlöder.
An efficient multiple shooting based reduced SQP strategy for large-scale
dynamic process optimization (Parts I and II). Computers and Chemical
Engineering, 27:157–174, 2003.
[108] R.J. LeVeque. Finite volume methods for hyperbolic problems. Cambridge
Texts in Applied Mathematics. Cambridge University Press, Cambridge,
2002.
[109] R.J. LeVeque. Finite difference methods for ordinary and partial differen-
tial equations. Society for Industrial and Applied Mathematics (SIAM),
Philadelphia, PA, 2007.
[110] K. Lust, D. Roose, A. Spence, and A. R. Champneys. An adaptive Newton-
Picard algorithm with subspace iteration for computing periodic solutions.
SIAM Journal on Scientific Computing, 19(4):1188–1209, 1998.
[111] I. Maros and C. Mészáros. A repository of convex quadratic programming
problems. Optimization Methods and Software, 11:671–681, 1999.
212 Bibliography
[126] L.S. Pontryagin, V.G. Boltyanski, R.V. Gamkrelidze, and E.F. Miscenko.
The Mathematical Theory of Optimal Processes. Wiley, Chichester, 1962.
[128] A. Potschka, H.G. Bock, and J.P. Schlöder. A minima tracking variant of
semi-infinite programming for the treatment of path constraints within di-
rect solution of optimal control problems. Optimization Methods and Soft-
ware, 24(2):237–252, 2009.
[129] A. Potschka, C. Kirches, H.G. Bock, and J.P. Schlöder. Reliable so-
lution of convex quadratic programs with parametric active set meth-
ods. Technical Report 2010–11–2828, Heidelberg University, Interdisci-
plinary Center for Scientific Computing, Heidelberg University, Im Neuen-
heimer Feld 368, 69120 Heidelberg, Germany, 2010. URL http://www.
optimization-online.org/DB_HTML/2009/02/2224.html.
[130] A. Potschka, A. Küpper, J.P. Schlöder, H.G. Bock, and S. Engell. Opti-
mal control of periodic adsorption processes: The Newton-Picard inexact
SQP method. In Recent Advances in Optimization and its Applications in
Engineering, pages 361–378. Springer Verlag, 2010.
[131] A. Potschka, M.S. Mommer, J.P. Schlöder, and H.G. Bock. A Newton-
Picard approach for efficient numerical solution of time-periodic parabolic
PDE constrained optimization problems. Technical Report 2010–03–
2570, Interdisciplinary Center for Scientific Computing (IWR), Heidelberg
University, 2010. URL http://www.optimization-online.org/DB_
HTML/2010/03/2570.html.
[132] S.M. Robinson. Perturbed Kuhn-Tucker points and rates of convergence for
a class of nonlinear programming algorithms. Mathematical Programming,
7:1–16, 1974.
[133] Y. Saad. Numerical methods for large eigenvalue problems. Algorithms and
Architectures for Advanced Scientific Computing. Manchester University
Press, Manchester, 1992.
[135] Y. Saad. Iterative Methods for Sparse Linear Systems. SIAM, Philadelpha,
PA, 2nd edition, 2003.
[136] Y. Saad and M.H. Schultz. GMRES: A generalized minimal residual algo-
rithm for solving nonsymmetric linear systems. SIAM Journal on Scientific
and Statistical Computing, 7:856–869, 1986.
[157] R.J. Vanderbei. LOQO: An interior point code for quadratic programming.
Optimization Methods and Software, 11(1–4):451–484, 1999.
[161] T. Warburton and M. Embree. The role of the penalty in the local dis-
continuous galerkin method for Maxwell’s eigenvalue problem. Computer
Methods Applied Mechanical Engineering, 195:3205–3223, 2006.
[162] A.J. Wathen. Realistic eigenvalue bounds for the Galerkin mass matrix.
IMA Journal of Numerical Analysis, 7(4):449–457, 1987.
[163] R.C. Whaley, A. Petitet, and J.J. Dongarra. Automated empirical optimiza-
tion of software and the ATLAS project. Parallel Computing, 27(1–2):3–35,
2001.
[165] J.H. Wilkinson. The Algebraic Eigenvalue Problem. Clarendon Press, Ox-
ford, 1965.
[166] L. Wirsching. An SQP algorithm with inexact derivatives for a Direct Mul-
tiple Shooting method for optimal control problems. Diploma thesis, Uni-
versität Heidelberg, 2006.
[168] X. Zhang and Y. Ye. User’s guide of COPL QP, computational optimization
program library: Convex quadratic programming, 1998. URL http://
www.stanford.edu/~yyye/Col.html.