Existence and Regularity of Solutions
Existence and Regularity of Solutions
where, as usual, u(k) signifies the kth derivitave with respect to the inde-
pendent variable x. We can only expect this equation to have a unique
solution if we supply, in addition to the equation, other data specifying the
value of u and its first n − 1 derivatives at some value x0 of the independent
variable. Moreover, in general, we can only expect that a solution exists on
some interval, possibly small, containing the point x0 , as examples show.
Example 6.0.3 The first-order equation u0 = 2xu2 with initial data u (0) =
−1
u0 has the solution u = u0 1 − u0 x2 . If u0 > 0 this becomes infinite at
−1/2
x = ±u0 , which defines the interval of existence for this solution.
Unless otherwise indicated, we shall assume in this chapter that the vari-
ables, both dependent and independent, are real.
131
132
The standard technique for converting the nth order equation (6.1) to a
system of n first-order equations is to write
U 0 = F (x, U ) (6.4)
where the vector field F appearing on the right-hand side need not have the
special form of equation (6.3), but can have components each of which is an
arbitrary function of the n + 1 variables x, u1 , . . . , un . Conclusions drawn
for such a general system are then necessarily valid for the special system
(6.3), from which they can be immediately translated into the language of
the nth-order equation (6.1).
Among the advantages of the system approach is that its treatment is
independent of n, and therefore its treatment for arbitrary n closely mimics
that of a single first-order equation, n = 1. We therefore investigate this
relatively simple case first.
where x and y are real variables. We shall assume that the function f is
continuous in a domain1 D of the xy-plane, and satisfies a Lipschitz condi-
tion there, as defined in Chapter 1, equation (1.49). The basic conclusion,
sometimes called local existence theory, is that there is an interval I of the
x axis, possibly very small, containing the point x0 , on which there is a
solution y (x). We address later the issue of extending this local solution to
a longer interval.
1
By a domain we understand a connected, open set.
133
in the sense that any continuous solution of the latter is a solution of the
former, and conversely. The plan is to construct a solution of the latter.
A number of ideas from real analysis are used in the proof of the basic
existence theorem below. One of the most important of these is idea of
uniform convergence of a sequence of functions. For a sequence of real
numbers, ak , say, the idea of convergence is expressed as follows: there is
a real number a (the limit of the sequence) with the property that, given
any positive number , there is an integer N such that |an − a| < for all
n > N . The notion of the convergence of a sequence of functions uk , each
defined on some interval I, to a function u on that interval is the same: for
each x ∈ I, the sequence of real numbers {uk (x)} should converge to the
real number u(x). In this convergence, the number N in general depends on
the choice of x ∈ I. A sequence of functions {uk } of functions defined on I
is said to converge uniformly on I if it converges to some function u there
and, given > 0, it is possible to find an integer N such that, for any x ∈ I,
|un (x) − u (x)| < if n ≥ N. The uniformity of the convergence consists in
that the choice of N depends only on and does not depend on x.
P The definition of uniform convergence for an infinite series of functions
vn (x) follows by considering the sequence of partial sums of the series.
A convenient criterion for the uniform convergence of the infinite
P series on
a set S is the existence of a convergent series of constants Mn such that
|vn (x)| ≤ Mn for each x in S. This criterion, the so-called Weierstrass
M -test, may be found (for example) in the texts by Taylor2 or Rudin3 .
Another idea used below is that of uniform continuity. A function u (x)
is uniformly continuous on an interval I if, given > 0, there is δ > 0 such
that |u (x) − u (y)| < for all points x, y ∈ I such that |x − y| < δ. Here
the uniformity consists in that the choice of δ depends only on and does
not depend on x or y.
Neither of these notions is restricted to functions of a single real variable,
but have immediate generalizations to functions of several variables.
Standard theorems used below, for which we refer the reader to textbooks
on real analysis (or advanced calculus), are the following:
2
Advanced Calculus
3
Principles of Mathematical Analysis
134
as follows:
Z x
y0 (x) = y0 , yn (x) = y0 + f (s, yn−1 (s)) ds , n = 1, 2, . . . . (6.7)
x0
since x − x0 ≤ α ≤ b/M. In the case when x < x0 the integral above would
taken from x to x0 and the conclusion would be the same except that x − x0
135
Z
|y2 (x) − y1 (x)| ≤ |f (s, y1 (s)) − f (s, y0 (s))| ds
(x − x0 )2
Z x Z x
≤L |y1 (s) − y0 (x)| ds ≤ LM (s − x0 ) ds = M L ,
x0 x0 2
where L is the Lipschitz constant (cf. the definition 1.4.2). We propose the
induction hypothesis that on the interval [x0 , x0 + α]
M (L (x − x0 ))n+1
|yn+1 (x) − yn (x)| ≤ . (6.9)
L (n + 1)!
We have verified this for the cases n = 0 and n = 1. Assume it’s true for a
value n − 1. Then we find
(s − x0 )n
Z x Z
|yn+1 (x) − yn (x)| ≤ L |yn (s) − yn−1 (s)| ds ≤ L M Ln−1 ds
x0 n!
n (x− x0 )n+1
= ML ,
(n + 1)!
P
The sequence {yn (x)} will converge if the corresponding series (yn+1 (x) − yn (x))
converges. But the latter is dominated by the series of positive constants
(M/L) (Lα)n /n! which is the convergent expansion for (M/L) exp (Lα) .
P
The sequence {yn (x)} therefore converges uniformly on the interval [x0 , x0 +
α] by the Weierstrass M -test. The restriction to this subinterval was for con-
venience in expressing the estimates; if we restrict to the interval [x0 − α, x0 ]
we find the same estimates, so we conclude that this sequence is uniformly
convergent on the entire interval I. We denote its limit function by y (x).
It is continuous by Theorem 6.1.1, and it satisfies the condition
Z x
y (x) = y0 + lim f (s, yn (s)) ds.
n→∞ x
0
The limit can be taken inside the integral if the hypotheses of Theorem
6.1.2 hold. But the integrand Fn (s) ≡ f (s, yn (s)) converges uniformly to
f (s, y (s)) by virtue of the continuity of the function f 4 . This verifies the
hypotheses of that theorem, and shows that the function y (x) so constructed
satisfies (6.6).
The uniqueness of this solution was established in Chapter 1 as Theorem
1.4.2.
In this theorem, the function f need not satisfy a Lipschitz condition on
the entire domain D. It suffices if it satisfies this condition on each closed
rectangle R ⊂ D since only this assumption was used in the proof.
{(xn , y(xn ))} lies in D the point (b, yb ) lies either in D or on ∂D. In fact
it must lie on ∂D for, if it were to lie in D, then solving the initial-value
problem with initial data (b, yb ) would extend the interval of existence to
the right of b, contradicting the assumption that b is right-maximal. We can
summarize this as follows:
Theorem 6.1.5 The solution of the initial-value problem (6.5) can be con-
tinued to the right to a maximal interval of existence [x0 , b) such that at least
one of the following holds:
1. b = +∞,
2. y (x) is unbounded as x → b, or
3. (x, y (x)) approaches the boundary of D as x → b.
Of course, the same considerations apply to the left. This provides a maxi-
mally extended interval of existence (a, b) for the solution of the initial-value
problem.
The uniqueness of solutions as stated in Theorem 6.1.4 applies to solu-
tions defined on possibly small intervals containing the initial point (x0 , y0 ),
but it extends to solutions on arbitrary intervals: either two solutions dis-
agree at every point of their common interval of existence or they are iden-
tical.
|U + V | ≤ |U | + |V | . (6.10)
Such a real-valued function of the vector U is called a norm. There are
various choices when n > 1. Two popular choices are:
• The Euclidean norm
q
|U | = U12 + U22 + · · · + Un2 . (6.11)
138
where, on the right-hand side, |Ui | signifies the usual absolute value of
a real number.
The Lipschitz condition is now stated in terms of the norm (we assume one
has been chosen). It takes the following form:
Proof: The proof of existence is virtually word-for-word the same as that for
theorem 6.1.4 if the norm is everywhere substituted for the absolute value;
this proof is therefore omitted. The proof of uniqueness is also virtually
word-for-word the same as that given in Chapter 1 for the case n = 1 again
substituting the norm for the absolute value; this proof is therefore likewise
omitted.
Theorem 6.2.1 is, like Theorem 6.1.4, a local existence theorem. How-
ever, the argument leading to the continuation of solutions to a maximal
interval of existence, Theorem 6.1.5, is independent of n and carries over to
the present case without modification. We state it, for the record:
Theorem 6.2.2 Under the conditions of Theorem 6.2.1, the solution can be
continued to a maximal interval (a, b). The possible behaviors of the solution
at either of these endpoints are described in Theorem 6.1.5.
The right-hand side of this equation has the form F (x, U ) of the existence
theorem 6.2.1, and we can conclude local existence provided that F is con-
tinuous and satisfies a Lipschitz condition with respect to U . If the matrix-
valued function A has entries that are continuous on the finite interval [a, b],
and if the components of R are likewise continuous there, then the function
F = AU + R is indeed continuous. To verify the Lipschitz condition, note
that
|AU | ≤ α |U | , (6.18)
which shows that F satisfies a Lipschitz condition.
We can now infer from the basic existence theorem that the initial-value
problem (6.16) possesses a solution on some interval, and therefore on a
maximal interval. It remains to show that the maximal interval coincides
with the prescribed interval [a, b] on which the coefficients are defined and
continuous.
Since each component of R is bounded on [a, b] we infer that there is a
positive number β such that
|R (x)| ≤ β (6.19)
for each x in [a, b]. Now consider in place of the initial-value problem (6.16)
the equivalent integral equation
Z x
U (x) = U0 + (A (s) U (s) + R (s)) ds. (6.20)
x0
Taking norms and defining φ (x) = |U (x)| leads to the inequality (for x >
x0 )
Z x Z x
φ (x) ≤ φ (x0 ) + (αφ (s) + β) ds ≤ K + α |U (s)| ds, (6.21)
x0 x0
i.e., U is bounded for all x in the interval [x0 , b]. It now follows from Theorem
6.2.2 that the graph (x, U (x)) extends to the boundary of the domain. Since
the boundaries are x = a and x = b in this case, the right-maximal interval
is [x0 , b).
This argument shows that the solution exists on [x0 , b) but, as it stands,
does not continue the solution to the closed interval: it is possible in principle
for the graph to approach the boundary without approaching a particular
point of the boundary. In fact, for the linear system (6.16), it does approach
a point of the boundary. For, from the equivalent integral formulation we
infer
|U (x̃) − U (x)| ≤ α |x̃ − x| .
Therefore, if {xn } is a sequence such that xn → b, then
Theorem 6.2.3 Let the matrix A and the vector function R be continuous
on the closed, bounded interval [a, b]. Then the solution of the initial-value
problem 6.16 exists on that interval.
1. Prove the statement following equation (6.6) to the effect that a continuous
solution of this equation is equivalent to a solution of the initial-value problem
(6.5).
2. For the initial-value problem y 0 = λy, y (0) = 1, formulate the equivalent
integral equation and find the first three approximations y0 , y1 , y2 in the
successive-approximations approach to a solution.
3. Same as the preceding exercise for the initial-value problem y 0 = y 2 , y (0) =
1.
Rx
4. Solve the integral equation u (x) = 1 + 0 s2 u (s) ds.
5. Suppose that f (x, y) is continuous on a closed, bounded set S, and that
the sequence of functions {yn (x)} converges uniformly on an interval I to
a function y (x). Suppose that (x, yn (x)) ∈ S for x ∈ I. Show that the
sequence Fn (x) = f (x, yn (x)) converges uniformly to f (x, y (x)) .
142
9. Take n = 2 and denote the norm (6.11) by k · k and the norm (6.12) by | · |.
Show that these norms are equivalent in the sense that there exist positive
constants a, b, c, d such that, for any vector U = (U1 , U2 ),
11. Prove that each of the two norms (6.11) and (6.12) satisfies the triangle
inequality.
For µ > 0 and arbitrary initial-value vector w0 , show that solutions exist on
(−∞, +∞).
Hint: Find a positive function of w1 ,w2 and w3 that is constant on solutions.
143
We consider this on an interval such that the graphs of both solutions lie in
D. Let F satisfy a Lipschitz condition there with Lipschitz constant L, and
suppose the difference F (x, U ) − G (x, U ) is bounded there, by a constant
M . Then
Z x
kU (x) − V (x)k ≤ (kU0 − V0 k + M (b − a)) + L kU (s) − V (s)k ds.
x0
144
6.4 Differentiablity
In order to be able to infer that the solution map is not only a continuous but
also a differentiable function of the initial data or of parameters, we must
place further requirements on the function f of the initial-value problem
(6.5) (in the case n = 1) or of the function F of problem (6.15) (for the
general case). We consider the treatment of the case n = 1, considering
only the differentiability with respect to initial data (but see the problem
set for an extension to differentiability with respect to parameters). Let
y = φ (x, y0 ) be the solution map. It satisfies
If these procedures are justified, the function φy0 satisfies the following linear
initial-value problem:
with a (x) = fy (x, φ (x, y0 )) . We are thus led to a linear equation for the
derivative (this is called the variational equation). Of course, for n = 1 we
can obtain a formula for its solution, but this is not our goal. Our goal is
to justify this procedure. The first observation is that it is not sufficient to
assume that f is continuous in D and satsifies a Lipschitz condition there.
We have taken a partial derivative with respect to y. We therefore make the
stronger assumption that f has a continuous partial derivative with respect
to y in the domain D.
With the aid of this assumption we can complete the proof that φy0
exists, and is equal to the solution of the variational equation, along the
following lines. Denote by ∆φ the difference
The object is to show that the limit of q exists as h → 0 and equals the
solution v of the variational equation (6.27). We proceed in two steps, of
which the first is to show that q is bounded as h → 0.
For ∆φ we have the integral equation
Z x
∆φ (x, y0 , h) = h + [f (s, φ (s, y0 + h)) − f (s, φ (s, y0 ))] ds
x0
Z x
=h+ [f (s, φ (s, y0 ) + ∆φ (s, y0 , h)) − f (s, φ (s, y0 ))] ds
y0
Z x
=h+ fy (s, φ (s, y0 ) + θ∆φ) ∆φ ds.
y0
In the last term above, we have used the mean-value theorem. The number
θ = θ (s, y0 , h) lies in the interval (0, 1). We have in this last term suppressed
the arguments of the function ∆φ = ∆φ (s, y0 , h) . From the uniform conti-
nuity of φ (x, y) we know that we may choose |∆φ| as small as we please,
by choosing h sufficiently small. Since by assumption fy is continuous, it
is bounded on any compact subset. We may confine its arguments to a
compact susbset of the domain D by choosing h small enough, and may
therefore assume that |fy | < M for some positive number M . Therefore
Z x
|∆φ| ≤ |h| + M |∆φ (s, y0 , h)| ds,
y0
This shows that ∆φ tends to zero like h or, equivalently, that q is bounded
as h → 0.
Now form the difference
q (x, y0 , h) − v (x)
Z x
f (s, φ (s, y0 + h)) − f (s, φ (s, y0 ))
= − fy (s, φ (s, y0 )) v (s) ds
x0 h
Z x Z x
= q [fy (s, φ + θ∆φ) − fy (s, φ)] ds + fy (s, φ) (q (s) − v (s)) ds.
y0 y0
147
The uniform continuity of fy shows that the term in square brackets in the
first integral can be made less than (say) if h is chosen sufficiently small,
and since q is bounded, we now find from Gronwall’s lemma that
In the statement of the theorem above, we have used the notation U (x, x0 , U0 )
for the solution to emphasize the dependence of the solution on the initial
values of both x and U . However, we have only considered the derivative
with respect to U0 . It is natural to inquire whether the dependence on x0
is continuous and differentiable as well under the appropriate conditions on
the vector field F . The answer is yes, but we do not pursue this further (but
see Problem 7 below).
Theorem 6.4.1 can be generalized to the case when the vector field f de-
pends differentiably on parameters. We easily obtain (see Problem 6 below)
Theorem 6.4.2 Consider the initial-value problem (6.25). Assume that F
is continuous on D × J and possesses continuous partial derivatives with
respect to U and µ there. Assume further that, if µ = µ0 , a solution W (x)
exists on [a, b]. Then there exists δ > 0 such that, for all (x0 , U0 ) in the set
Choose a basis u1 , u2 such that u1 (0) = 1, u01 (0) = 0 and u2 (0) = 0, u02 (0) =
1. Write the solution u (x, u0 , u00 ). Is it a continuous function of the inital
data? Does it possess partial derivatives with respect to the initial data?
4. Consider the linear initial-value problem on [1, 2] with parameter β:
where
−δ 1
A(δ, µ) =
0 −µ
and δ and µ are positive parameters.
(a) Solve this problem explicitly when µ = δ.
(b) Solve it when µ 6= δ.
(c) Show explicitly that, for any fixed t > 0, if we take the limit as µ → δ,
the two solutions become the same.
9. In equation (6.22) of Problem 13 of Problem Set 6.2.1, take
0
w0 = 0 .
1
[2] P.F. Byrd and M. D. Friedman. Handbook of elliptic integrals for engi-
neers and physicists. Springer, Berlin, 1954.
315
316