Calculus of Varitation and Integral Equation
Calculus of Varitation and Integral Equation
Equations
DEMTH655
Edited by:
Dr. Kulwinder Singh
Calculus of variation and
Integral Equations
Edited By
Dr. Kulwinder Singh
CONTENTS
Objectives
Students will be able to
Introduction
In real world, the solution of physical problems can be found by using the theory of ordinary or
partial differential equations or by the theory of integral equations. The solution of integral
equations is much easier than the original boundary or initial value problems.
Several situations of science lead to integral equations e.g., neutron diffusion problem, radiation
transfer problem etc.For example, while searching for the representation formula for the solution of
linear differential equation in such a manner to include boundary conditions or initial conditions
explicitly, an integral equation is obtained.
The Abel Prize is intended to give the mathematicians their own equivalent of a Nobel Prize.Today,
every mathematics undergraduate encounter Abel's name in connection with commutative groups,
which are more commonly known as "abelian groups."
Consider the following problem of mechanics.
Consider a smooth curve in a vertical plane and suppose a material point starts from rest at any
point P under the influence of gravity along the curve. Let T be the time taken by the particle from
P to the lowest point O.Treat O as the origin of coordinate axes; y-axis being horizontal and x-axis
being vertically upwards.
Let the coordinates of points P and Q be (x, y) and , respectively.Let the arc OQ= s. Then, at
the particle at Q, we have
1
= ℎ
2
=2 ℎ
=± 2 ℎ
=± 2 ( − )
We consider
= − 2 ( − )
=−
2 ( − )
= … (1)
2 ( − )
If the shape of the curve is given, then s can be expressed in terms of , and = ( )
(1) => =
( )
… (2)
2 ( − )
Abel treated this problem in modified form by finding the curve for which the time of descent T is a
given function of , say ( ). Thus, we are led to a problem of finding the unknown function
from the equation,
( )
( )= … (3)
2 ( − )
An integral equation is an equation in which an unknown function appears under one or more
integral signs.
e.g. For ≤ ≤ , ≤ ≤ , ∫ ( , ) ( ) = ( )
( )− ( , ) ( ) = ( )
And ( ) = ∫ ( , )[ ( )]
Where ( ) is an unknown function, ( ) is a known function ( , ) is a known functionand
, , are the constants.
The above mentioned functions may be complex valued functions of the real variables and .
An integral equation is called linear if only linear operations are performed in it upon the unknown
function. An integral equation which is not linear is known as a non-linear integral equation.
{ ( )+ ( )} = { ( )} + { ( )}
The most general type of the linear integral equation is of the form:
( ) ( )= ( )+ ( , ) ( )
Let ( ) = ∫ ( , ) ( )
And { ( )+ ( )} = { ( )} + { ( )}
Remarks:
1. can be incorporated into the kernel ( , ).
2. If ( ) is non zero then the equation ( ) ( ) = ( ) + ∫ ( , ) ( ) is known as linear
( )+ ( , ) ( ) =0
( )= ( )+ ( , ) ( )
V. Volterra, theItalian Mathematician and E. Fredholm, the Swedish mathematician are credited for
their work on integral equations.
( ) ( )= ( )+ ( , ) ( )
Special Cases
( ) ( )= ( )+ ( , ) ( )
Indefinite integrals
If the differential of x3 is 3x2, then
3x .dx = x
2 3
But 3x2 is also the differential of x3- 1 and x3+ 8, etc. so that this reversal is not unique - we've 'lost'
the constant! So in general, 3x2 is the differential of (x3+ k) where k is any constant – this is known as
(If we’re given more information, we can work out the value for k).
Since integration is the reverse of differentiation, for any polynomial y(x) = xn, we can simply
reverse the differentiation procedure, so that the integral is given by
x .dx = x .dx
½ =
( / )
/
=
√
+k
abx .dx =
3
ab x3.dx= ab +k
x 2
x13 .dx = x .dx
-3 =
2
= -
1
2x 2
+k
2m2.dm =
2m 3
3
+k
10 3
5 .d = +k
3
Definite Integrals
We now know how to integrate simple polynomials, but if we want to use this technique to
calculate areas, we need to know the limits of integration. If we specify the limits x = ax = b, we
call the integral a definite integral.
To solve a definite integral, we first integrate the function as before (i.e. find its indefinite integral),
then feed in the 2 values of the limits. Subtracting one from the other gives the area.
What is the area under the curve y(x) = 2x2 between x=1 and x=3?
Area =∫ 2 .
x 3
2x3
= 3 k
x 1
= (18 + k) - (2/3 + k)
= 18 -2/3
What is the area under the curve y(x) = 2x3- 6x between x = -1 and x = 0?
A=∫ 2 −6 . y
4
3
area
0
x4 2
2 3x
2
=
1 1
0 x
= (0 - 0) - (½ - 3) -1
-2
= 2½ sq.units
-3
-4
-2 -1 0 1 2
2
What is the area under the curve y(x) = between x = 1 and x = ?
x2
[If you draw the graph, you’ll see that although x goes to infinity, the curve is getting closer and
closer to the y axis and so the area is getting smaller. So, in this case, it is possible to calculate a
finite area, even though we are integrating to infinity].
∞
A=∫ .
∞
=∫ 2 .
2
=
1 = x
[-2x-1]
1
= (0) - (-2)
= 2 sq.units.
Just as in differentiation, a function can by integrated term-by-term, and we only need one constant
of integration.
Our rules of integration are of no help here. We note that the integrand is of the form
where and
Since we are looking for an antiderivative of and we know that we can re-write our
integral as
In practice, we use the following substitution scheme to verify that we can integrate in this way:
1. Let
2. Differentiate both sides so
3. Change the original integral in to an integral in :
4. Integrate with respect to :
where and
Solution:
We note that the derivative of is ; hence, the current problem is not of the form
But we notice that the derivative is off only by a constant of and we
know that constants are easy to deal with when differentiating and integrating. Hence
Let
Then
Then and we are ready to change the original integral from to an integral in and
integrate:
We can also use this substitution method to evaluate definite integrals. If we attach limits of
integration to our first example, we could have a problem such as
The method still works. However, we have a choice to make once we are ready to use the
Fundamental Theorem to evaluate the integral.
where
We are not able to state a rule for integrating products of functions, but we can get
a relationship that is almost as effective. Recall how we differentiated a product of functions:
or
( )
= + + + ⋯+ + where
( ) ( ) ( ) ( ) ( )
1. Substitute the roots of the distinct linear factors of the denominator into the basic equation (the
equation obtained after eliminating the fractions on both sides of the equation) and find the
resulting constants.
2. For repeated linear factors, use the coefficients found in step 1 and substitute other
convenient values of x to find the other coefficients.
3. Integrate each term.
Integrating terms using Partial Fractions with Irreducible Quadratic Terms + + (quadratic
terms that cannot be factored) in the Denominator.
Summary
An integral equation is an equation in which an unknown function appears under one or
more integral signs.
An integral equation is called linear if only linear operations are performed in it upon the
unknown function. An integral equation which is not linear is known as a non-linear
integral equation.
A linear integral equation of the form
( ) ( )= ( )+ ( , ) ( )
( ) ( )= ( )+ ( , ) ( )
A solution of the integral equation is a function ( ) which when substituted into the
equation, reduces it into an identity.
Key words
Fredholm integral equation, Volterra integral equation
Self Assessment
1. Which of the following statement is correct?
A. A solution of the integral equation is a function ( ) which when substituted into the
equation, reduces it into an identity.
B. A solution of the integral equation is a function ( ) which when substituted into the
equation, reduces it into another equation of same type but reduced order.
C. A solution of the integral equation is a function ( ) which when substituted into the
equation, reduces it into a constant.
D. None of these
2. Which of the following is true for linear functions?
( )
A. = − + +⋯+ +
( ) ( ) ( ) ( ) ( )
( )
B. = + + + ⋯+ +
( ) ( ) ( ) ( ) ( )
( )
C. = + + +⋯+ +
( ) ( ) ( ) ( ) ( )
( )
D. = + + + ⋯+ +
( ) ( ) ( ) ( ) ( )
3. In ( ) ( ) = ( ) + ∫ ( , ) ( ) , K(x,t) is called
A. Factor
B. Kernel
C. Kernal
D. Spectrum
4. ( )+ ∫ ( , ) ( ) = 0 is
5. ( )+ ∫ ( , ) ( ) = 0 is
A. ( )+ ( ) = { ( )} . { ( )}
B. ( )+ ( ) = { ( )}/ { ( )}
C. ( )+ ( ) = { ( )} − { ( )}
D. ( )+ ( ) = { ( )} + { ( )}
7. Which of the following is a solution of the integral equation
1
( )= − ( )
1+ 1+
A. ( ) = (1 + )
B. ( )= 1+
C. ( ) = (1 + )
D. ( ) = (1 + )
( ) ( )= ( )+ ( , ) ( )
A. ( )
B. ( )a ( , )
C. ( )a ( , )
D. ( ), ( ) a ( , )
2
10. What is the area under the curve y(x) = between x = 1 and x = ?
x2
A. 1
B. 3
C. 2
D. 4
6. D 7. A 8. B 9. D 10. C
Review Questions
1. What is meant by Fredholm integral equation?
2. What is meant by Volterra integral equation?
3. What is meant by integral equation?
4. State the rules of integration of various elementary functions and derive if possible.
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
1. https://www.wolfram.com/language/11/symbolic-and-numeric-calculus/solve-a-
volterra-integral-equation.html
2. https://mathworld.wolfram.com/FredholmIntegralEquationoftheSecondKind.html
Objectives
Students will be able to
Introduction
In this chapter, we will learn about the singular and non singular integral equations, regularity
conditions, square integrable function and special kind of kernels. Moreover, we will learn about
the eigenvalues and eigenfunctions in the context of integral equations.
If f is integrable on [a, b] then the number V is denoted by ∫ and is called “the integral from a to
b of f.”
In practical terms, integrability depends on continuity. If a function is continuous on a given
interval, it’s integrable on that interval. Moreover, if a function has only a finite number of some
kinds of discontinuities on an interval, it’s also integrable on that interval.
Many functions — such as those with discontinuities, sharp turns, and vertical slopes — are non
differentiable. Discontinuous functions are also non differentiable. However, functions with sharp
turns and vertical slopes are integrable.
For example, the function y = |x| contains a sharp point at x = 0, so the function is non
differentiable at this point. However, the same function is integrable for all values of x.
So, the set of differentiable functions is actually a subset of the set of integrable functions.
The function y(x) will be a square integrablefunction, if it is measurable (i.e. finite) in the interval
[a, b] and ∫ | ( )| < ∞ , where the integral being taken in the sense of Lebesgue. The above
conditions are called as the 'regularity conditions’.
The function ( ) satisfying regularity conditions is called as 'square integrable function’ or simply
an -function.
≤ ≤ , ≤ ≤ such that
( , ) <∞
| ( , )| <∞
| ( , )| <∞
A function ( , ) satisfying the above three regularity conditions is called as square integrable
kernel’ or an ' – kernel’.
While dealing with -kernels, we shall often have occasion to use Fubini's theorem and its
extension by Tonelli and Hobson in order to justify changes in the order of integration in multiple
integrals.
G. Fubini (1879-1943) gave an important theorem related to changes in the order of integration in
multiple integrals, known as Fubini’s theorem which is stated as follows:
(i) The integral ∫ ( , ) exists for almost all x and is an integrable function of x such that
∫∫ ( , ) =∫ ∫ ( , )
(ii) The integral ∫ ( , ) exists for almost all t and is an integrable function of t such that
∫∫ ( , ) =∫ ∫ ( , )
Tonelli and Hobson extended the facts given by Fubini's theorem, known as Tonelli and Hobson’s
theorem which is stated as follows:
| ( , )| <∞
( , )= ( ) ( )+ ( ) ( )+ ⋯+ ( ) ( )= ( ) ( )
The functions ( ) can be regarded as linearly independent, otherwise the number of terms in
above can be further reduced. Recall that the set of functions ( ) is said to be 'linearly
independent’ if
( )+ ( ) +⋯+ ( ) = 0, where , ,…, are arbitrary constants gives = =
⋯= =0
For example, ( , ) = +
Difference Kernel:
Convolution integral:
The following non-singular linear integral equations
( )= ( )+ ∫ ( − ) ( ) and
( )= ( )+ ( − ) ( )
are called integral equations of the convolution or Faltung type (closed cycle type), where ( − )
is the difference kernel.
Let ( ) and ( ) be two continuous functions defined for ≥ 0. Then, the convolution or
Faltung of and is denoted and defined by
∗ = ( − ) ( ) = ( ) ( − )
Transposed Kernel:
A kernel ( , ) is said to be a transposed kernel if ( , )= ( , ) where ( , ) denotes the
transpose of ( , ).
( )= ( )+ ( , ) ( )
Then in this case, the iterated kernels ( , ); = 1,2,3, …are defined as follows:
( , )= ( , )
( , )= ( , ) ( , ) , = 2,3, …
( )= ( )+ ∫ ( , ) ( ) and
( )= ( )+ ( , ) ( )
( )= ( )+ ∫ ( , ; ) ( ) and
( ) = ( ) + ∫ Γ( , ; ) ( )
Then, ( , ; ) and
Γ( , ; ) are respectively called the resolvent kernel or reciprocal kernel of the given integral equations.
( )= ∫ ( , ) ( ) --- (1)
This equation has the obvious solution y(x) = 0, which is called as zero solution or trivial solution .
The values of the parameter for which (1) has a non-zero solution i.e. ( ) ≠ 0 , are called as the
eigenvalues of (1) or the eigenvalues of the kernel K (x, t) and each non-zero solution of (1) is called
an eigenfunction corresponding to the eigenvalue .
In other words, If ( ) is a continuous function such that ( ) ≠ 0, in the interval [a, b] and if
( )= ∫ ( , ) ( )
Then, the function ( ) is known as an eigenfunction of the integral equation (1) corresponding to
the eigenvalue .
Eigenvalues are also called as characteristic values or characteristic numbers or proper values or
fundamental values. Consequently, eigenfunctions are termed as characteristic functions or proper
functions or fundamental functions.
The following remarks can be made regarding the eigenvalues and eigenfunctions:
Remarks:
1. The number = 0 is not an eigenvalue of (1), since for = 0 it follows from (1) that ( ) = 0 ,
which is a trivial solution.
If G and H are absolute constants, then Leibnitz's rule takes the following form :
( , ) =
x
Summary
When one or both limits of integration in an integral equation becomes infinite or when
the kernel of an integral equation becomes infinite at one or more points within the range
of integration, then the integral equation is called as singular integral equation.
An integral equation which is not singular is known as non- singular integral equation.
Let f be a bounded function from an interval [a, b] to R. We say that f is integrable on [a, b]
if there is a number V such that for every sequence of partitions { } on [a, b] such that
{ ( )} → 0 and every sequence { } where is a sample for
{∑( , , )} →
If f is integrable on [a, b] then the number V is denoted by ∫ and is called “the integral
from a to b of f.”
I. The integral ∫ ( , ) exists for almost all x and is an integrable function of x such that
∫∫ ( , ) =∫ ∫ ( , )
II. The integral ∫ ( , ) exists for almost all t and is an integrable function of t such that
∫∫ ( , ) =∫ ∫ ( , )
Consider the following homogeneous Fredholm integral equation:
( )= ∫ ( , ) ( ) --- (1)
This equation has the obvious solution y(x) = 0, which is called as zero solution or trivial solution .
The values of the parameter for which (1) has a non-zero solution i.e. ( ) ≠ 0 , are called as the
eigenvalues of (1) or the eigenvalues of the kernel K (x, t) and each non-zero solution of (1) is called
an eigenfunction corresponding to the eigenvalue .
Keywords
Singular integral equation, L2 kernel, integrable function, Fubini’s theorem, reciprocal kernel,
eigenvalues of an integral equation
Self Assessment
1. Which of the following statement is correct about the following equation:
( )= ( )+ ( − ) ( )
A. It is a linear integral equation.
B. It is a non singular integral equation
C. It is a convolution type integral
D. All of these
2. ( , )= ( ) ( )+ ( ) ( ) + ⋯+ ( ) ( )=∑ ( ) ( ) is called
A. PG- kernel
B. Separable kernel
C. Degenerate kernel
D. All the above
4. ( , )=e is a
A. Difference kernel
B. Degenerate kernel
C. Both (a) and (b)
D. None of these
6. B 7. C
Review Questions
1. What is meant by a non singular integral equation?
2. Define kernel and elaborate.
3. Mention the various types of kernels of linear integral equations.
4. What is meant by eigenvalue in general and in the context of integral equations?
5. State Fubini’s theorem and contemplate on its role in solving the integral equations.
Further Readings
1. M. Gelfand, S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
https://web.ma.utexas.edu/users/m408m/Display15-2-3.shtml
Pratiksha, Lovely Professional University Unit 03: Differential Equation to Integral Equation
Objectives
Students will be able to
Introduction
While searching for the representation formula for the solution of an ordinary linear differential
equation in such a manner so as to include the boundary or initial conditions explicitly, we always
arrive at an integral equation.Thus, an ordinary differential equation may be converted to an
integral equation. Similarly, a partial differential equation may also be converted to an integral
equation.
Here, we shall describe in what ways the integral equations can arise from ordinary differential
equations. For the purpose, we shall start from initial and boundary value problems.
After converting an initial value problem or a boundary value problem to an integral equation, it
can be solved by shorter methods of solving integral equations.
For example,
+ = ; (1) = 2, (1) = 2
+ = 0; (0) = 0, ′(0) = 0.
( )− (0) = − ( )
Or
( )= − ( )
( ) − (0) = − ( )
( )=− ( − ) ( )
Also derive the original differential equation with initial conditions from the obtained integral
equation.
Solution. The given differential equation can be written as
( )=3 ( )−2 ( )+4
Integrating both sides w.r.t. x from 0 to x, we get
( ) = 3 ( )−2 ( ) − 4 cos − 1
( ) − (0) = 3 ( ) −2 ( − ) ( ) − 4 sin −
Or ( ) = 1 − − 4 sin + ∫ 3 − 2( − ) ( )
( ) = −1 − 4 cos + −2 ( ) +3 ( )
( ) = 4 sin − 2 ( ) + 3 ′( )
( )=4 − 2 ( ) + 3 ′( )
Or
( )− 3 ( )+2 ( )=4
that (0) = 1 and from ( ) = −1 − 4 cos + ∫ −2 ( ) + 3 ( ) , it can be seen that (0) = −2.
Thus we got the initial conditions also and obtained a complete initial value problem.
+ = 0; ( ) = , ( )=
is a boundary value problem. Here, it must be noted that the two different values x = a and x = b of
the independent variable x are involved.
+ = 0; (0) = 0, ( ) = 0.
( )− (0) = − ( )
Let (0) =
( )= − ( )
( ) − (0) = − ( )
( )= − ( − ) ( )
We know that ( ) = 0
()= − ( − ) ( )
= ( − ) ( )
( )= . ( − ) ( ) − ( − ) ( )
= ( − ) ( ) + ( − ) ( ) − ( − ) ( )
( − ) ( − )
= ( ) + ( )
= ( , ) ( )
( ) ( )
where ( , ) = ,0 < < and ( , )= , < <
+ = 1; (0) = 0, (1) = 1.
Also recover the boundary value problem from the obtained integral equation.
Solution. We can write the given problem as
( )=1− ( ); (0) = 0, (1) = 1
Integrate both sides w.r.t. x
( )− (0) = − ( )
( )= + − ( )
where (0) =
On integration in the usual manner, we can write
( )= + − ( − ) ( )
2
( )
(Using∫ ( ) =∫ ( ) ))
( )!
Put =1
1
(1) = 1 = + − (1 − ) ( )
2
1
= + (1 − ) ( )
2
So ( ) = + ∫ (1 − ) ( ) + −∫ ( − ) ( )
+
( )= + (1 − ) ( ) + (1 − ) ( )
2
Summary
An initial value problem when converted to integral equation, takes the Volterra form.
A boundary value problem when converted to integral equation, takes the Fredholm form.
Keywords
IVP, BVP, Volterra integral equation, Fredholm integral equation
Self Assessment
1. The equation ( )+ ( ) = 1, (0) = 0, (1) = 1 is
A. an IVP
B. a BVP
C. Both (a) and (b)
D. None of these
2. A boundary value problem when converted to integral equation, takes the ________ form.
A. Volterra
B. Cauchy
C. Fredholm
D. Gaussian
( )
B. ∫ ( ) =∫ ( )!
( )
( )
C. ∫ ( ) =∫ ( )
( )
( )
D. ∫ ( ) =∫ ( )!
( )
Review Questions
1. What is meant by IVP?
2. Define a BVP.
3. Differentiate in IVP and BVP.
4. Can we say that IVP and BVP are equivalent to the integral equations? Explain in
detail.
5. Write an initial value problem and convert it into the corresponding integral equation.
6. Write any boundary value problem and convert it into corresponding integral
equation.
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
1. https://www.youtube.com/watch?v=Z_i5ibKjSAg
2. https://www.youtube.com/watch?v=0VWaBJz3t1s
3. https://www.youtube.com/watch?v=WPlBrzjI1KI
4. https://projecteuclid.org/journals/differential-and-integral-equations
Unit 04: Fredholm Integral Equations of the Second Kind with Separable Kernels
Pratiksha, Lovely Professional University
Objectives
Students will be able to
Introduction
In this chapter, we will learn to solve the Fredholm integral equations of second kind in both
homogeneous and non-homogeneous forms. Prior to that, we need to learn about the orthogonality
of two functions.
( )= ∫ ( , ) ( ) --- (1)
This equation has the obvious solution y(x) = 0, which is called as zero solution ortrivial solution .
The values of the parameter for which (1) has a non-zero solution i.e.( ) ≠ 0 , are called as the
eigenvalues or characteristic values of (1) or the eigenvalues or characteristic values of the kernel K
(x, t).
Then, the function ( ) is known as ancharacteristic function of the integral equation (1)
corresponding to the eigenvalue .
The number = 0 is not an eigenvalue of (1), since for = 0 it follows from (1) that ( ) = 0 ,
which is a trivial solution.
( )= ( , ) ( ) − − − (1)
and ( )= ∫ ( , ) ( ) − − − (3)
Again, since we are interested in symmetric kernel, therefore, the kernel can be written as follows :
( , )= ( , ) − − − (4)
( ) ( )= ( )∫ , ( )
∫ ( ) ( ) = ∫ ( )∫ , ( )
∫ ( ) ( ) = ∫ ( )∫ , ( )
∫ ( ) ( ) = ∫ ( )∫ ( , ) ( ) − −(5)
( )= ( , ) ( )
( )= ( , ) ( )
Unit 04: Fredholm Integral Equations of the Second Kind with Separable Kernels
( )= ∫ ( , ) ( ) − − − (6)
( )
( ) ( ) = ( ) = ( ) ( )
Or
( ) ( ) = ( ) ( )
Or
∫ ( ) ( ) = ∫ ( ) ( )
Or
( − ) ( ) ( ) =0
Hence, the Eigenfunctions ( ) and ( ) corresponding to two distinct eigenvalues and are
orthogonal to each other over the interval [ , ].
Theorem 02: The eigenvalues of a symmetric kernel of a homogeneous Fredholm integral equation
are real.
Consider the following homogeneous Fredholm integral equation of the second kind:
( )= ( , ) ( ) − − − (1)
Now, let us suppose that equation (1) has an eigenvalue , which is not real, therefore, we may
assume
= + − − − (2)
Also let
( )= + − − − (3)
be the complex eigenfunction corresponding to the eigenvalue . Then, the complex conjugate
of would necessarily be an eigenvalue corresponding to the eigenfunction ( ) which is
complex conjugate of ( ). Thus, we have
= − and ( )= −
Now, since ( ) and ( ) are the eigenfunctions corresponding to the eigenvalues and
respectively, they must satisfy equation (1). Thus, we have
( )= ( , ) ( ) − − − (4)
And
( )= ( , ) ( ) − − − (5)
Multiplying both sides of (4) by ( ) and then integrating w.r.t. x over the interval [a, b], we get
( ) ( ) = ( ) ( , ) ( )
( ) ( ) = ( ) ( , ) ( )
Or
( ) ( ) = ( ) ( , ) ( ) − − − (6)
Or ( )= ∫ ( , ) ( )
( )
Therefore, =∫ ( , ) ( )
( )
( ) ( ) = ( )
Or
( ) ( ) = ( ) ( )
Or
− ( ) ( ) =0
Or
2 ∫ ( + ) =0
Unit 04: Fredholm Integral Equations of the Second Kind with Separable Kernels
So, ∫ ( + ) ≠0
Thus, =0
Hence, the eigenvalues of a symmetric kernel of a homogeneous Fredholm integral equation are
real.
( )= ( , ) ( )
( , )= ( ) ( )
Putting ( , ) in ( ), we get
( )= ( ) ( ) ( )
( )= ( ) ( ) ( )
( ) ( ) = ; = 1,2, … ,
( )= ( ) ( ) ( )
becomes ( ) = ∑ ( ); = 1,2, … ,
where, constants are to be determined to find the complete solution of the given integral equation.
Multiply both sides of
( )= ∑ ( )∫ ( ) ( ) successively by ( ), ( )… ( ) and integrating over [ , ],
and define
= ( ) ( ) ; , = 1,2, … ,
We can write
= ; = 1,2, … ,
or(1 − ) − − ⋯− =0
Similarly, we get
(1 − ) − − ⋯− =0
(− ) + (1 − ) − ⋯− =0
………………………………………………………….
(− ) + (− ) − ⋯ + (1 − ) =0
Thus, we get a homogeneous linear equations which may be used to find the values of , and its
determinant can be written as:
1− − … −
− 1− … −
( )=
… … … …
− − … 1−
If this determinant is non zero, then the system of equations from which it is made up of, has only trivial
solution given by = = ⋯ = 0. Thus the integral equation also has a trivial solution i.e. a zero solution.
If this determinant is zero, then at least one of the can be assigned in an arbitrary manner and the other
′ can be obtained accordingly. Therefore infinitely many solutions exist for the integral equation in the
second case.
The eigenvalues of the given integral equation can be obtained from
( )=0
or
1− − … −
− 1− … −
=0
… … … …
− − … 1−
The degree of this equation in is less than or equal to n. So, if the integral equation has separable kernel,
then the integral equation has at most ‘n’ eigenvalues and corresponding eigenfunctions will be its
solutions.
( )= ( )+ ( , ) ( )
( , )= ( ) ( )
Putting ( , ) in ( ), we get
( )= ( )+ ( ) ( ) ( )
( )= ( )+ ( ) ( ) ( )
( ) ( ) = ; = 1,2, … ,
( )= ( )+ ( ) ( ) ( )
becomes ( ) = ( ) + ∑ ( ); = 1,2, … ,
where, constants are to be determined to find the complete solution of the given integral equation.
Unit 04: Fredholm Integral Equations of the Second Kind with Separable Kernels
= ( ) ( ) ; , = 1,2, … ,
And =∫ ( ) ( ) ; = 1,2, … ,
We can write
= + ; = 1,2, … ,
= + + +⋯+
or (1 − ) − − ⋯− =
Similarly, we can simplify the equations further to
get
(1 − ) − − ⋯− =
(− ) + (1 − ) − ⋯− =
………………………………………………………….
(− ) + (− ) − ⋯ + (1 − ) =
Thus, we get a homogeneous linear equations which may be used to find the values of , and its
determinant can be written as:
1− − … −
− 1− … −
( )=
… … … …
− − … 1−
If this determinant is non zero, then the system of equations from which it is made up of, has only trivial
solution given by = = ⋯ = 0. Thus the integral equation also has a trivial solution i.e. a zero solution.
If this determinant is zero, then at least one of the can be assigned in an arbitrary manner and the other
′ can be obtained accordingly. Therefore infinitely many solutions exist for the integral equation in the
second case.
The eigenvalues of the given integral equation can be obtained from
( )=0
or
1− − … −
− 1− … −
=0
… … … …
− − … 1−
The degree of this equation in is less than or equal to n.
Now three cases arise:
Case 1: When at least one is non zero
If ( ) is non zero, then a unique non zero solution of the system of equations exists and hence the integral
equation has a unique non zero solution.
If ( ) is zero, then either infinitely many solutions or no solution of the system of equations exists and
hence the integral equation has infinitely many solutions or no solution.
Summary
The equation ( ) = ∫ ( , ) ( ) has the obvious solution y(x) = 0, which is called as
zero solution or trivial solution.The values of the parameter for which the equation has a
non-zero solution i.e.( ) ≠ 0 , are called as the eigenvalues or characteristic valuesof the
integral equation or the eigenvalues or characteristic values of the kernel K (x, t).
The number = 0 is not an eigenvalue of (1), since for = 0 it follows from (1) that
( ) = 0 , which is a trivial solution.
If ( ) is an eigenfunction of (1), then ( ), where c is an arbitrary constant, is also an
eigenfunction of (1), which corresponds to the same eigenvalue .
If the kernel ( , ) is continuous in the rectangle : ≤ ≤ , ≤ ≤ and the
numbers and are finite, then to every eigenvalue , there exists a finite number of
linearly independent eigenfunctions. The number of such functions is known as the index
of the eigenvalue. Different eigenvalues have different indices.
Two functions ( ) and ( ), both continuous in the interval [a, b] are said to be
Keywords
Orthogonal functions, characteristic functions,Fredholm integral equation of second kind,
homogeneous integral equation, non-homogeneous integral equation
Self Assessment
1. The equation ( ) = ∫ ( , ) ( ) has the obvious solution y(x) = 0, which is called
A. Zero solution
B. Trivial solution
Unit 04: Fredholm Integral Equations of the Second Kind with Separable Kernels
2. The values of the parameter for which the equation has a zero solution i.e. ( ) = 0 , are
called as the eigenvalues or characteristic values
A. True
B. False
A. True
B. False
4. Two functions ( ) and ( ), both continuous in the interval [a, b] are said to be
orthogonal to each other over the interval [a, b] if
A. ∫ ( ) ( ) ≠0
B. ∫ ( ) ( ) =0
C. ∫ ( ) ( ) =0
D. ∫ ( ) ( ) =0
6. B
Review Questions
1. Find the characteristic values and characteristic functions of the homogeneous
integral equation:
( )= ( )
2. Show that ( ) = ∫ (3 − 2) ( ) has no characteristic values and characteristic
functions.
3. Solve ( ) = ∫ sin( + ) ( )
4. Solve ( ) = ( ) + ∫ ( )
5. Solve ( ) = + ∫ ( + ) ( )
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
1. https://www.youtube.com/watch?v=rCWzF1yvZlQ
2. https://www.youtube.com/watch?v=K0t53t7RLWY
Objectives
Students will be able to
assure that the solution of integral equation exists
define the iterated and resolvent kernel
solve the non-homogeneous integral equations
Introduction
Separable kernel in an integral equation implies integral equation of second kind becomes the
algebraic system of equations
For more general kernels, what can be the consequence? To answer this we must know that any
well behaved kernel can be expressed as infinite series of separable kernels. When an integral
equation cannot be solved in closed form,we use approximation methods.
For any approximation method, existence of solution should be known in advance and Fredholm
theorem provides the assurance that the solution of integral equation exists.
( )= ( )+ ( , ) ( )
( , )= ( ) ( )
Putting ( , ) in ( ), we get
( )= ( )+ ( ) ( ) ( )
( )= ( )+ ( ) ( ) ( )
( ) ( ) = ; = 1,2, … ,
( )= ( )+ ( ) ( ) ( )
becomes ( ) = ( ) + ∑ ( ); = 1,2, … ,
where, constants are to be determined to find the complete solution of the given integral equation.
Multiply both sides of
( )= ∑ ( )∫ ( ) ( ) successively by ( ), ( )… ( ) and integrating over [ , ],
and define
= ( ) ( ) ; , = 1,2, … ,
And =∫ ( ) ( ) ; = 1,2, … ,
We can write
= + ; = 1,2, … ,
= + + +⋯+
or (1 − ) − − ⋯− =
Similarly, we can simplify the equations further to
get
(1 − ) − − ⋯− =
(− ) + (1 − ) − ⋯− =
………………………………………………………….
(− ) + (− ) − ⋯ + (1 − ) =
Thus, we get a homogeneous linear equations which may be used to find the values of , and its
determinant can be written as:
1− − … −
− 1− … −
( )=
… … … …
− − … 1−
The required solution of (1) depends on the determinant given by (9).
Two cases arise:
Case I. If ( ) ≠ 0, then the system of equations (8) has only one solution given by Cramer's rule,
⋯
= ( )
, = 1, 2, … , …(10)
Substituting the value from (10) in (5), the unique solution of (1) is given by
⋯
( )= ( )+ ∑ ( )…(11)
( )
Now by developing this determinant by the elements of the first row and the corresponding minor
of the first column, we have
{ ( )+ ( )+ ⋯+ ( )} ( ) = ( , : )
( , : )
( )= ( )+ ( )
( )
Or
( )= ( )+ ∫ ( , : ) ( ) …(14)
where
( , : )
( , : )= is known as the resolvent or reciprocal kernel of the given integral equation.
( )
The only possible singular points of ( , : ) in the plane are the roots of the equation ( ) = 0.
( )= ( )+ ∫ ( , : ) ( ) .
, :
The resolvent kernel ( , : ) is given by the quotient of two polynomials ( )
.
Case II. If ( ) = 0, then (8) has no solution in general, because an algebraic system with ( ) = 0
can be solved only for some particular values of , , … . For this situation, we re write the
system of equations (8) in matrix form as follows:
…
( − ) = , where = … … , = … ,
…
= …
And is the identity matrix of order . Since ( ) = 0, for each non trivial solution of the
homogeneous system ( − ) = 0, there corresponds a non-trivial solution (eigenfunction) of the
homogeneous integral equation (12).
Then to each eigenvalue of index , there corresponds a solution ( ) of (3) of the form
( )= ( )
= geometric multiplicity
= algebraic multiplicity
For the situation when for non-homogeneous Fredholm integral equation ( ) = 0 we define the
transpose of equation (1) as
( )= ( )+ ∫ ( , ) ( ) …(15)
Equation (1) and (15) are transpose of each other and the relation between them is symmetric.
We can write its kernel as
( , )=∑ ( ) ( ) …(16)
And proceeding on this , we get the algebraic system
( − ) = …(17)
where =∫ ( ) ( ) and =∫ ( ) ( )
The determinant ( ) is the same function, except that there is an interchange of the rows and
columns. Hence, the eigenvalues of (15) are same as that of the original integral equation (1). So (15)
possesses a unique solution, whenever (1) does.
The number of LI eigenfunctions will be for the transposed homogeneous integral equation . Let
these LI solutions be denoted by , ,…, and let us assume that they have been
normalized. Then any solution ( ) of the transposed integral equation
( )= ∫ ( , ) ( ) …(18)
We now prove that the eigenfunctions ( ) and ( ) corresponding to the distinct eigenvalues
and , are orthogonal.
By definition,
( )= ∫ ( , ) ( ) …(19)
and
( )= ∫ ( , ) ( ) …(20)
Multiplying both sides of (19) by ( ) and then integrating w.r.t. over the interval ( , ), we get
( ) ( ) = ( ) ( , ) ( )
Or
∫ ( ) ( ) = ∫ ( )∫ ( , ) ( ) …(21)
Again multiplying both sides of (20) by ( ) and then integrating w.r.t. over the interval
( , ), we get
∫ ( ) ( ) = ∫ ( )∫ ( , ) ( ) …(22)
( ) ( ) = ( ) ( )
The solution of (1) for the case ( ) = 0: The necessary and sufficient condition for (1) to possess a
solution for = , a root of ( ) = , is that ( ) be orthogonal to the eigenfunctions of the
transposed equation (15).
( )= ( )+ ( , ) ( )
Or
( )= ( )− ( , ) ( )
Multiplying both sides of above equation by ( ) and then integrating w.r.t. x over the interval (a,
b)
( ) ( ) = ( ) ( ) − ( ) ( , ) ( ) =0
Finally we know that the difference of any two solutions of (1) is a solution of the homogeneous equation
(12) . Hence the most general solution of (1) has the form
( )= ( )+ ( )+ ( ) + ⋯+ ( )
Thus, we have proved that if = is a root of multiplicity ≥ 1 of the equation ( ) = 0, then the
inhomogeneous equation (1) has a solution if and only if ( ) is orthogonal to all the eigenfunctions of the
transposed equation.
( )= ( )+ ( , ) ( )
and ( , )=∫ ( , ) ( , ) , = 2, 3, … ,
or ( , )=∫ ( , ) ( , ) , = 2, 3, … ,
The iterated kernels are also called as the approximate kernels or approximate functions.
( )= ( )+ ( , ) ( )
( )= ( )+ ∫ ( , ; ) ( )
where
( , ; )= ( , )
where ( , ) is the iterated kernel and ( , ; ) is the resolvent kernel of the non-
homogeneous Fredholm integral equation.
( )= ( )+ ( , ) ( )
( , )= ( , )
and ( , )=∫ ( , ) ( , ) , = 2, 3, … ,
or ( , )=∫ ( , ) ( , ) , = 2, 3, … ,
The iterated kernels are also called as the approximate kernels or approximate functions.
( )= ( )+ ( , ) ( )
( )= ( )+ ( , ; ) ( )
where
( , ; )= ( , )
where ( , ) is the iterated kernel and ( , ; ) is the resolvent kernel of the non-
homogeneous Volterra integral equation.
Summary
Either the integral equation
( )= ( )+ ∫ ( , ) ( ) …(i)
with fixed possesses one and only one solution ( ) for arbitrary functions ( ) and ( , ), in
particular, the solution = 0 for = 0; or the homogeneous equation
( )= ∫ ( , ) ( ) …(ii)
possesses a finite number of linearly independent solutions , = 1, 2, … , .
In the first case, the transposed inhomogeneous equation
( )= ( )+ ∫ ( , ) ( ) …(iii)
also possesses a unique solution.
In the second case, the transposed homogeneous equation
( )= ∫ ( , ) ( ) …(iv)
also has linearly independent solutions , = 1, 2, … , ;
The inhomogeneous integral equation (i) has a solution if and only if the given function ( )
satisfies the conditions
∫ ( ) ( ) = 0, = 1, 2, … , …(v)
In this case the solution of (i) is determined only up to an additive linear combination ∑
and ( , )=∫ ( , ) ( , ) , = 2, 3, … ,
( )= ( )+ ( , ) ( )
( )= ( )+ ( , ; ) ( )
where
( , ; )= ( , )
where ( , ) is the iterated kernel and ( , ; ) is the resolvent kernel of the non-
homogeneous Volterra integral equation.
Keywords
Fredholm Alternative Theorem,
Iterated kernel,
Resolvent kernel,
Non-homogeneous integral equations
Self Assessment
1. Which of the following statement is correct?
A. ( , ; )=∑ ( , )
B. ( , ; )=∑ ( , )
C. ( , ; )=∑ ( , )
D. ( , ; )=∑ ( , )
A. diagonal
B. symmetric
C. Hermitian
D. Orthogonal
Review Questions
1. Find the iterated kernel for the following:
(i) ( , ) = sin( − 2 ) , 0 ≤ ≤2 , 0≤ ≤2
( , ) = sin( − 2 ) , 0 ≤ ≤2 , 0≤ ≤2
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
1. https://mathworld.wolfram.com/FredholmAlternative.html
2. https://terrytao.wordpress.com/2011/04/10/a-proof-of-the-fredholm-alternative/
Objectives
Students will be able to
Introduction
In this chapter we will mainly learn to solve the Fredholm and Volterra integral equation of second
kind by the method of successive substitution and successive approximation. We establish the
methods using the theorems and then the problems can be solved using those methods.
( )= ( )+ ∫ ( , ) ( ) ---------- (1)
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) + ⋯ --------------(5)
( )= ( )+ ∫ ( , ) ( ) ---------- (6)
( ) = ( )+ ∫ ( , ) ( ) -----------(7)
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) -----------(8)
( ) = ( )+ ∫ ( , ) ( ) --------------(9)
( )= ( )+ ∫ ( . ) ( ) ---------------(10)
( )= ( )+ ( , ) ( ) + ( , ) ( , )[ ( ) + ( . ) ( ) ]
Or
( )=
( )+
∫ ( , ) ( ) +
∫ ( , )∫ ( , ) ( ) + ∫ ( , )∫ ( , )∫ ( , ) ( ) -------------- (11)
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) + ⋯ -----------(14)
Because of (i) and (ii), each term of the series (14) is continuous in I. Therefore the series (14) is also
continuous in I, if it converges uniformly in I.
Let ( ) be the general term of the series (14), then we can write
( )= ∫ ( , )∫ ( , )… ∫ ( , ) ( ) … … -------------(15)
| ( )| = | ∫ ( , )∫ ( , )…∫ ( , ) ( ) … |------------(16)
And
| ( )| ≤ | | ( − )
Therefore the series (14) converges absolutely and uniformly only when
1
| |<
( − )
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) + ⋯ --------------(5)
( ) = ( )+ ∫ ( , ) ( ) -----------(7)
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) -----------(8)
( ) = ( )+ ∫ ( , ) ( ) --------------(9)
( )= ( )+ ∫ ( . ) ( ) ---------------(10)
( )= ( )+ ( , ) ( ) + ( , ) ( , )[ ( ) + ( . ) ( ) ]
Or
( )=
( )+
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) + ⋯ -----------(14)
Let ( ) be the general term of the series (14), then we can write
( )= ∫ ( , )∫ ( , )… ∫ ( , ) ( ) … … -------------(15)
| ( )| = | ∫ ( , )∫ ( , )…∫ ( , ) ( ) … |------------(16)
And
| | ( )
| ( )| ≤ ; ≤ ≤ -------------(17)
!
( )= ( )+ ∫ ( , ) ( ) ----------(1)
Consider ( ) = ( )--------(2)
as a zero order approximation for the solution of (1). Further, if ( ) and ( ) are respectively
the nth order and (n-1)th order approximations of y(x), then
( )= ( )+ ∫ ( , ) ( ) ---------(3)
( )= ( )+ ∫ ( , ) ( ) --------- (5)
( )= ( )+ ∫ ( , ) ( ) -------------(7)
( )= ( )+ ∫ ( , ) ( ) or
( )= ( )+ ∫ ( , ) ( ) -----------(8)
Replacing x by s in (7)
( ) = ( )+ ∫ ( , ) ( ) -----------(9)
( )= ( )+ ∫ ( , ) ( ) + ∫ ( , )[∫ ( , ) ( ) ] ----------(10)
Or
( ) = ( ) + ∑{ } ∫ ( , ) ( ) -------------(11)
( ) = ( ) + ∑{ } ∫ ( , ) ( ) -------------(12)
( ) = lim → ( ) = ( ) + ∑{ } ∫ ( , ) ( ) -------------(13)
BY changing the order of integration in third term on RHS of (10), we can write
( )= ( )+ ( , ) ( ) + ( , ) ( )
We now determine the resolvent kernelor reciprocal kernel in terms of iterated kernels ( , ). For
this purpose, by changing the order of integration and summation in the Neumann series, we get
( ) = ( ) + ∫ [∑ ( , )] ( ) ------------(14)
We get ( , ; ) = ∑ ( , )]--------------(16)
Consider ( ) = ( )--------(2)
as a zero order approximation for the solution of (1). Further, if ( ) and ( ) are respectively
the nth order and (n-1)th order approximations of y(x), then
( )= ( )+ ∫ ( , ) ( ) ---------(3)
( )= ( )+ ∫ ( , ) ( ) -----------(8)
Replacing x by z in (7)
( )= ( )+ ∫ ( , ) ( ) -----------(9)
( )= ( )+ ( , ) ( ) + ( )[ ( , ) ( , ) ]
( ) = ( ) + ∑{ } ∫ ( , ) ( ) -------------(12)
BY changing the order of integration in third term on RHS of (10), we can write
( )= ( )+ ( , ) ( ) + ( , ) ( )
We now determine the resolvent kernel or reciprocal kernel in terms of iterated kernels ( , ).
For this purpose, by changing the order of integration and summation in the Neumann series, we
get
( ) = ( ) + ∫ [∑ ( , )] ( ) ------------(14)
We get ( , ; ) = ∑ ( , )]--------------(16)
The series converges absolutely and uniformly when K(x,t) is continuous inR.
Summary
Consider the following non-homogeneous Fredholm integral equation:
( )= ( )+ ∫ ( , ) ( ) ---------- (1)
Then, the integral equation (1) has a unique continuous solution in I, which is given by the
following uniformly and absolutely convergent series:
( )= ( )+ ( , ) ( ) + ( , ) ( , ) ( ) +⋯
( )= ( )+ ∫ ( , ) ( ) ---------- (2)
Then, the integral equation (2) has a unique continuous solution in I, which is given by the
following uniformly and absolutely convergent series:
( )= ( )+ ( , ) ( ) + ( , ) ( , ) ( ) +⋯
( )= ( )+ ∫ ( , ) ( ) ----------(3)
Let ( ) = ( )
as a zero order approximation for the solution of (3). Further, if ( ) and ( ) are respectively
the nth order and (n-1)th order approximations of y(x), then
( )= ( )+ ( , ) ( )
( )= ( )+ ∫ ( , ) ( ) ----------(4)
Let ( ) = ( )
as a zero order approximation for the solution of (4). Further, if ( ) and ( ) are respectively
the nth order and (n-1)th order approximations of y(x), then
( )= ( )+ ( , ) ( )
Key words
Fredholm integral equation,
Volterra Integral equation,
Integral equation of second kind,
successive substitution, successive approximation
Self Assessment
1. Which of the following statement is correct in context of integral equations?
B. ( )= ( )+ ∫ ( , ) ( )
C. ( )= ( )+ ∫ ( , ) ( )
D. ( )= ( )+ ∫ ( , ) ( )
4. The series ( ) = ( ) + ∫ ( , ) ( ) + ∫ ( , )∫ ( , ) ( ) +⋯
A. Converges absolutely
B. Converges uniformly
C. Converges absolutely and uniformly
D. Does not converge
| | ( )
5. term in context of the integral equation ( )= ( )+ ∫ ( , ) ( )
!
A. Constant
B. Positive constant
C. Negative constant
D. Arbitrary variable
Review Questions
1. What is the method of successive substitution to solve integral equations?
2. State the Fredholm integral equation of second kind.
3. What is the method of successive approximation to solve the integral equations?
4. State the Volterra integral equation of second kind.
6. Solve ( ) = sin + 2 ∫ ( ) )
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc.
(1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley
(1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press
(1947)
Web Links
1. https://www.youtube.com/watch?v=Qo9riz9NZ64
2. https://www.youtube.com/watch?v=u4yhu8QMC2M
Pratiksha, Lovely Professional University Unit 07: Iterated and Resolvent Kernels
Objectives
Students will be able to
Introduction
In this chapter we will learn about some theorems on iterated kernels and resolvent kernels. We
will solve some integral equations also. We will also learn about the solution when the kernel takes
two specific polynomial forms.
( , ; )= ( , )+ ( , ) ( , ; )
2. Let ( , ; )be the resolvent kernel of the following non homogeneous Volterra integral equation:
( )= ( )+ ( , ) ( )
( , ; )= ( , )+ ( , ) ( , ; )
( , )= ( , ) ( , )
Therefore,
( , )= . =
3
( , )= . =
3 3
.
.
.
( , )=
3
Therefore, ( , ; ) = ∑ ( , )
=
3
=
6
6
=
5
The solution of given equation is
5 1 6 5
( )= +
6 2 5 6
6
= =
6
Example:Solve ( ) = 1 + ∫ ( )
Here ( ) = 1, = 1, ( , ) = 1 = ( , )
( , )= ( , ) ( , )
= 1 = −
( , )= ( , ) ( , )
( − )
= ( − ) = |
2
( , )= ( , ) ( , )
( − )
=
2
( − )
=
3!
⋮
( − )
( , )=
( − 1)!
( , ; )= ( , )
( − )
= =
( − 1)!
( )=1+
=1+
= 1 + (− )( − 1)
=1−1+
=
Example:Solve ( ) = 1 + ∫ ( + ) ( )
Here
( , )= + = ( , )
( , )= ( + )( + )
+ 1
= + +
2 3
1
( , )= ( + ) + + +
2 3 2
7 1
= ( + )+ +
12 3
7 7 1
( , )= ( + ) + + +
12 12 3
15 13 13
= ( + )+ +
24 12 36
And proceeding like this the nth iterated kernel can be calculated and hence the resolvent kernel
too can be calculated.
When the resolvent kernel cannot be obtained in the closed form, we use the method
of successive approximations to find solution up to third order.
Let the given Fredholm integral equation of second kind be
( )= ( )+ ∫ ( , ) ( ) ---------- (1)
( )= ( )+ ∫ ( , ) ( ) ---------------(3)
( )= + ( − )
= −
6
( )= + ( − ) −
6
= − +
3! 5!
7.2 Volterra integral equation of the second kind
Consider the equation,
( )= ( )+ ∫ ( , ) ( ) -------(1)
Where ( , ) is
( , ; )
For Case I. ( , ; ) =
( ) ( )
− + +⋯+ ( ) =0
( )= ( )+ ( , ; ) ( )
( , ; )
For Case II. ( , ; ) = −
ℎ ( ) ℎ ( )
− + + ⋯+ ( )ℎ = 0
( )= ( )+ ( , ; ) ( )
Example:Solve ( ) = 29 + 6 + ∫ {5 − 6( − )} ( )
Comparewith ( ) = ( ) + ∫ ( , ) ( )
( ) = 29 + 6
=1
( , ) = 5 − 6( − ) = ( )+ ( )( − )
( , ; )
So = 5, = −6 Then ( , ; ) =
Or ( , ; 1) =
Or − +6 =0
and = 1 when = .
( ) ( )
=3 −2
( ) ( )
=9 −4
( ) ( )
So ( , ; 1) = 9 −4
Therefore the solution of the integral equation is
( )= ( )+ ( , ; ) ( )
( ) = 29 + 6 , = 1, ( , ; 1) = 9 ( ) ( )
−4
( ) ( )
∴ ( ) = 29 + 6 + [9 −4 ] [29 + 6 ]
= 93 − 64
( )= ( )+ ( , ) ( )
Example:Solve ( ) = ∫ ( ) , (0) = 0
∫ ( ) = ( ) ------(1)
( ) + ( ) = ′( )
( )= ( )+ ∫ ( , ) ( ) ----- (4)
where ( ) = ( )
= −1
( , )=
( , )= ( , )=
( , )= ( , ) ( , )
( , )= = |
=( − )
( − )
( , )=
2!
( − )
( , )=
3!
( − )
( , )=
( − 1)!
( , ; )= ( , )
= (−1) ( , )
= ( , )− ( , )+ ( , )−⋯
− ( − )
= {1− + − ⋯}
1! 2!
( )
= . =1
( )= ( )+ ( , ; ) ( )
= ( )− ( )
= ( ) − ( )|
= ( ) − ( ) + (0)
= ( )− ( )
Summary
Let ( , ; )be the resolvent kernel of the following non-homogeneous Fredholm
integral equation:
( )= ( )+ ( , ) ( )
( , ; )= ( , )+ ( , ) ( , ; )
Let ( , ; )be the resolvent kernel of the following non homogeneous Volterra integral equation:
( )= ( )+ ( , ) ( )
( , ; )= ( , )+ ( , ) ( , ; )
When the resolvent kernel cannot be obtained in the closed form, we use the method of
successive approximations to find solution up to third order.
( , ; )
For Case I. ( , ; ) =
( ) ( )
− + +⋯+ ( ) =0
( )= ( )+ ( , ; ) ( )
( , ; )
For Case II. ( , ; ) = −
ℎ ( ) ℎ ( )
− + + ⋯+ ( )ℎ = 0
( )= ( )+ ( , ; ) ( )
Key words
Iterated kernel,
Resolvent kernel,
Kernel in polynomial form
Self Assessment
1. Which of the following statement is correct?
A. The resolvent kernel satisfies the integro-differential equation
( , ; )
= ( , ; ) ( , ; )
( )= ( )+ ( , ) ( )
A. ( , ; ) = ( , )+ ∫ ( , ) ( , ; )
B. ( , ; ) = ( , )+ ∫ ( , ) ( , ; )
C. ( , ; ) = ( , )+ ∫ ( , ) ( , ; )
D. ( , ; ) = ( , )+ ∫ ( , ) ( , ; )
A.
B.
C.
D.
( )
4. For = −1 and ( , )= , the resolvent kernel is given as
( )!
A. 1
B. 2
C.
D.
5. The equation ( ) = ( ) + ∫ ( , ) ( ) is
Review Questions
1. What is the method of successive approximation?
2. Solve ( ) = +∫ ( + ) ( )
4. Solve ∫ ( ) = .
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
1.https://www.youtube.com/watch?v=Qo9riz9NZ64
2.https://www.youtube.com/watch?v=u4yhu8QMC2M
Objectives
Students will be able to
define the Green’s function
solve the integral equations using Green’s function
find the solution of boundary value problem using Green’s function
Introduction
In this chapter, we will learn some fundamental functions and then the Green’s function that is
used to solve the integral equations. Let us begin with the adjoint of a differential equation.
( ) + ( ) + ( ) =0 (1)
≡ ( ) + ( ) + ( ) (2)
( )= { ( ) ( )} − { ( ) ( )} (4)
Also
( )=0
Definition 2: Self adjoint equationif the adjoint equation of any linear homogeneous equation is
identical with the equation itself, then the given equation is known as self adjoint equation.
Theorem1: The necessary and sufficient condition that the second order homogeneous linear
differential equation + + = 0 as defined in (1) on ≤ ≤ to be self adjoint is that
( ) = ( ) on ≤ ≤ .
( ) + ( ) + ( ) =0 (1)
is
( ( ) )− ( ( ) )+ ( ) =0 (2a)
i.e.
( ) + {2 ( )− ( )} +{ ( )− ( )+ ( )} = 0 (2b)
Necessary condition: Let (1) be a self adjoint equation. Then (2b) must be identical with (1)
∴2 ( )− ( )= ( ) (3)
And
( )− ( )+ ( )= ( )
⇒ ( )= ( )
Or
( )= ( )+
(3) ⇒ 2( ( ) + ) − ( )= ( )
⇒ =0
Hence
( )= ( )
( ) + ( ) + ( ) =0
( ) + ( ) =0
( ) + ( ) = 0,
( ) ( )
∫ ( ) ∫
where = ( ) and = ( )
( ) .
( )− ( )+ =0
⇒ ( ) − + + =0
Or
( +2 )− − + =0
Or
+ 4 − (2 + 1) + (3 − 6 ) =0 (2)
− = { ( − )} + { ( − ′)}
∴ ( − ) = { ( − )+ ( − )}
=[ ( − )+ ( − )] (3)
Then ∫ ( ) =∫ =1
And ∫ ( − ) =1
1, ∈( , )
Also ∫ ( − ) =
0, ∉( , )
( ) ( ) = ( ) ( )| − ( ) ( )
= − ′(0)
∴ ( ) ( ) = − ′(0)
( ) ( )( ) ( ) (0)
= (−1)
And
( )=− ( )
George Green, English mathematician who was first to attempt to devise a theory of
electricity and magnetism, gave this function.
Green’s Function
Consider a linear homogeneous differential equation of order :
( ) = 0, …(1)
where is the differential operator
≡ ( ) + ( ) + ⋯+ ( )…(2)
where
( ) ( ) ( )( ( ) ( ) ( )(
( )≡ ( )+ ( )+ ⋯+ )+ ( )+ ( ) + ⋯+ ) (4)
( )( ( )(
where, the linear forms , ,…, in ( ), ( ), … , ), ( ), ( ), … , )are linearly
independent.
Suppose that the homogeneous BVP (1) to (4) has only a trivial solution ( ) = 0. Then the Green's
function of this BVP is the function ( , ) constructed for any point , < < and which has the
following properties:
(i) In each of the intervals [ , ), ( , ] the function ( , ) considered as a function of is a solution
of (1), that is
( )=0 (5)
(iii) ( − 1) derivative of ( , ) with respect to at the point = has discontinuity of the first
kind , the jump being equal to − ( ) , i.e.
− =− ( )
… (6)
Theorem: If the boundary value problem (1) to (4), has only a trivial solution ( ) ≡ 0, then the
operator has a unique Green's function ( , ).
Proof :Let ( ) , ( ), … , ( ) be linearly independent solutions of the equation (1).
Then, property (i) of ( , ) ⇒ ( , ) must have the following representation
( )+ ( ) + ⋯+ ( ), ≤ <
( , )=
( )+ ( ) + ⋯+ ( ), < ≤
(8)
where , ,…, , , ,…, are some functions of .
Property (ii) of ( , ) ⇒ at point =
[ ( ) + ⋯+ ( )] − [ ( ) +⋯+ ( )] = 0 ( )
[ ( ) + ⋯+ ( )] − [ ( ) +⋯+ ( )] = 0 ( )
⋮
[ ( ) + ⋯+ ( )] − [ ( ) + ⋯+ ( )] = 0 ( )
Property (iii) of ( , ) ⇒
[ ( ) + ⋯+ ( )] − [ ( ) + ⋯+ ( )] = − ( )
( )
= ( )− ( ), where = 1,2,3, … ,
∴( ), ( ), … , ( )⇒
( )+ ⋯+ ( )=0 ( )
( )+ ⋯+ ( )=0 ( )
⋮
( )+ ⋯+ ( )=−
( )
( )
Equation(9) [ = ( )− ( )] ⇒
= − , = 1,2,3, … , (16)
(15) and (16)⇒
( − ) ( )+ ⋯+( − ) ( )+ ( ) + ⋯+ ( )=0 (17)
Or
[ ( )+ ( )] + ⋯ + [ ( )+ ( )] = ( )+ ⋯+ ( )
Or
( ) +⋯+ ( )= ( )+ ⋯+ ( ), = 1,2, … , (18)
Note: 1. If the BVP (1) to (4)is self adjoint then Green’s function is symmetric, i.e. ( , ) = ( , ).
The converse is also true.
2. If at one of the extremities of an interval [ , ], the coefficient of the highest derivative vanishes,
then the natural boundary condition for boundedness of the solution at = is imposed for
( ) = 0 and at the other extremity the ordinary boundary condition is specified.
≡ ( ) + ( ) +⋯+ ( )
( ) ( ) ( )( ( ) ( ) ( )(
( )≡ ( )+ ( )+ ⋯+ )+ ( )+ ( ) + ⋯+ )
( ) = 0, … (1)
( ) = 0, = 1, 2, … , … (2)
involving homogeneous boundary conditions (2) at the end points = and = of the interval
[ , ].
( )=∫ ( , ) ( ) (5)
( )=∫ ( , ) ( , ( )) (9)
( )= ∫ ( , ) ( ) ( ) −∫ ( , ) ( ) (12)
Example: When the prescribed end conditions are not homogeneous we’d use the modified
method as follows:
Let ( , ) be the Green’s function corresponding to the associated homogenous end conditions.
We look for a function ( ) such that the relation
( )= ( )+∫ ( , ) ( ) (13)
is equivalent to
( )+ ( )=0 (14)
together with the prescribed non-homogeneous end conditions
∫ ( , ) ( ) =− ( ) (15)
( ) =0 (16)
Since the second term in (13) satisfies the associated homogeneous end conditions.
⇒ ( ) in (13) must be a solution of (16) which satisfies the prescribed non-homogeneous end
conditions.
When ( , ) exists, ( ) exists.
Special Case:
Consider a linear homogeneous differential equation of order two:
( ) + ( ) = 0, ≤ ≤ -------(1)
where L is a self adjoint operator
= + = + + -----------(2)
(i) =0 ℎ ≤ <
And =0 ℎ < ≤ ---------------(5)
(ii) and satisfythe equations
( )+ ( )=0
( )+ ( )=0
where A is a constant independent of x and t; then ( ) is a solution of the boundary value problem
given by (1) and (3) iff
( )= ( , ) ( )
Summary
Adjoint equation of 2nd order linear differential equation
( ) + ( ) + ( ) =0 is
{ ( ) ( )} − { ( ) ( )} + ( ) ( )=0
∫ ( − ) =∫ { ( − )+ ( − )}
=[ ( − )+ ( − )]
is called Green’s formula or operator
≡ ( ) + ( ) + ⋯+ ( )…(2)
where
( ) ( ) ( )( ( ) ( ) ( )(
( )≡ ( )+ ( )+ ⋯+ )+ ( )+ ( ) + ⋯+ ) (4)
( )( ( )(
where , the linear forms , ,…, in ( ), ( ), … , ), ( ), ( ), … , )are
linearly independent.
Suppose that the homogeneous BVP (1) to (4) has only a trivial solution ( ) = 0. Then the Green's
function of this BVP is the function ( , ) constructed for any point , < < and which has the
following properties:
(i) In each of the intervals [ , ), ( , ] the function ( , ) considered as a function of is a solution
of (1), that is
( )=0 …(5)
(iii) ( − 1) derivative of ( , ) with respect to at the point = has discontinuity of the first
kind , the jump being equal to − ( ) , i.e.
− =− ( )
… (6)
If the boundary value problem (1) to (4), has only a trivial solution ( ) ≡ 0, then the
operator has a unique Green's function ( , ).
Key words
Adjoint equation,
Dirac delta function,
Green’s function,
Green’s theorem
Self Assessment
1. Which of the following statement is correct?
A. The necessary and sufficient condition that the second order homogeneous linear
differential equation + + = 0 as defined in (1) on ≤ ≤ to be self
adjoint is that ( ) = ( ) on ≤ ≤ .
B. The necessary and sufficient condition that the second order homogeneous linear
differential equation + + = 0 as defined in (1) on ≤ ≤ to be self
adjoint is that ( ) = ( ) on ≤ ≤ .
C. The necessary and sufficient condition that the second order homogeneous linear
differential equation + + = 0 as defined in (1) on ≤ ≤ to be self
adjoint is that ( ) = ( ) on ≤ ≤ .
D. The necessary and sufficient condition that the second order homogeneous linear
differential equation + + = 0 as defined in (1) on ≤ ≤ to be self
adjoint is that ( ) = ( ) on ≤ ≤ .
2. If the BVP (1) to (4) as in summary, is self adjoint then Green’s function is
A. symmetric
B. Hermitian
C. anti-symmetric
D. skew symmetric
, | |<
3. ( )=
0, | | >
and ∫ ( ) =∫ = 1, then
4. The expression − =− ( )
means
Review Questions
1. What are the conditions needed for construction of Green’s function?
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc.
(1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley
(1964)
Web Links
1. https://www.youtube.com/watch?v=C2ZLHf1B_Jc
2. https://www.youtube.com/watch?v=3GJDXCgVM98
Objectives
Students will be able to
understand the nature of problems of calculus of variation
define the function space
derive the properties of the differentiable functional
Introduction
The calculus of variations was created as a tool to calculate minimizers (or maximizers) of certain
mathematical quantities taking values in the set of real numbers.
In the 18th century mathematicians and physicists, such as Maupertuis, d’Alembert, Euler, and
Lagrange, postulated variational principles stating that nature acts with minimal expenditure
(“principle of least action”).
The notion “calculus of variations” goes back to the year 1744 and was introduced by Euler: For
externals of a function, the derivative vanishes, and in the 18 th century that derivative was called
the “first variation.”
The “mathematical quantities” involved in this topic are not ranges of mappings from into , i.e.
the domain of definition of the mappings is infinite-dimensional. And such mappings are called
“functional”.
9.1 Functionals
Variable quantities called functionals play an important role in many problems arising in analysis,
mechanics, geometry, etc.
Example: Consider the set of all rectifiable plane curves. A definite number is associated with
each such curve, namely, its length. Thus, the length of a curve is a functional defined on the set of
rectifiable curves.
Example: Suppose that each rectifiable plane curve is regarded as being made out of some
homogeneous material. Then if we associate with each such curve the ordinate of its center of mass,
we again obtain a functional.
Example: Consider all possible paths joining two given points A and B in the plane. Suppose
that a particle can move along any of these paths, and let the particle have a definite velocity ( , )
at the point ( , ). Then we obtain a functional by associating with each path, the time the particle
takes to traverse the path.
[ ]= ( )
[ ]=∫ [ , ( ), ( )] (1)
Where ( ) ranges over the set of all continuously differentiable functions defined on the interval
[ , ], defines a functional.
By choosing different functions ( , , ), we obtain different functionals.
∫
(ii) [ ]=
∫
Thus, in the space ( , ), the distance between the function ∗ ( ) and the function ( ) does not
exceed if the graph of the function ∗ ( ) lies inside a strip of width 2 (in the vertical direction)
“bordering" the graph of the function ( ).
The space ( , ) consists of all functions ( ) defined on an interval [ , ] which are continuous
and have continuous first derivatives. The operations of addition and multiplication by numbers
are the same as in ( , ) , but the norm is defined by the formula
| | = | ( )| + | ′( )|
The space ( , ) consists of all functions ( ) defined on an interval [ , ] which are continuous
and have continuous derivatives up to order, where is a fixed integer. The operations of
addition and multiplication by numbers are the same as in ( , ) , but the norm is defined by the
formula
| | ( )
= | ( )|
When the Phoenician Princess Dido had fled to North Africa, she asked the Berber king for a bit of
land. She was offered as much land as could be encompassed by an oxhide.
Dido cut the oxhide into fine stripes so that she got a long enough rope to encircle an entire hill
named Byrsa, the center of Carthage.
What shape was the region that she encircled so that area was maximal?
( ) ℎ( ) = ( )( − )( − ) >0
Let be defined as in ∫ [ ( ) − ] =0
And let ℎ( ) = ∫ [ ( ) − ]
Then ∫ [ ( ) − ] ℎ ( ) =∫ ( )ℎ ( ) − [ℎ( ) − ℎ( )] = 0
Also
[ ( ) − ]ℎ ( ) = [ ( )− ]
⇒ ( )− =0
⇒ ( )= ∀ ∈[ , ]
Then [ℎ] ≡ 0∀ ℎ
[ ]
Suppose [ℎ ] ≠ 0 for some ℎ ≠ 0. Then setting ℎ = , = | |
.
We see |ℎ | → 0 as →∞
But
[ℎ ] [ℎ ]
lim = lim = ≠0
→ |ℎ | → |ℎ |
Which is contradiction.
Now let [ ] is not uniquely defined, then
Δ [ℎ] = [ℎ] + |ℎ|
∴ [ ] is uniquely defined.
Theorem 2: A necessary condition for the differentiable functional [ ] to have an extremum for
= i.e. [ℎ] = 0 for = and all admissible ℎ.
Let [ℎ] has a minimum for = ,
By definition of variation,
Δ [ℎ] = [ℎ] + ‖ℎ‖ (1)
Where → 0 as ‖ℎ‖ → 0
∴ for sufficiently small ‖ℎ‖
The sign of Δ [ℎ] and of [ℎ] will be same.
Let [ℎ ] ≠ 0 for some ℎ , then for any > 0, however small ,, we have
[− ℎ ] = − [ ℎ ]
(1) can be made to have either sign for arbitrarily small ‖ℎ‖.
But this is impossible ∵ [ ] is supposed to have a minimum for =
i.e.
Δ [ℎ] = [ + ℎ] − [ ] ≥ 0
For all sufficiently small ‖ℎ‖, which is a contradiction.
Hence the theorem gets proved.
Having understood the definition of functional, we can now proceed to solve the problems on
calculus of variation.
Summary
The calculus of variations was created as a tool to calculate minimizers (or maximizers) of
certain mathematical quantities taking values in the set of real numbers.
Functionals of type ∫ ( , , ) have a localization property consisting of the fact that
if we divide the curve = ( ) into parts and calculate the value of the functional for each
part, the sum of the values of the functional for the separate parts equals the value of the
functional for the whole curve.
The functional [ ] is said to be continuous at the point ∈ if for any > 0, there is a
> 0 such that | [ ] − [ ]| < provided that ‖ − ‖< .
Let Δ [ℎ] = [ℎ] + |ℎ| , where [ℎ] is a linear functional and → 0 as |ℎ| → 0. Then, the
functional [ ] is said to be differentiable and the principal linear part of the increment
Δ [ℎ] i.e. [ℎ] is called the variation of [ ] and is denoted as [ℎ].
The differential of a differentiable functional is unique.
A necessary condition for the differentiable functional [ ] to have an extremum for =
i.e. [ℎ] = 0 for = and all admissible ℎ.
Key words
Dido’s problem, function space, functional, differentiable functional
Self Assessment
1. Which of the following statement is correct?
A. linear functional on ( , )
B. non-linear functional on ( , )
C. linear functional on ( , )
D. nonlinear functional on ( , )
Review Questions
1. What are the motivating problems of calculus of variations?
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
https://mathoverflow.net/questions/109717/functional-variational-derivative-and-the-
leibniz-rule
https://mathworld.wolfram.com/CalculusofVariations.html
Objectives
Students will be able to
analyze the necessary and sufficient condition for the extremum of a functional
remember the Euler’s equation for various forms of functional
create the Euler’s equation to several variables
understand the problem of shortest distance
analyze the invariance of the Euler’s equations
Introduction
In this chapter, we will give necessary conditions for an extremum of a function of the type
[ ] = ∫ ( , , ) with various types of boundary conditions.
The necessary condition is in the form of adifferentialequation that the extremal curve should
satisfy, and this differential equation is called the Euler-Lagrange equation.
Consider the case where the curves are allowed to varybetweentwo fixed points
The functional [ ] has a (relative) extremum for = if [ ] − [ ] does not change its sign in
some neighbourhood of the curve = ( ).
Subsequently, we shall be concerned with functionals defined on some set of continuously
differentiable functions, and the functions themselves can be regarded either as elements of the
space or elements of the space . Corresponding to these two possibilities, we can define two
kinds of extrema:
The functional [ ] has a weak extremum for = if there exists an > 0 such that [ ] − [ ] has
the same sign for all y in the domain of definition of the functional which satisfy the condition
‖ − ‖ < where‖ ‖ is the norm in the space .
And
The functional [ ] has a weak extremum for = if there exists an > 0 such that [ ] − [ ] has
the same sign for all y in the domain of definition of the functional which satisfy the condition
‖ − ‖ < where‖ ‖ is the norm in the space .
= [ + ℎ] − [ ] = ( , + ℎ, +ℎ ) − ( , , )
= [ ( , + ℎ, + ℎ′ ) − ( , , )]
The variation of [ ] is
= [ ( , , )ℎ + ( , , ) ℎ′ ]
= [ ( , , )ℎ + ( , , ) ℎ′ ] =0
Or
− =0
The integral curves of Euler's equation are called extremals. Since Euler's equation is a second-order
differential equation, its solution will in general depend on two arbitrary constants, which are
determined from the boundary conditions ( ) = , ( ) = .
1. When the integrand does not depend on i.e. the functional has the form ∫ ( , )
=0
Or
= ⇒ = ( , )
And then can be calculated.
2. When the integrand does not depend on i.e. if the functional is of the form ∫ ( , ′)
then
− = − +
Or
− = − − ′′ (1)
Multiply (1) by ′
′ − = − − ′′ ′
= ( − )
− =0
⇒ − =
=0
Here integrand =
i.e.
( , )= (1)
=
1+
Squaring both sides
= (1 + )
− =
Or
1
− =
Or
=
1−
Or
=
√1 −
Or
=
√1 −
Separating the variables and integrating
=
√1 −
= − √1 − + (*)
Or
1
− =− 1−
Or
+( − ) = (**)
Using condition
(1) = 0 ⇒ 1 + = (2)
Also
(2) = 1 ⇒ 4 + (1 − ) = (3)
1 − 4 = (1 − ) − =1+ −2 −
Or
−4 = −2 ⇒ =2
− =0
[ ]= , , , ,
= =
= =
− − =0
Or
− − =0
[ ]=∬ ( + ) , here = +
Euler’s equation:
− − =0
0− (2 ) − (2 ) = 0
⇒2 +2 =0
Or
+ =0
− =0
Where = .
If the functional [ ] has an extremum on the function ( ), then ( ) satisfies the differential
equation
− + =0
[ ] = ∫ (16 − + ) , Here = 16 − +
− + =0
Or
32 − 2 =0 (1)
Or
− 16 = 0
− 16 = 0, where =
Auxiliary equation,
− 16 = 0
⇒ = ±2, ±2
∴ = + + sin 2 + cos 2
[ ]= ( , , ,…, , , ,…, )
− = 0, = 1,2, … ,
− = 0 and − =0
⇒2 − (2 ) = 0 and 2 − (2 ) = 0
Or
− ′′=0 and − =0
Or
= ⇒( ) =
Symbolically,
( − 1) = 0, here =
Auxiliary equation,
−1=0
Or
= ±1, ±
∴ = + + cos + sin
=− + − sin + cos
And
= + − cos − sin
Using (0) = 0 ⇒ (0) = 0 ⇒ + − =0 (2)
(0) = 0 ⇒ + + =0 (3)
= −1 ⇒ = −1 ⇒ + − = −1 (4)
=1⇒ + + =1 (5)
Given two points in a plane, what is the shortest path between them?
We certainly know the answer: Straight line. Let’s prove it using the variation method.
Consider an arbitrary path ( ).
Let be the distance function and elementary length
be .
then,
= ( ) +( ) = 1+
= 1+
− =0 (1)
Here ( , , )= 1+
∴ from (1)
1
0− (1 + ) .2 =0
2
Or
=
1+
⇒ = (1 + )
Or
(1 − )=
⇒ = = (say)
Or
= √ = (say)
Or
= ⇒ =
Here =
By Euler’s equation
− =0
Or
2
= ⇒ =
Or
=
2
Or
=
2
⇒ =
2
Integrating both sides, we get
= . +
2 4
Or
= + , where =
Here ( , , )= + +2
− =0
⇒2 +2 − (2 ) = 0
Or
2 +2 −2 =0
Or
− − =0
Or
− =
Symbolic form is
( − 1) = , here ≡
Auxiliary equation,
−1=0
⇒ = ±1
∴ . .= +
For particular Integral, P.I.
P.I.= = ^
=
2
∴ = . .+ . .
= + +
2
[ ]= ( , , ^′ ) , ( )= , ( )=
corresponding to the simplest variational problem. Our approach is to first go from the variational
problem to an n-dimensional problem, and then pass to the limit → ∞ .
Thus, we divide the interval [ , ] into + 1 equal subintervals by introducing the points
= , ,… , = ,( − = Δ ),
and we replace the smooth function ( ) by the polygonal line with vertices
( , ), ( , ), … , ( , ), ( , ),
Where = ( )
Then
[ ]=∫ ( , , ) , ( )= , ( )= ,
( , ,…, )=∑ , , Δ ,
= , , Δ + , , Δ − , , Δ .
Or
= , , Δ + , , Δ − , , .
We note that the expression appearing in the denominator on the left has a direct geometric
meaning, and is in fact just the area of the region lying between the solid and the dashed curves in
Figure
≡ ( , , )− ( , , ′),
We see that the variational derivative is just the left-hand side of Euler's equation and hence the
meaning of Euler's equation is just that the variational derivative of the functional under
consideration should vanish at every point.
This is the analogue of the situation encountered in elementary analysis, where a necessary
condition for a function of n variables to have an extremum is that all its partial derivatives vanish.
Next, we let the area go to zero in such a way that both max |ℎ( )| and the length of the interval
in which ℎ( ) is non vanishing go to zero. Then,
if the ratio
[ + ℎ] − [ ]
Δ
converges to a limit as → 0, this limit is called the variational derivative of the functional [ ] at
the point [for the curve = ( )], and is denoted by
It can be shown that the analogues of all the familiar rules obeyed by ordinary derivatives (e.g., the
formulas for differentiating sums and products of functions, composite functions, etc.) are valid for
variational derivatives.
Suppose that instead of the rectangular plane coordinates and , we introduce curvilinear
coordinates and , where
= ( , ), = ( , ), ≠ 0.
Then the curve given by the equation = ( ) in the -plane corresponds to the curve given by
some equation = ( ) in the - plane. When we make the change of variables, the functional
[ ]= ( , , )
=∫ ( , , ) ,
Where
+
( , , )= ( , ), ( , ), ( + ).
+
− =0
− =0
Let denote the area bounded by the curves = ( ) and = ( ) + ℎ( ), and let denote
the area bounded by the corresponding curves = ( ) and = ( ) + ( ) in the -plane. By
the standard formula for the transformation of areas, the limit as → 0, → 0 of the ratio
approaches the non zero Jacobian
− =0
If ( ) satisifies
− =0
In other words, whether a curve is an extremal is a property which is independent of the choice of
the coordinate system.
In solving Euler's equation, changes of variables can often be used to advantage. Because of the
invariance property just proved, the change of variables can be made directly in the integral
representing the functional rather than in Euler's equation, and we can then write Euler's equation
for the new integral.
Summary
Every strong extremum is simultaneously a weak extremum.
For the functional = ( , , ′), the necessary condition for it to have an extremum is
− = 0.
When the integrand does not depend on i.e. the functional has the form ∫ ( , )
then − =
=0
− + =0
− =0
Key words
Functional, extremal, Euler Lagrange equation, Euler Poisson equation, Euler Ostrogradsky
equation
Self Assessment
1. Which of the following statement is correct?
(a) Every strong extremum is simultaneously a weak extremum.
(b) Every weak extremum is simultaneously a strong extremum.
(c) Strong extremum can never be weak.
(d) None of these
2. For the integrand of the functional = ( , , ′), the necessary condition for it to have an
extremum is
(a) − =0
(b) + =0
(c) − =0
(d) − =0
3. When the integrand of the functional does not depend on , the Euler’s equation takes the
form
(a) − =0
(b) = ( , )
(c) − =0
(d) − =0
4. When the integrand of the functional does not depend on , Euler’s equation takes the
form
(a) = ( , )
(b) = ( , )
(c) = ( , )
(d) of a finite equation
(b) −∑ =0
(c) − + =0
(a) − − =0
(b) −∑ =0
(c) − + =0
8. The extremals of ∫
(a) = +
(b) = +
(c) = +
(d) = +
9. The concept of the ______________ plays the same role for functionals as the concept of the
partial derivative plays for functions of n variables.
(a) variation
(b) variational derivative
(c) extremal
(d) strong extremum
10. The Euler equation for the extremals of ∫ ( + +2 ) results in the differential
equation:
(a) + =
(b) − =
(c) + =
(d) − =
Answers:
a
1
c
2
b
3
d
4
a
5
c
6
b
7
a
8
b
9
d
10
Review Questions
1. What is the necessary condition for the extremal of a functional?
2. Find the curve that gives the minimum distance between two points on a plane
surface?
3. Discuss the weak extremum and the strong extremum.
4. What are the various forms of Euler’s equation when the functional takes different
forms?
5. Discuss the invariance of Euler’s equation.
6. Define the variational derivative.
Further Readings
1. I. M. Gelfand, S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
https://mathworld.wolfram.com/FunctionalDerivative.html
http://liberzon.csl.illinois.edu/teaching/cvoc/node26.html
https://www.ucl.ac.uk/~ucahmto/latex_html/chapter2_latex2html/node5.html
Objectives
Students will be able to
define the brachistochrone
solve the brachistochrone problem
visualize the surfaces of revolution
solve the problem of minimum surface of revolution
Introduction
Many problems involve finding a function that maximizes or minimizes an integral expression.
One example is finding the curve giving the shortest distance between two points — a straight line,
in Cartesian geometry and a difficult one would be if the two points lie on a curved surface. The
mathematical techniques developed to solve this type of problem are collectively known as the
calculus of variations. In this unit we will learn about two famous problems of calculus of
variations, namely, the brachistochrone problem and the problem of minimum surface of
revolution.
Let A and B be two fixed points. Then the time it takes a particle to slide under the influence of
gravity along some path joining A and B depends on the choice of the path (curve), and hence is
functional.
The curve such that the particle takes the least time to go from A to B is called the brachistochrone.
The word brachistochrone is from the Greek for "shortest time".The problem was solved by John
Bernoulli, James Bernoulli, Newton, and L'Hospital.
The brachistochrone turns out to be a cycloid, lying in the vertical plane and passing through A and
B. (The cycloid is the locus of a point on the rim of a circle of radius 'a' rolling along a straight line.)
We'll choose a coordinate system with the origin at point A and the y axis directed downwards.
Points A and B are in this plane as shown.
Let be the velocity of the particle when it has reached point P. Let be the mass of the particle.
= ( ) +( )
= 1+
= 1+
Thus, the element of time to traverse ds is
1+
= =
2 ( )
and the total time of descent is
1+
=
2 ( )
1 1+
=
2 ( )
1
= 1+ ( )
2
Different functions ( ) will typically yield different values for T; we call T a "functional" of .
Our goal is to find the minimum of this functional with respect to possible functions .
Here
= (1)
√
− = (2)
1+ 1 2
− × =
2 1+
1+
− =
1+
Or
1+ −
=
1+
1
⇒ =
1+
Squaring both sides
1
=
(1 + )
Or
1
(1 + )= = ( )
⇒1+ =
Or
−
=
Or
−
=
=
−
Or
= +
−
sin
= 2 sin cos +
− sin
Or
= 2 ∫ tanθ sinθ cos +
=2 sin θ +
1 − cos 2
=2
2
or
sin 2
= − +
2
and
= [1 − cos 2 ]
2
which are the parametric equations ofa cycloid. Thus, the brachistochrone is a cycloid.
Symmetries
Azimuthally or cylindrical symmetry is symmetry around a straight line. Spherical
symmetry is symmetry around a point.
Surface of Revolution
A surface of revolution is a surface generated by rotating a 2-D curve about an axis.
The resulting surface therefore always has azimuthal symmetry.
Example:Among all the curves joining the two given points ( , ) and ( , ), find the one
which generates the surface of minimum area when rotated about the y-axis.
(∵ = + = 1+ )
Hence = 1+
=2 ∫ 1+ (1)
Here integrand = 1+
should be satisfied
1
×2 =
2 1+
Or
= 1+
Squaring both sides
= (1 + )
Or
( − ) =
Or
=
−
Or
=
√ −
⇒ =
√ −
Separating the variables and integrating
=
√ −
⇒ = ∫√ + (2)
= ln + − +
Or
= cosh +
Or
− = cosh
Or
−
= cosh
Example:Among all the curves joining the two given points ( , ) and ( , ), find the one
which generates the surface of minimum area when rotated about the -axis.
The surface area of the strip at A = 2
=2 1+
If − =0 (2)
Here = 1+ = ( , ′)
− =
2
⇒ 1+ − × =
2 1+
Or
(1 + )− = 1+
Hence
= 1+
or
= (1 + )
Or
−1 =
−
⇒ =
Or
−
=
= +
−
⇒ cosh = +
Or
+
= cosh
Or
= cosh , where =
which is catenary.
A note on catenary
The catenary is the equilibrium shape assumed by a chain suspended from two points. One speaks
of a chain rather than any kind of string since a chain with very small links is fully flexible and un
stretchable, as the idealized physical model assumes.
Catena in Latin means “chain.” A chain is very strong, so it does not stretch very much, and
because of the links, it does not support any bending force beyond a single link.
Stretching a soap film between two parallel circular wire loops generates a catenoidal minimal
surface of revolution.
Summary
The brachistochrone is a cycloid.
The solution to the problem of minimum surface of revolution is a catenary.
Key words
Brachistochrone, brachistochrone problem, cycloid, surface of revolution, minimum surface,
catenary
Self Assessment
1. If a particle in the absence of friction will slide from one point to another in the shortest time
under the action of gravity, then the path is
A. a right circular cone
B. a cone
C. a cylinder
D. a cycloid
A. Johann Bernoulli
B. James Bernoulli
C. Issac Newton
D. Leonhard Euler
4. Stretching a soap film between two parallel circular wire loops generates a
A. catenary
B. cycloid
C. catenoid
D. circle
A. catenary
B. helix
C. cycloid
D. great circle
A. straight line
B. point
C. corner
D. curve
A. spherical symmetry
B. cylindrical symmetry
C. rectangular symmetry
D. line symmetry
8. Which of the following is the equilibrium shape assumed by a chain suspended from
two points?
A. Parabola
B. Catenary
C. Hyperbola
D. Cycloid
9. The locus of a point on the rim of a circle rolling along a straight line is
A. paraboloid
B. hyperboloid
C. cycloid
D. catenary
A. paraboloid
B. hyperboloid
C. cycloid
D. catenary
6. A 7. B 8. B 9. C 10. D
Review Questions
1. State the brachistochrone problem.
2. Solve the brachistochrone problem.
3. State the problem of minimum surface of revolution.
4. Among all the curves joining the two given points ( , ) and ( , ), find the one which
generates the surface of minimum area when rotated about the -axis.
5. Among all the curves joining the two given points ( , ) and ( , ), find the one which
generates the surface of minimum area when rotated about the -axis.
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
https://mathworld.wolfram.com/BrachistochroneProblem.html
https://mathworld.wolfram.com/MinimalSurfaceofRevolution.html
Objectives
Students will be able to
Introduction
So far, we have considered functionals of curves given by explicit equations,
e.g. = ( ) ---(1)
in the two-dimensional case.
However, it is often more convenient to consider functional of curves given in parametric form.
Moreover, in problems involving closed curves, it is usually impossible to get along without
representing the curves in parametric form.
∫ ( , , ) ---(2)
we wish to regard the argument as a curve which is given in parametric form, rather than in the
form (1). Then (2) can be written as
̇( )
∫ ( ), ( ),
̇( )
̇( ) = ∫ Φ( , , ̇ , ̇ ) (3)
̇( )
Here ̇ ( ) = , ̇( ) = and = = = ̇( )
(where the over dot denotes differentiation with respect to t) i.e. as functional depending on two
unknown functions ( ) and ( ). The function Φ appearing in the right-hand-side of (3) does not
involve explicitly, and is positive homogenous of degree 1 in ̇ ( ) and ̇ ( ), which means that
Φ( , , ̇, ̇ ) ≡ Φ( , , ̇ , ̇ ) , ∀ > 0.
Conversely, let
Φ( , , ̇ , ̇ )
be a functional whose integrand Φ does not involve explicitly and is positive homogenous of
degree 1 in ̇ and ̇ . We now show that the value of such a functional depends on the curve in the
- plane defined by the parametric equations = ( ), = ( ), and not on the functions ( ),
( ) themselves, i.e. that if we go from to some new parameter by setting
= ( ),
∫ Φ , , , = ∫ Φ( , , ̇ , ̇ ) .
Φ , , , = Φ , , ̇ , ̇
= Φ( , , ̇ , ̇ ) = Φ( , , ̇ , ̇ )
Φ( , , ̇ , ̇ )
to depend only on the curve in the -plane defined by the parametric equations = ( ), = ( )
and not on the choice of the parametric representation of the curve, is that the integrand Φ should
not involve explicitly and should be a positive- homogenous function of degree 1 in ̇ and ̇ .
Now, suppose some parameterization of the curve = ( ) reduces the functional (2) of the form
̇
∫ , , ̇
̇ = ∫ Φ( , , ̇ , ̇ ) . (4)
The variational problem for the right-hand side of (4) leads to the pair of Euler equations
− ̇
= 0, − ̇
=0 (5)
− = 0, here =
̇ Φ − Φ ̇ + ̇ Φ − Φ ̇ =0
∂Φ ∂Φ
Here Φ ̇ = ̇
,Φ ̇ = ̇
,Φ =
∂
and Φ = ∂
[ ]=∫ ( , ,∇ ) (6)
And a transformation
∗
= Φ ( , , ∇ ),
∗
= Ψ( , , ∇ ) (7)
Definition. The functional (6) is said to be invariant under the transformation (7) if [ ∗] = [ ], i.e. if
( ∗, ∗ , ∇∗ ∗ ) ∗
= ( , ,∇ )
∗
[ ]=∫ ∫ + (8)
∗ ∗
= cos − sin + sin + cos
∗
( ∗, ∗)
= + = +
∗ ( , ) ∗
( ∗, ∗)
cos − sin
And = ∗ ∗ = =1
( , ) sin cos
[ ]= ( , ,∇ )
+ =0
= −
What shape was the region that she encircled so that area was maximal?
Thisoperimetric problem is the problem to find the curve = ( ) for which the functional
[ ]=∫ ( , , ) (1)
has an extremum, where the admissible curves satisfy the boundary conditions
( )= , ( )= ,
And are such that,another functional
[ ]=∫ ( , , ) (2)
[ ]= ( , , ′) ,
( )= , ( )= , [ ]=∫ ( , , ′) = , (3)
( + ) ,
− + − =0 (4)
Proof: Let [ ] have an extremum for the curve = ( ), subject to the conditions (3). We choose
two points and satisfies a condition to be stated belowbut is otherwise arbitrary. Then we give
( ) an increment ( )+ ( ) is nonzero only in a neighborhood of .
Using variational derivatives, we can write the corresponding increment Δ of the functional in the
form
Δ = + Δ + + Δ , (5)
where
Δ =∫ ( ) ,Δ =∫ ( ) ,
and , → 0 as Δ , Δ → 0.
Δ = [ ∗] − [ ] = + Δ + + Δ =0 (6)
≠ 0.
Such a point exists, since by hypothesis = ( ) is not an extremal of the functional . With this
choice of , we can write the condition (6) in the form
Δ =− + ′ Δ (7)
Where → 0 as Δ → 0. Setting
=− ,
Δ = + Δ + Δ , (8)
δ = + Δ .
Since a necessary condition for an extremum is that = 0, and since Δ is non zero while is
arbitrary, we finally have
+ = + − + = 0,
= − + − =0
which is precisely equation (4). This completes the proof of the theorem.
Remarks:
∵ Let = + , where =Lagrangian multiplier
The using Euler-Lagrange equation
− =0
⇒ ( + )− ( + ) =0
Or
− + − =0
Example: Find the plane curve of the fixed perimeter and maximum area.
We know that
Area =∫
Perimeter=∫ =∫ 1+ =
Here = and = 1+
Define = + 1+ (1)
To have an extremum,
Euler-Lagrange equation is
− =
1
+ 1+ − 2 =
2 1+
Or
1+ + (1 + )− =
Or
1+ + = ⇒ 1+ = −
Squaring both sides
(1 + )=( − )
Or
( − )
1+ =
( − )
⇒ = −1
Or
( − ) −
=
Or
= (2)
( )
Put ( − ) − = ⇒ −2 =
(2)⇒ − =
√
= −2
√
2√ = −2 +
Or
√ =− + , =
which is a circle.
So now you can answer, what shape was the region that Queen Dido encircled so that area was
maximal!
such that
∫ = 36, (2)
is a parabola.
Here = , =
Let = + , where = Lagrangian multiplier
= +
For (1) to have extremum, given (2)
− =0
Or here,
− =
Or
+ − [2 ] =
⇒ − =
− =
Or
= −
Or
=
−
Integrating both sides
=
−
2
⇒ − = +
Or
4 = ( + ) + (3)
Hence
∫ = 36 along with
4
(1 + ) + = 12
4
and
4
(4 + ) + = 96
4
the exact equation of the parabola can be obtained.
Summary
Suppose that in the functional
∫ ( , , ) ---(2)
we wish to regard the argument as a curve which is given in parametric form, rather than in the
form (1). Then (2) can be written as
̇( )
∫ ( ), ( ),
̇( )
̇( ) = ∫ Φ( , , ̇ , ̇ ) (3)
The function Φ appearing in the right-hand-side of (3) does not involve explicitly, and is positive
homogenous of degree 1 in ̇ ( ) and ̇ ( ), which means that
Φ( , , ̇, ̇ ) ≡ Φ( , , ̇ , ̇ ) , ∀ > 0.
Conversely, let
Φ( , , ̇ , ̇ )
be a functional whose integrand Φ does not involve explicitly and is positive homogenous of
degree 1 in ̇ and ̇ . The value of such a functional depends on the curve in the - plane defined
by the parametric equations = ( ), = ( ), and not on the functions ( ), ( ) themselves.
Φ( , , ̇ , ̇ )
to depend only on the curve in the -plane defined by the parametric equations = ( ), = ( )
and not on the choice of the parametric representation of the curve, is that the integrand Φ should
not involve explicitly and should be a positive- homogenous function of degree 1 in ̇ and ̇ .
If the functional
[ ]= ( , ,∇ )
+ =0
= −
Keywords
Variational problem, isoperimetric problem, Dido’s problem, Noether theorem
Self Assessment
̇( )
1. For the functional ∫ ( ), ( ),
̇( )
̇( ) = ∫ Φ( , , ̇ , ̇ ) , the function Φ is
Φ( , , ̇ , ̇ )
to depend only on the curve in the -plane defined by the parametric equations =
( ), = ( )and not on the choice of the parametric representation of the curve, is that
I. the integrand Φ should not involve explicitly
II. the integrand should be a positive- homogenous function of degree 1 in ̇ and ̇ .
A. Only I is true
B. Only II is true
C. Both I and II are true
D. None is true
5. For the functional at the point ,the necessary condition for an extremum is that
A. =0
B. ≠0
C. >0
D. <0
Review Questions
1. Define an isoperimetric problem with example.
2. Explain the solution of Dido’s problem.
3. State the Noether’s theorem.
4. Find the necessary and sufficient condition for the functional∫ Φ( , , ̇ , ̇ ) to depend
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
https://www.youtube.com/watch?v=EaPyCNUbu40
Objectives
Students will be able to
define a geodesic
distinguish between geodesics of various manifolds
find the geodesics for a plane, sphere, cylinder and cone
Introduction
The problem of finding geodesics, is the problem of finding shortest curves joining two points of
some manifold.
The same kind of problem arises in geometric optics, in finding the paths along which light rays
propagate in an inhomogeneous medium. In fact, according to Fermat's principle, light goes from a
point to a point along the path for which the transit time is the smallest.
Geodesics are curves of shortest distance on a given surface. Apart from their intrinsic interest, they
are of practical importance in the transport of goods and passengers at minimal expense of time
and energy. They are also of paramount importance as escape routes during flights.
Finding geodesics can be accomplished using the methods of differential geometry as well as the
calculus of variation.
As an application, consider the calculation of great circle routes on a sphere, in particular, for any
two cities on the Earth. The airlines have long discovered this, with flight paths between cities as
sections of planes through the Earth’s center.
− = 0, ! ≤ ≤ (1)
In differential calculus, we are looking for those values of which give some function ( ) its
maximum or minimum values. In the calculus of variation, we are seeking the function itself that
makes some integral of , satisfying certain conditions, a maximum or minimum. In the geodesic
problem, we wish to find the curve = ( , , ′) that joins the two points on a given surface such
that the distance between them is minimized. Thus, the problem is to find that integrand which
minimizes the integral of the arc length.:
Here, = ( , , ′),
And = + + .
There are first integrals of (1) based on two simplifying assumptions: (i) if is independent of , i.e.,
= 0. Then
=0⇒ = (4)
(ii) If is independent of :
− = − − =− − − , (5)
− = (5’)
= ( , ), = ( , ), = ( , ) (6)
With = + + , = + + , = + + .
= ∫ [ +2 + ] (8)
2. A Sphere
To find the geodesic for any two points on a
Sphere, the ideal way will be to work with
spherical coordinate system.
The Cartesian and spherical coordinates are
Related as follows:
= sin cos ,
= sin sin ,
= cos
The respective variations are
= cos cos + sin (− sin ) ;
= cos sin + sin cos ;
= (− sin )
The elementary arc length is given as,
( ) =( ) +( ) +( ) = [( ) + sin ( ) ]= ( ) 1 + sin
= 1 + sin
= = 1 + sin
=∫ (let = ′)
Here
= 1 + sin ′ (1)
− =0
=
′
sin 2 ′
∴ (1) ⇒ =
2 1 + sin ′
Or
sin ′
=
1 + sin ′
sin = 1 + sin
Or
[sin − C sin ]=
∴ =
sin (sin − )
Or
csc
=
(sin − )
csc
=
(1 − csc )
Or
csc
=
(1 − (1 + cot ))
Hence
csc
=
1− (1 + cot )
Put = , ⇒ − csc =
∴ =−
1− (1 + )
Or
= −∫
√1 − −
= − sin +
√1 −
− = − sin
√1 −
Or
sin( − )= =
√1 − √1 −
cos
=
√1 − sin
cos
sin sin( − )=
√1 −
Or
cos
sin sin( − )=
√1 −
cos
sin (sin cos − cos sin ) =
√1 −
Or
sin − cos =
√1 −
Which is the equation of a plane passing through the origin, known as the Great Circle.
Thus, the geodesics for a sphere are the great circles as shown.
= ( ) +( ) +( )
= (− sin ) + ( cos ) +( )
= ( ) +( )
= ( ) 1+
Or
= √1 + , where = ′( )
( , , ) = √1 + (1)
− =0
⇒ =
′
(1) ⇒ =
√1 +
= 1+
On squaring both sides, we get
= (1 + )
Or
( − )=
Or
=
−
=
√ −
Or
= , where =
√
⇒ = +
which is the equation of helix. Therefore the geodesic for two points on the cylinder
is a helix.
= ( ) +( ) +( )
And
= sin (− sin ) + sin cos ;
= sin cos + sin sin ;
= cos
∴( ) = + ( sin )
Now
=∫
=∫ + ( sin )
Or
= ∫ 1 + ( sin )
= ∫ 1 + ( sin ) , where =
Here
( , , )= 1 + ( sin ) (*)
− =0
⇒ =
Or
sin ′
=
1+ sin
Or
sin = 1 + ( sin )
Squaring both sides, we get
sin = (1 + ( sin ) )
Or
[ sin − sin ]=
Or
=
sin ( sin − )
Or
=
sin √ sin −
Or
=
sin √ sin −
Integrating both sides
= ∫ (using∫ = sec + )
√
sin
= × sec +
sin
1 sin
= sec +
sin
which is the equation of a helix. Thus, the geodesic for two points on a cone
is a helix.
Summary
The problem of finding geodesics, is the problem of finding shortest curves joining two
points of some manifold.
Geodesic on a plane is a straight line.
Geodesic on a sphere is a great circle.
Geodesic on a cylinder and cone are the helix.
Key words
Geodesic, manifold, straight line, helix, great circle
Self Assessment
1. The geodesic for two points in a plane is
A. Helix
B. Straight line
C. Great circle
D. Cycloid
Review Questions
1. Derive the geodesic for two points on a plane.
2. Define geodesic with example.
3. Derive the geodesic for two points on a sphere.
4. Derive the geodesic for two points on a right circular cylinder.
5. Derive the geodesic for two points on a right circular cone.
Further Readings
1. I. M. Gelfand,S. V. Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
3. G.A. Bliss, "Lectures on the calculus of variations", Chicago Univ. Press (1947)
Web Links
https://www.youtube.com/watch?v=NfqrCdAjiks
Pratiksha, Lovely Professional University Unit 14: Conditional Extremum under Constraints
Objectives
Students will be able to
Introduction
Variational problems involving a conditional extremum are problems in which we find an
extremum of a functional v such that certain constraints are imposed on the functions on which the
functional v is dependent.
Recall the solution of a similar problem dealing with the investigation of the function =
( , , … , ) for an extremum of the constraints
[ , ,…, ] = 0 ( = 1, 2, … , ; < )
The natural way is to solve the system
[ , ,…, ] = 0 ( = 1, 2, … , )
= ( , ,…, );
= ( , ,…, );
⋮
= ( , ,…, );
for , ,…, and substituting their expressions into [ , , … , ], we get the functional
[ , ,…, ] which depends only on n−m arguments that are already independent.
However, both for the functions and functionals, we may apply a more convenient method known
as method of undetermined coefficients.
If we investigate the extremum of the function = ( , ,…, ). with the given constraints.
( , ,…, ) = 0, = 1,2, … ,
Then using the method of undetermined coefficients, we construct a new auxiliary function
̅= +
are the constant factors and the function ̅is now investigated for an unconditional extremum i.e.
we form a system of equations
̅
= 0, = 1, 2, … ,
= ( , , ,…, , , ,…, )
[ , , ,…, ] = 0, = 1,2, …
Then the functional ̅ =∫ ( +∑ ( ) )
− =0 = 1, 2, … , (1)
( , , ,…, ) = 0, = 1,2, … ,
The number of equations + is sufficient to determine the + unknown functions , , … ,
and , , … , and the boundary conditions ( ) = and ( ) = ; = 1,2, … which
must not contradict the constraint equations and will help to determine the 2n arbitrary constants
in the general solution of the system of Euler's equations.
The curves thus found on which a minimum or a maximum of the functional ̅ is achieved will also
be solutions of the original variational problem.
In fact the functions ( ), = 1,2, … , and , = 1,2, … found from the system (1), all the
= 0 and hence ̅ = and if for = ( ), = 1,2, … determined from the system (1) there is
achieved an unconditional extremum of the functional v ̅ i.e. an extremum relative to all close
lying curves (both which satisfy the constraint equations and those who do not), then, in particular
an extremum is also achieved w.r.t. a subclass of curves that satisfy the constraint equations.
= ( , , ,…, , , ,…, )
̅= ( + ( ) ) =
− =0, = 1, 2, … ,
and
( , , ,…, ) = 0, = 1,2, … ,
Proof : In this case, the condition for extremum = 0 takes the form
( + ) =0
( − ) =0
( , , ,…, ) = 0, = 1,2, … ,
It follows that the variations are not arbitrary and the fundamental lemma cannot be applied in
the present form. The variations must satisfy the following conditions obtained by means of
varying the constraint equations =0
= 0, = 1,2, … ,
Multiplying each of these equations term by term by ( ) and integrating from to , we get
( ) = 0, = 1,2, … ,
Adding term wise all these m equations, which are satisfied by the permissible variations , with
the equation
( − ) =0
we will have
+ ( ) − ) =0
Let us assume = +∑ ( )
We get
( − ) =0
( − ) = 0, = 1,2, … ,
or
+ ( ) − = 0, = 1,2, … ,
( , ,…, )
≠0
( , ,…, )
Hence this system has the solution ( ), ( ), … , ( ). Given this choice of ( ), ( ), … , ( ),
the basic necessary condition for an extremum
( − ) =0
( − ) =0
Now for the extremizing functions , , … , of the functional , this functional equation reduces
to an identity for an arbitrary choice of , = + 1, … , it follows that the fundamental lemma is
now applicable
We get
( − ) =0
( − ) = 0, = 1,2, … ,
or
+ ( ) − = 0, = 1,2, … ,
( − ) = 0, = + 1, … ,
( − ) = 0, = 1,2, … ,
And finally we find the functions which achieve a conditional extremum of the functional v and the
factors ( ) must satisfy the system of equations
( − ) = 0, = 1,2, … ,
( , , ,…, ) = 0, = 1,2, … ,
Find the shortest distance between the two points ( , , ) and ( , , ) on the surface
( , , ) = 0.
= + +
= + +
= 1+ +
= 1+ +
By Euler-Lagrange’s equation,
∫ ( , . ) then − =0
And if
↔ ( , , , , ′)
then
− = 0 and − =0
− = 0---- (2)
− = 0---- (3)
And
( , , ) = 0---- (4)
(2), (3) and (4) ⇒ = ( ) and = ( ) on which the conditional minimum of the functional ( )
can be achieved as well as all the ’s.
= ( , , ,…, , , ,…, )
( )= , ( )= , = 1,2, … …
given the finite constraints
( , , ,…, ) = 0, = 1,2, … ,
Now if the constraint equations are the differential equations
This type of constraints is called non holonomic. In this case we will prove a conditional extremum
of a functional is achieved on the same curves on which is achieved an unconditional extremum
of the functional.
̅= + ( ) =
where
= + ( )
Suppose that one of the functional determinants of order m is different from zero say
( , ,…, )
≠0
( , ,…, )
( , ,…, )
≠0
( , ,…, )
Thus, , ,…, are arbitrarily differentiable functions with fixed boundary values and
hence, their variations are arbitrary in the same sense.
+ = 0, = 1,2, … ,
Multiply term by term each of the equations obtained by the undetermined factor ( ) and
integrate from to ; this gives
( ) + ( ) = 0, = 1,2, … ,
Integrating each term of the second integral by parts and taking into consideration that =
and = = 0, we will have
∫ ∑ ( ) − ( ) =0 (1)
=∫ ∑ − =0 (2)
Since =∫ ∑ + =∫ ∑ −
Adding term wise all the equations (1) and equation (2) and introducing the notation
= +∑ ( ) , we have
∫ ∑ ( − ) =0 (3)
Since the variations , = 1,2, … , are not arbitrary, we cannot yet use the fundamental lemma.
Choose factors ( ), ( ), … , ( ) so that they satisfy the equations
− = 0, = 1,2, … ,
( − ) =0
where the variations , = + 1, … , are now arbitrary and hence assuming all variations =
0, except one, and applying the fundamental lemma we obtain
− = 0, = + 1, … ,
Thus, the functions ( ), … , ( ) that render the functional a conditional extremum, and the
factors ( ), ( ) … , ( ) must satisfy the system of + equations
− = 0, = 1,2, … ,
and
= 0, = 1,2, … ,
i.e. they must satisfy Euler equations of the auxiliary functional ̅ , which is regarded as a functional
dependent on the n+m functions , , … , , , … , .
∗ = +
where are constants and write the Euler equations for it. The arbitrary constants , , … in
the general solution of a system of Euler's equations and the constants ,…, are determined
from the boundary conditions
( )= , ( )= , = 1,2, … ,
= , = 1,2, … ,
∗ does ∗
The system of Euler's equations for the functional not vary if is multiplied by some
constant factor and hence is given in the form
∗
=∫ ∑ where = and = , = 1,2, … , .
Now all the functions enter symmetrically and therefore the extremals in the original variational
problem and in the problem involving finding an extremum of the functional ∫ given the
isoperimetric conditions
= , = 1,2, … − 1, + 1, … ,
coincide with any choice of , = 0,1, … , . This property is called the reciprocity principle.
NOTE:
• The finite constraints are
( , , ,…, ) = 0, = 1,2, … ,
=∫ ( , , ,…, ) , ( ) = 0, ( )= = 1,2, … ,
Prove that among all the triangles with fixed area, equilateral triangle has the smallest
perimeter.
We will use the variables , and ℎ as indicated in the figure below:
With the variables , anf ℎ the aarea and the perimeter are given as follow:
1
= ( + )ℎ
2
= + + +ℎ + +ℎ
1+ = ℎ (1)
√
1+ = ℎ (2)
ℎ ℎ
+ = ( + )
√ +ℎ +ℎ
Equations (1) and (2) yield = . So, the above four equations reduce to the following three
equations
ℎ= (3)
= (5)
√
1 1 3 2
= +ℎ = + √3 = + = √
√3 √3 √3 √3
1 1 3 2
= +ℎ = + √3 = + = √
√3 √3 √3 √3
Summary
Given the conditions
= ( , , ,…, , , ,…, )
̅= ( + ( ) ) =
− =0, = 1, 2, … ,
and
( , , ,…, ) = 0, = 1,2, … ,
Among all the triangles with fixed area, equilateral triangle has the smallest perimeter.
Key words
Integral constraints,
geometric constraints,
isoperimetric problem,
Lagrange method of undetermined multipliers
Self Assessment
1. Among all the triangles with fixed area, which triangle has the smallest perimeter
A. Right angled
B. Scalane
C. Isosceles
D. Equilateral
D. None of these
3. The externals in the original variational problem and in the problem involving finding an
extremum of the functional ∫ given the isoperimetric conditions
= , = 1,2, … − 1, + 1, … ,
( , ,…, )
( )
≠ 0 guarantees the
, ,…,
A. Independence of constraints
B. Dependence of constraints
C. Independence of functional
D. Dependence of functional
with the constraint equations ( , , ,…, ) = 0, = 1,2, … , , how many equations are
needed
to determine the unknown functions?
A. m
B. n
C. m+n
D. m-n
Review Questions
1. Isoperimetric problem is an example of (a) integral constraints or (b) geometric constraints?
2. Prove that among all the triangles with fixed area, equilateral triangle has the smallest
perimeter.
3. To obtain the necessary condition in an isoperimetric problem involving finding an
Further Readings
1. I. M. Gelfand,S. V.Fomin, “Calculus of Variations”, Prentice-Hall Inc. (1963)
2. G.F. Hadley, "Nonlinear and dynamic programming", Addison-Wesley (1964)
Web Links
https://www.ucl.ac.uk/~ucahmto/latex_html/pandoc_chapter2.html