Lectures On Stochastic Control and Its Applications To Finance Chap 4 Martingale Approach Pham
Lectures On Stochastic Control and Its Applications To Finance Chap 4 Martingale Approach Pham
Huyên PHAM
University Paris Diderot, LPMA
and CREST-ENSAE
[email protected]
https://sites.google.com/site/phamxuanhuyen/home
I. Introduction
V. Verification theorem
VI. Applications : Merton portfolio selection (CRRA utility functions and general
utility functions by duality approach), Merton portfolio/ consumption choice
or
Z ∞
−βt
J(X, α) = E e f (Xt , αt )dt on an infinite horizon
0
• Basic goal: find the optimal control (which achieves the optimum of the objective
functional) if it exists and the value function (the optimal objective functional)
• Mathematical tools
• A process α in the form αs = a(s, Xst,x ) for some measurable function a from
[0, T ] × Rn into A is called Markovian or feedback control.
III. Dynamic programming principle
“An optimal policy has the property that whatever the initial state and initial decision
are, the remaining decisions must constitute an optimal policy with regard to the
state resulting from the first decision”
for any stopping time θ ∈ Tt,T (set of stopping times valued in [t, T ]).
• Stronger version of the DPP
In a stronger and useful version of the DPP, θ may actually depend on α in (2). This
means:
Z θ
v(t, x) = sup sup E f (Xst,x , αs )ds + v(θ, Xθt,x )
α∈A θ∈Tt,T t
Z θ
= sup inf E f (Xst,x , αs )ds + v(θ, Xθt,x ) .
α∈A θ∈Tt,T t
⇐⇒
(ii) For all ε > 0, there exists α ∈ A such that for all θ ∈ Tt,T
Z θ
t,x t,x
v(t, x) − ε ≤ E f (Xs , αs )ds + v(θ, Xθ ) .
t
Proof of the DPP. (First part)
1. Given α ∈ A, we have by pathwise uniqueness of the flow of the SDE for X, the
Markovian structure
θ,Xθt,x
Xst,x = Xs , s ≥ θ,
for any θ ∈ Tt,T . By the law of iterated conditional expectation, we then get
Z θ
t,x
J(t, x, α) = E f (s, Xst,x , αs )ds + J(θ, Xθ , α) ,
t
=⇒
Z θ
v(t, x) ≤ sup inf E f (s, Xst,x , αs )ds + v(θ, Xθt,x ) . (3)
α∈A θ∈Tt,T t
Proof of the DPP. (Second part)
2. Fix some arbitrary control α ∈ A and θ ∈ Tt,T . By definition of the value functions,
for any ε > 0 and ω ∈ Ω, there exists αε,ω ∈ A, which is an ε-optimal control for
t,x
v(θ(ω), Xθ(ω) (ω)), i.e.
t,x t,x
v(θ(ω), Xθ(ω) (ω)) − ε ≤ J(θ(ω), Xθ(ω) (ω), αε,ω ). (4)
By using again the law of iterated conditional expectation, and from (4):
Z θ
t,x t,x ε
v(t, x) ≥ J(t, x, α̂) = E f (s, Xs , αs )ds + J(θ, Xθ , α )
t
Z θ
t,x t,x
≥ E f (s, Xs , αs )ds + v(θ, Xθ ) − ε.
t
The HJB equation is the infinitesimal version of the dynamic programming principle:
it describes the local behavior of the value function when we send the stopping time θ
in the DPP (2) to t. The HJB equation is also called dynamic programming equation.
We assume that the value function is smooth enough to apply Itô’s formula, and we
postpone integrability questions.
I For any α ∈ A, and a controlled process X t,x , apply Itô’s formula to v(s, Xst,x )
between s = t and s = t + h:
Z t+h
t,x ∂v
v(t + h, Xt+h ) = v(t, x) + + L v (s, Xst,x )ds
αs
+ (local) martingale,
t ∂t
where for a ∈ A, La is the second-order operator associated to the diffusion X with
constant control a:
1
La v = b(x, a).Dx v + tr σ(x, a)σ 0 (x, a)Dx2 v .
2
I Plug into the DPP:
Z t+h
∂v
+ Lαs v (s, Xst,x ) + f (Xst,x , αs )ds
sup E = 0.
α∈A t ∂t
I Divide by h, send h to zero, and “obtain” by the mean-value theorem, the so-called
HJB equation:
∂v
+ sup [La v + f (x, a)] = 0, (t, x) ∈ [0, T ) × Rn . (5)
∂t a∈A
Moreover, if v is continuous at T , we have the terminal condition
• Show if possible the existence of a smooth solution to HJB, or even better obtain
an explicit solution
• Verification step: prove that this smooth solution to HJB is the value function of
the stochastic control problem, and obtain as a byproduct the optimal control.
Remark.
In the classical verification approach, we don’t need to prove the DPP, but “only” to
get the existence of a smooth solution to the HJB equation.
V. Verification approach
Theorem
(and satisfying eventually additional growth conditions related to f and g). Suppose
there exists a measurable function â(t, x), (t, x) ∈ [0, T ) × Rn , valued in A, attaining
the supremum in HJB i.e.
admits a unique solution, denoted by X̂st,x , given an initial condition Xt = x, and the
process α̂ = {â(s, X̂st,x ) t ≤ s ≤ T } lies in A. Then,
w = v,
I For any α ∈ A, and a controlled process X t,x , apply Itô’s formula to w(s, Xst,x )
between s = t and s = T ∧ τn , and take expectation:
Z T ∧τn
t,x ∂w αs t,x
E w(T ∧ τn , XT ∧τn ) = w(t, x) + E + L w (s, Xs )ds
t ∂t
where (τn ) is a localizing sequence of stopping times for the local martingale appearing
in Itô’s formula.
I Apply Itô’s formula to w(s, X̂st,x ) for the feedback control α̂. By same arguments
as in the first part, we have now the equality (after an eventual localization):
Z T
∂w α̂s t,x t,x
w(t, x) = −E + L w (s, X̂s )ds + w(T, X̂T )
t ∂t
Z T
t,x t,x
= E f (X̂s , α̂s )ds + g(X̂T ) (≤ v(t, x)).
t
I Together with the first part, this proves that w = v and α̂ is an optimal feedback
control.
Probabilistic formulation of the verification approach
Suppose that the measurable function w on [0, T ] × Rn satisfies the two properties:
• there exists a control α̂ ∈ A with associated controlled process X, such that the
process
Z t
w(t, X̂t ) + f (X̂s , α̂s )ds is a martingale. (11)
0
• Infinite horizon:
hZ ∞ i
−rt
v(x) = sup E e f (Xsx , αs )ds .
α∈A 0
I Assuming a Black-Scholes model for S (with constant rate of return µ and volatility
σ > 0), the dynamics of the controlled wealth process is:
Xt αt Xt (1 − αt ) 0
dXt = dSt + dSt
St St0
= Xt (r + αt (µ − r)) dt + Xt αt σdWt .
• The preferences of the agent is described by a utility function U : increasing and con-
cave function. The performance of a portfolio strategy is measured by the expected
utility from terminal wealth → Utility maximization problem at a finite horizon T :
h 1 2 2 2 i
vt + rxvx + sup a(µ − r)xvx + x a σ vxx = 0, (t, x) ∈ [0, T ) × (0, ∞)
a∈A 2
v(T, x) = U (x), x > 0.
where
1
ρ = rp + p sup a(µ − r) − a2 (1 − p)σ 2 ,
a∈A 2
→
When A = R (no portfolio constraint), the values of ρ and â are explicitly given by
(µ − r)2 p
ρ = + rp.
2σ 2 1 − p
and
µ−r
â = ,
σ 2 (1
− p)
• General utility functions:
U is C 1 , strictly increasing and concave on (0, ∞), and satisfies the Inada conditions:
U 0 (0) = ∞, U 0 (∞) = 0.
I Convex conjugate of U :
• Assume that A = R (no portfolio constraint and complete market) and for simplicity
r = 0 so that HJB is also written as
1 µ2 vx2
vt − = 0,
2 σ 2 vxx
with a candidate for the optimal feedback control:
µ vx
â(t, x) = − .
σ 2 xvxx
Recall the terminal condition:
v(T, x) = U (x).
ṽ(t, y) = E Ũ (yYTt ) ,
dSt = St σdWtQ ,
• The primal value function is obtained by duality relation:
v(t, x) = inf ṽ(t, y) + xy , x > 0.
y>0
Recalling that Ũ 0 = −(U 0 )−1 =: I, the infimum in (12) is attained at ŷ = ŷ(t, x) s.t.
E YTt I(ŷYTt ) = x,
(saturation budget constraint) (13)
and we have
h i
t t t
v(t, x) = E Ũ (ŷYT ) + ŷYT I(ŷYT ) .
- From the martingale representation theorem (or since the market is complete), there
exists α̂ ∈ A s.t.
X̂st,x α̂s
dX̂st,x = X̂st,x σ α̂s dWsQ = dSs ,
Ss
which means that X̂ t,x is a wealth process controlled by the proportion α̂, and starting
from initial capital x at time t.
I From the representation (14) of the value function, this proves that X̂ t,x is the
optimal wealth process:
h i
t,x
v(t, x) = E U (X̂T ) .
2. Merton portfolio/consumption choice on infinite horizon
In addition to the investment α in the stock, the agent can also consume from his
wealth:
→ (ct )t≥0 consumption per unit of wealth
dXt = Xt (r + αt (µ − r) − ct ) dt + Xt αt σdWt .
• The preferences of the agent is described by a utility U from consumption, and the
goal is to maximize over portfolio/consumption the expected utility from intertem-
poral consumption up to a random time horizon:
Z τ
−βt x
v(x) = sup E e U (ct Xt )dt , x > 0.
(α,c) 0
1
(β + λ)v − rxv 0 − sup[a(µ − r)v 0 + a2 x2 σ 2 v 00 ] − sup[U (cx) − cxv 0 ] = 0, x > 0.
a∈A 2 c≥0
1−p
1−p
v(x) = K U (x), with K = .
β+λ−ρ
→ Applications in finance: Bielecki, Pliska, Fleming, Sheu, Nagai, Davis, etc ...
The control decision is a stopping time τ where we decide to stop the process
→ Value function:
"Z #
T X
v(t, x) = sup E f (Xst,x )ds + g(XTt,x ) + c(Xτn− , ζn ) .
(τn ζn ) t n
• Transaction costs and liquidity risk models, execution in limit order books, where
trading times take place discretely
• Real options and firm investment problems, where decisions represent change of
regimes or production technologies
Chapter 2 : Viscosity solutions and stochastic control
dXs = αs Xs dWs ,
But, for the supremum in a ∈ R to be finite and HJB equation (1) to be well-posed,
we must have
v(t, x) ≥ g(x),
I Therefore,
• Need to consider the case where the supremum in HJB can explode (singular case)
and to define weak solutions for HJB equation
M ≤M
c =⇒ F (t, x, r, q, p, M ) ≥ F (t, x, r, q, p, M
c), (3)
F (t, x, r, q, p, M ) = −q − H(x, p, M ),
I From the ellipticity and parabolicity conditions (3) and (4), we deduce that
∂ϕ
F (t̄, x̄, ϕ(t̄, x̄), (t̄, x̄), Dx ϕ(t̄, x̄), Dx2 ϕ(t̄, x̄))
∂t
∂w
≥ F (t̄, x̄, w(t̄, x̄), (t̄, x̄), Dx w(t̄, x̄), Dx2 w(t̄, x̄)) ≥ 0,
∂t
• Similarly, if w is a classical subsolution to (2), then for all test functions ϕ, and
(t̄, x̄) ∈ [0, T ) × O such that (t̄, x̄) is a maximum point of w − ϕ, we have
∂ϕ
F (t̄, x̄, ϕ(t̄, x̄), (t̄, x̄), Dx ϕ(t̄, x̄), Dx2 ϕ(t̄, x̄)) ≤ 0.
∂t
General definition of (discontinuous) viscosity solutions
H is continuous on int(dom(H))
and there exists a continuous function G on Rn × Rn × Sn such that
(x, p, M ) ∈ dom(H) ⇐⇒ G(x, p, M ) ≥ 0.
and so
G(x, p, M ) = −M.
• Viscosity property inside the domain
Remark. In the regular case when the Hamiltonian H is finite on the whole domain
Rn × Rn × Sn (this occurs typically when the control space is compact), the condition
(DH) is satisfied with any choice of strictly positive continuous function G. In this
case, the HJB variational inequality is reduced to the regular HJB equation:
∂v
− (t, x) − H(x, Dx v, Dx2 v) = 0, (t, x) ∈ [0, T ) × Rn ,
∂t
which the value function satisfies in the viscosity sense. Hence, the above Theorem
states a general viscosity property including both the regular and singular case.
Proof of viscosity supersolution property
• Let (t̄, x̄) ∈ [0, T ) × Rn and let ϕ ∈ C 2 ([0, T ) × Rn ) be a test function such that
By definition of v∗ (t̄, x̄), there exists a sequence (tm , xm )m in [0, T ) × Rn such that
when m goes to infinity. By the continuity of ϕ and by (5) we also have that
γm := v(tm , xm ) − ϕ(tm , xm ) → 0.
• Let α ∈ A, a constant process equal to a ∈ A, and Xstm ,xm the associated controlled
process. Let τm = inf{s ≥ tm : |Xstm ,xm − xm | ≥ η}, with η > 0 a fixed constant. Let
(hm ) be a strictly positive sequence such that
γm
hm → 0 and → 0.
hm
We apply the first part of the DPP for v(tm , xm ) to θm := τm ∧ (tm + hm ) and get
Z θm
t ,x
v(tm , xm ) ≥ E f (s, Xstm ,xm , a)ds + v(θm , Xθmm m ) .
tm
Therefore,
h ∂ϕ i
2 2
min − (t̄, x̄) − H(x̄, Dx ϕ(t̄, x̄), Dx ϕ(t̄, x̄)) , G(x̄, Dx ϕ(t̄, x̄), Dx ϕ(t̄, x̄)) ≥ 0,
∂t
which is the required supersolution property.
Proof of viscosity subsolution property
• Let (t̄, x̄) ∈ [0, T ) × Rn and let ϕ ∈ C 2 ([0, T ) × Rn ) be a test function such that
• We will show the result by contradiction, and assume on the contrary that
∂ϕ
− (t̄, x̄) − H(x̄, Dx ϕ(t̄, x̄), Dx2 ϕ(t̄, x̄)) > 0,
∂t
and G(x̄, Dx ϕ(t̄, x̄), Dx2 ϕ(t̄, x̄)) > 0.
Apply Itô’s formula to ϕ(s, X̂stm ,xm ) between tm and θm , we then get after noting that
the stochastic integral vanishes in expectation:
Z θ m
δ ∂ϕ m
γm − + E − − Lα̂s ϕ − f (s, X̂stm ,xm , α̂sm )ds ≤ −δ. (11)
2 tm ∂t
Now, from (8) and definition of H, we have
∂ϕ m
− (s, X̂stm ,xm ) − Lα̂s ϕ(s, X̂stm ,xm ) − f (X̂stm ,xm , α̂sm )
∂t
∂ϕ
≥ − (s, X̂stm ,xm ) − H(s, Dx ϕ(s, X̂stm ,xm ), Dx2 ϕ(s, X̂stm ,xm ))
∂t
> 0, for tm ≤ s ≤ θm .
Due to the singularity of the Hamiltonian H, the value function may be discontinuous
at T , i.e. v(T − , x) may be different from g(x). The right terminal condition is given
by the relaxed terminal condition:
Remark. Denote by ĝ the upper G-envelope of g, defined as the smallest lsc function
above g, and viscosity supersolution to
Then v∗ (T, x) ≥ ĝ(x). On the other hand, since ĝ is a viscosity supersolution to (14),
and if a comparison principle holds for (13), then v ∗ (T, x) ≤ ĝ(x). This implies
• We say that a strong comparison principle holds for (16)-(17) when the follow-
ing statement is true:
Suppose that v and w are two viscosity solutions to (16)-(17). This means that v ∗ is
a viscosity subsolution to (16)-(17), and w∗ is a viscosity supersolution to (16)-(17),
and vice-versa. By the strong comparison principle, we get:
v ∗ ≤ w∗ and w∗ ≤ v∗ .
v ∗ = v∗ = w∗ = w∗ .
Therefore,
v = w, i.e. uniqueness
v∗ = v ∗ , i.e. continuity of v on [0, T ) × Rn .
• Conclusion
The value function of the stochastic control problem is the unique continuous viscosity
solution to (16)-(17) (satisfying some growth condition).
V. Application: superreplication in uncertain volatility model
dXs = αs Xs dWs , t ≤ s ≤ T,
Financial interpretation
α represents the uncertain volatility process of the stock price X, and the function
g represents the payoff of an European option of maturity T . The value function
v is the superreplication cost for this option, that is the minimum capital required
to superhedge (by means of trading strategies on the stock) the option payoff at
maturity T whatever the realization of the uncertain volatility.
The Hamiltonian of this stochastic control problem is
1 2 2
H(x, M ) = sup a x M , (x, M ) ∈ (0, ∞) × R.
a∈[a,ā] 2
→ We shall then distinguish two cases: ā finite or not.
I v is continuous on [0, T ] × (0, ∞), and is the unique viscosity solution with linear
growth condition to the so-called Black-Scholes-Barenblatt equation
1
vt + â2 (vxx ) x2 vxx = 0, (t, x) ∈ [0, T ) × (0, ∞),
2
satisfying the terminal condition
Remark. If g is convex, then v is equal to the Black-Scholes price with volatility ā,
which is convex in x, so that â(vxx ) = ā.
• Unbounded volatility: ā = ∞.
min[ĝ − g , −ĝxx ] = 0.
Let us consider the Black-Scholes price with volatility a of the option payoff ĝ, i.e.
h i
t,x
w(t, x) = E ĝ X̂T ,
where
Then,
∂w
→ This shows that θt = (t, Xt ) is the superhedging strategy associated to the
∂x
superreplication cost w(0, X0 ) of the option payoff g(XT ) in the uncertain volatility
model.
Chapter 3 : Backward Stochastic Differential Equations and
stochastic control
I. Introduction
• Emergence of the theory since the seminal paper by Pardoux and Peng (90): general
BSDEs
Notations
Definition of BSDE
where the data is a pair (ξ, f ), called terminal condition and generator (or driver):
ξ ∈ L2 (Ω, FT , P) , f (t, ω, y, z) is P ⊗ B(R × Rd )-measurable.
• f is uniformly Lipschitz in (y, z), i.e. there exists a positive constant C s.t. for
all (y, z, y 0 , z 0 ):
0 0 0 0
|f (t, y, z) − f (t, y , z )| ≤ C |y − y | + |z − z | , dt ⊗ dP a.e.
Theorem (Pardoux and Peng 90) Under (H1), there exists a unique solution (Y, Z)
to the BSDE (1).
Proof. (a) Assume first the case where f does not depend on (y, z), and consider
the martingale
Z T
Mt = E ξ + f (t, ω)dt Ft ,
0
This pair (Y, Z) exists from Step (a). We then see that (Y, Z) is a solution to the
BSDE (1) if and only if it is a fixed point of Φ.
This shows that Φ is a strict contraction on the Banach space S2 (0, T ) × H2 (0, T )d
endowed with the norm
hZ T i 12
βs 2 2
k(Y, Z)kβ = E e |Ys | + |Zs | ds .
0
We conclude that Φ admits a unique fixed point, which is the solution to the BSDE
(1).
• f is continuous in (y, z) and satisfies a linear growth condition on (y, z). Then,
there exists a minimal solution to the BSDE (1). (Lepeltier and San Martin 97)
Then, by Itô’s formula to v(t, Xt ) between t and T , with v smooth solution to (2)-(3):
Z T Z T
v(t, Xt ) = g(XT ) + f (s, Xs )ds − Dx v(s, Xs )0 σ(Xs )dWs .
t t
It follows that the pair (Yt , Zt ) = (v(t, Xt ), σ 0 (Xt )Dx v(t, Xt )) solves the linear BSDE:
Z T Z T
Yt = g(XT ) + f (s, Xs )ds − Zs dWs .
t t
Remark
We can compute the solution v(0, X0 ) = Y0 by the Monte-Carlo expectation:
Z T
Y0 = E g(XT ) + f (s, Xs )ds .
0
Non linear Feynman-Kac formula
I The time discretization and simulation of the BSDE (6) provides a numerical
method for solving the semilinear PDE (4)-(5)
Simulation of BSDE: time discretization
To get the Z-component, multiply (7) by Wti+1 − Wti and take expectation:
1 h π i
Ztπi = E Yti+1 (Wti+1 − Wti ) Xtπi
∆t
Simulation of BSDE: numerical methods
Here Ē is the empirical mean based on Monte-Carlo simulations of Xtπi , Xtπi+1 , Wti+1 −
Wt i .
→ Efficiency enhanced by using the same set of simulation paths to compute all
conditional expectations.
• Other methods:
→ Important literature: Kohatsu-Higa, Pettersson (01), Ma, Zhang (02), Bally and
Pagès (03), Bouchard, Ekeland, Touzi (04), Gobet et al. (05), Soner and Touzi (05),
Peng, Xu (06), Delarue, Menozzi (07), Bender and Zhang (08), etc ...
IV. Application: CRRA utility maximization
• Consider a financial market model with one riskless asset S 0 = 1, and n stocks of
price process
dSt = diag(St ) µt dt + σt dWt ,
Consider an agent investing in the stocks a fraction α of his wealth X at any time:
• Given a utility function U on (0, ∞), and starting from initial capital X0 > 0, the
objective of the agent is:
I In the sequel, we exploit the DP in the case of CRRA utility functions: U (x) =
xp /p, p < 1. The key observation is the property that the F-adapted process
Vt (α)
Yt := > 0 does not depend on α ∈ A, and YT = 1.
U (Xtα )
• (Ỹ , Z̃) satisfy a quadratic BSDE. Then, we rely on results by Kobylanski (00)
• We conclude that V0 := supα∈A E[U (XTα )] = U (X0 )Y0 , and α̂ is an optimal control.
Markov cases
• Merton model: the coefficients of the stock price µ(t) and σ(t) are deterministic
In this case, the BSDE for (Ỹ , Z̃) = (ln Y, Z/Y ) is written as:
Z T Z T
Yt = f˜(s, Ls , Ỹs , Z̃s )ds − Z̃s dWs
t t
˜
h 1−p i 1
f (t, `, y, z) = p sup (µ(t, `) + σ(t, `)z).a − |σ(t, `)a| + z 2 .
2
a∈A 2 2
→ Ỹt = ϕ(t, Lt ), with a corresponding semilinear PDE for ϕ:
∂ϕ 1
+ η(`).D` ϕ + tr(D`2 ϕ) + f˜(t, `, ϕ, D` ϕ) = 0, ϕ(T, `) = 0.
∂t 2
→ Value function:
V0 = U (X0 ) exp ϕ(0, L0 ) .
Remark The BSDE approach and dynamic programming is also well-suitable for
exponential utility maximization:
→ Many papers: El Karoui, Rouge (00), Hu, Imkeller, Muller (04), Sekine (06),
Becherer (06), etc ...
Chapter 4 : Martingale approach to portfolio optimization problem
2. representation problem: find the dynamic portfolio strategy leading to the optimal
terminal wealth
• utility maximization
sup E U (XTx,θ ) ,
v(x) = x ≥ 0. (2)
θ∈Θ(x)
• Key point: In the complete market framework, any nonnegative contingent claim
is attainable by a terminal wealth if and only if its unique arbitrage price is equal to
the initial capital of the wealth. In other words, for any x ≥ 0, H ∈ L0+ (FT ), the set
of nonnegative FT -measurable random variables, we have:
We seek a zero of the gradient of L(H, y), which leads formally to the equations:
We denote by I = (U 0 )−1 the inverse of U 0 , which is then one to one strictly decreasing
from (0, ∞) into (0, ∞). We then have:
Fix now x > 0. Then, for any H ∈ H(x), y > 0, we have by (3):
E U (H) − yZT H ≤ E Ũ (yZT )] = E U (I(yZT )) − yZT I(yZT ) ,
and so
E U (H) ≤ E U (I(yZT )) − y E ZT I(yZT ) − x . (4)
We shall assume that E[ZT I(yZT )] < ∞ for all y > 0. Then, since I(0) = ∞, I(∞)
= 0, there exists a unique y ∗ = y ∗ (x) > 0 s.t.
E[ZT I(y ∗ ZT )] = x.
Therefore, by setting:
H ∗ = I(y ∗ ZT ) ∈ H(x),
∗
and we notice in particular for t = 0, that x = X0x,θ = EQ [I(y ∗ ZT )] = M̂0 . The
general method for determining the optimal portfolio strategy θ∗ consists then of
computing the positive Q-martingale M̂t , and write it by derivation in terms of the
Brownian motion W Q under Q:
and then to identify with the general form for the dynamics of the wealth process
∗
dXtx,θ = θt∗ dSt = θt∗ St σt dWtQ , (6)
so that
φt M̂t θt∗ St φt
θt∗ = , or equivalently in proportion αt∗ := x,θ∗
= , 0 ≤ t ≤ T.
σt St Xt σt
In some particular cases (see Examples below), the derivation (5) can be made explicit
with Itô’s formula. In the general case, the representation (5) is not explicit and
involves Clark-Ocone formula.
Example: Logarithmic utility function
We consider an utility function of the form
U (x) = ln x.
Therefore, by identifying with the dynamics (6) of the wealth process, we obtain the
optimal portfolio strategy in terms of number of shares :
∗
∗ λt δ M̂t µt Xtx,θ
θt = = , 0 ≤ t ≤ T,
σt St (1 − γ)σt2 St
or equivalently in proportion :
∗ θt∗ St µt
αt = x,θ∗
= , 0 ≤ t ≤ T.
Xt (1 − γ)σt2
This generalizes the result obtained by Bellman method in Chapter 1 for constant
coefficients.
III. VaR hedging problem
• In the context of complete market model as in the Black-Scholes model, the price of
any contingent claim is equal to its replication cost, that is the initial capital allowing
to construct a portfolio strategy whose terminal wealth replicates at maturity its
payoff. This price is computed as the expectation of the (discounted) claim under
the unique risk-neutral martingale measure. In particular, by putting up an initial
capital at least equal to the cost of replication, we can stay on the safe side with a
portfolio strategy whose terminal wealth superhedges the payoff of the option.
• What if the investor is unwilling to put up the initial amount required by the
replication? What is the maximal probability of a successful hedge the investor can
achieve with a given smaller amount? Equivalently, one can ask how much initial
capital an investor can save by accepting a certain shortfall probability, i.e. by being
willing to take the risk of having to supply additional capital at maturity in e.g. 1%
of the cases. This question corresponds to the familiar “Value at Risk” (VaR)
concept, which is very popular among investors and practioners. Just as in VaR a
certain level of security (e.g. 99%) is chosen, and the investor is looking for the most
efficient allocation of capital.
VaR optimization problems
• Given a contingent claim represented by a nonnegative FT -measurable random
variable H ∈ L0+ (FT ), and for some fixed initial capital x stricty smaller than the
price of the claim, i.e. x < EQ [H], we are looking for:
The set {XTx,θ ≥ H} is called the success set associated to a strategy θ ∈ Θ(x), and
v(x) the maximal probability of success sets.
I Given shortfall probability ε ∈ (0, 1), we are looking for the least amount of initial
capital which allows us to stay on the safe side with probability 1 − ε, i.e. we want
to determine the minimal initial capital x s.t. there exists an admissible strategy θ
∈ Θ(x) with
P[XTx,θ ≥ H] ≥ 1 − ε.
EQ [H1A ] ≤ x. (9)
H ∗ := H1A∗
solves the VaR optimization problem (7), and the corresponding success set coincides
almost surely with A∗ .
Proof. 1) Let θ ∈ Θ(x) be an admissible strategy, and denote by A = {XTx,θ ≥ H}
its corresponding success set. Then, we have
EQ [H1A ] ≤ EQ [XTx,θ ] = x.
then A∗ is solution to the Neyman-Pearson test, i.e. (8)-(9). Indeed, let A ∈ FT s.t.
Q∗ [A] ≤ α. By definition of A∗ , we then get
dP
∗
(1A∗ − 1A ) ∗
−c ≥ 0, Q∗ a.s.
dQ
and so
∗ Q∗
h dP i
P[A ] − P[A] = E (1A − 1A ) ∗
∗
dQ
Q∗
≥ E [(1A∗ − 1A )c∗ ] = c(Q∗ [A∗ ] − Q[A]) = c(α − Q[A]) ≥ 0,
Remark. The solution to Neyman-Pearson problem relies on the condition that the
set A∗ of (10) satisfies (11). This condition is clearly satisfied whenever the function
h i h i
c → Q dQ∗ > c is continuous at c , i.e. Q dQ∗ = c = 0. Since Q∗ is absolutely
∗ dP ∗ ∗ dP ∗
where
1 1 √ √
d1 = √ ln(S0 /K) + σ T , d2 = d1 − σ T ,
σ T 2
and N (.) is the distribution function of the standard normal random variable.
Now, suppose we start from an initial capital x smaller than the Black-Scholes price,
and we want to maximize the probability of success set. From the above result, the
optimal strategy consists in replicating a knock-out option H1A where the success
set A is of the form
n dP o
A = > const.H .
dQ
Due to (12), we can write
µ/σ 2
A = ST > a(ST − K)+ ,
EQ [H1A ] = x. (13)
where
1 2
c = S0 exp µ − σ T + σb .
2
Hence, the knock-out option H1A can be written as
H1A = (ST − K)+ 1{ST <c} = (ST − K)+ − (ST − c)+ − (c − K)1{ST >c} ,
which is a combination of two call options and of a binary option. We then calculate
the maximal probability of success set
√
P[A] = N (b/ T ),
for some c1 < c2 . Hence, the knock-out option H1A can be written again as a
combination of call options and digital options. We calculate the maximal probability
of success set
√ √
P[A] = N (b1 / T ) + N (−b2 / T ),
with constants b1 and b2 determined from the condition (13), which can be written
explicitly (after some straightforward computations) as :
x = EQ [H1A ]
−b − µ T + σT −b − µ T
1 σ 1
= S0 N (d1 ) − KN (d2 ) − S0 N √ + KN √ σ
T T
−b − µ T + σT −b − µ T
2 2
+S0 N √σ − KN √ σ .
T T
Numerical illustration
We illustrate the amount of initial capital that can be saved by accepting a certain
shortfall probability.