P1.T2. Quantitative Analysis
P1.T2. Quantitative Analysis
The information provided in this document is intended solely for you. Please do not freely distribute.
2
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
3
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
1. Generate the data, x(i) = {x(1i), x(2i), …, x(ni)} according to the specified data
generating process (DGP) with random errors drawn from some given distribution; for
example, a common error is the random standard normal, ~N(0,1), but any plausible
distribution is possible.
Each (i) is a single simulation trial or replication. Because the MCS of often simulated
over time, we can think of the trial/replication as a single row where the columns
represent time intervals. For example, if the MCS is a simulation of a daily stock price
going forward one year, the first trial/replication is a single row: [1 row × 250 columns of
future simulated prices; S(1), S(2), …, S(250)].
2. Calculate the test statistic, function, or regression based on trial (i), such that g(i) =
g[x(i)]. For example, the final price (at the end of the row); or if we are valuing an Asian
option, the arithmetic average of the replication’s (row’s) price series.
3. Repeat the first two steps to produce (b) replications. If we continue the matrix view,
this is where we add one additional row for each replication. It is common to generate
many replications, so perhaps the matrix adds 1,000 or 10,000 rows; e.g., if we conduct
1,000 replications, then the MCS matrix looks like: [1,000 rows/replications/trials × 250
columns of future simulated prices; S(1), S(2), …, S(250)].
4. From the replications, estimate the quantity of interest from {g(1), g(2), …, g(b)} which is
likely to be summary statistic. For example, this might be quantile such as the 95th-
percentile of the {g(1), g(2), …, g(b)} so that we have estimated as 95.0% value at risk
(VaR) based on the simulated data!
5. Compute the standard error (SE) to evaluate the accuracy of the quantity of interest.
According to Brooks, we can abstract from the many steps and think about two key stages2:
The first stage is the specification of the (data generating) model. This model may be either
a pure time series model or a structural model.
o Pure time series models are simpler to implement.
o A full structural model is harder because it requires (in addition) the
specification of the DGP for the explanatory variables also.
Once the time series model is selected, the next choice is the probability distribution
used to specify the random errors.
The second stage involves estimation of the parameter of interest in the study.
Viable parameters include, for example, coefficient value in a regression; option value at
maturity; or portfolio value under a set of scenarios that govern asset price dynamics.
1
Chapter 13 (GARP 2020). But also informed by (prior author-specific FRM assignment): Brooks, Chris,
Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
2
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
4
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
Consider interest rate paths. With relatively few assumptions, simulation allows us to randomize
possible future interest rate paths. The two charts below can be found in the Learning
Spreadsheet associated with this reading. Each chart simulates only ten (10) different interest
rate paths over a ten-year (120 month) horizon. The two models employed are Cox-Ingersoll-
Ross (CIR, on the left-hand side) and the Vasicek model (on the right-hand side). In each case,
the “error” (the random component) is a random standard normal variable, N(0,1). These
interest rate paths become inputs into a credit exposure (in the XLS) or can feed into a value at
risk (VaR) model. Such is the flexibility of simulations, which allow us to manufacture data!
3
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
5
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
Sampling variation scales with the square root of the sample size
Let’s denote with x(i) the parameter value of interest for the replication (i). If we generate N =
10,000 replications and retrieve the average, we expect a slightly different result than another
researcher who also conducts 10,000 replications, under an identical model. The difference is
due to the fact that each experiment’s random number matrix (aka, the error matrix) will be
different.
The sampling variation in a Monte Carlo study is measured by the standard error estimate, Sx:
( )
=
where var(x) is the variance of the estimates of the quantity of interest over the N replications.
The key relationship here is the embedded 1/SQRT(n) such that to reduce the Monte Carlo
standard error by a factor of 10, the number of replications must be increased by a factor
of 100. In general, to achieve acceptable accuracy, the number of replications may have to be
set at an very high level; e.g., 10,000 replications.
An alternative way to reduce Monte Carlo sampling error is to use a variance reduction
technique. Two of the intuitively simplest and most widely used variance reduction methods are:
Antithetic variate technique: This involves taking the complement of a set of random
numbers and running a parallel simulation on those so that the covariance is negative,
and therefore the Monte Carlo sampling error is reduced.
Control variate technique: This involves using employing a variable similar to that
used in the simulation, but whose properties are known prior to the simulation. The
effects of sampling error for the problem under study and the known problem will be
similar, and hence can be reduced by calibrating the Monte Carlo results using the
analytic ones.
6
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
where x1 and x2 are the average parameter values for replications sets 1 and 2,
respectively.
The variance of ̅ will be given by:
( )= ( ( )+ ( )+ ( , ))
If no antithetic variates are used, the two sets of MCS replication will be independent,
such that their covariance will be zero, i.e.
( )= ( ( )+ ( ))
However, if we employ antithetic variates, then the covariance will be negative and
the MCS sampling error will be reduced. It might seem that the reduction in MCS
sampling variation from using antithetic variates will be huge since, by definition,
( ,− ) = ( ,− ) = − .
But the relevant covariance is between the simulated quantity of interest; i.e., between
the standard replications and the antithetic variate replications. Most pricing applications
(e.g., option prices) involve a non-linear transformation of u(t) such that covariances
between the terminal prices of the underlying assets based on the draws and based on
the antithetic variates will be negative, but not perfectly negative.
Other similar variance reduction techniques are available, including low discrepancy
sequencing, stratified sampling, and moment-matching. Low discrepancy sequencing is a quasi-
random sequence that involves the selection of a specific sequence of representative samples
from a given probability distribution. Random samples are selected to fill the unselected gaps
left by the probability distribution. In this way, the result is a set of random draws, but they are
deliberately distributed across all of the outcomes of interest. This approach creates MCS
standard error that are reduced in direct proportion to the number of replications rather
than in proportion to the square root of the number of replications; to reduce the MCS standard
error by a factor of 10, we only need to increase the number of replications by a factor of 10,
rather than the typical 10^2 = 100.
4
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
5
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
7
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
The control variate technique employs a whose properties are known prior to the simulation.
Suppose we want to simulate the variable denoted by (x) but we already know the properties of
the variable denoted by (y). The MCS is conducted on both (x) and (y) with the same error
vector; ie, vector of random numbers. Denoting the simulation estimates of x and y by and ,
respectively, a new estimate of x can be derived from:
∗
= +( − ) (Brooks 13.5) 6
The MCS sampling error of this quantity, x∗, will be lower than that of x under certain condition:
the correlation between and must be sufficiently high. For this reason, as GARP explains, “a
good control variate should have two properties:
1. First, it should be [cheap] to construct from x(i). If the control variate is slow to compute,
then larger variance reductions—holding computational cost fixed— may be achieved by
increasing the number of simulations (b) rather than constructing the control variates;
2. Second, a control variate should have a high correlation with g(x). The optimal
combination parameter β that minimizes the approximation error is estimated using the
regression g[x(i)] = α + β*h[x(i)] + v(i)” 7
In this way, control variates reduce the MCS variation by using a single error vector (random
draws) on a related function whose solution is already known. Under the hypothesis that the
sampling error effects for both (i.e., the study function and the already known function) will be
similar, the sampling error can be reduced on the study function
According to the second property above, control variates reduces the MCS sampling error only
if the control and simulation functions are correlated. If this correlation is too low, the variance
reduction will be ineffective. As Brooks explains, we can take the variance of the above (13.5):
( ∗) = ( + − )) (Brooks 13.6)
Because (y) is the known and without sampling variation, var(y) = 0 so that:
( ∗) = ( )+ ( )− ( , ) (Brooks 13.7)
To be effective, the MCS sampling variance must be lower with the control variate than without:
( )− ( , )< → ( , )> ( )
Divide both sides by (var( ),var( ))1/2 obtains the necessary correlation condition:
( )
( , )>
( )
6
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
7
GARP Chapter 13 (2020)
8
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
The advantage of bootstrapping over analytical results is that we can avoid making
strong, or even any, distributional assumptions, since the distribution is empirical. (Further,
variance reduction techniques are also available).
8
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
9
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
“Pseudo-random numbers are generated by complex but deterministic functions that produce
values that are difficult to predict, and so appear to be random. The functions that produce
pseudo-random values are known as pseudo-random number generators (PRNGs) and are
initialized with what is called a seed value. It is important to note that each distinct seed value
produces an identical set of random values every time the PNRG is run.
9
GARP Chapter 13 (2020)
10
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
According to GARP, the two big disadvantages (aka, limitations of simulation) of simulations are
the first two disadvantages above (of the four above): model risk, and potential cost. With
respect to the model risk (i.e., simulations are experiment-specific), GARP says (emphasis
ours), “Monte Carlo simulation is a straightforward method to approximate moments or to
understand estimators’ behavior. The biggest challenge when using simulation to
approximate moments is the specification of the DGP. If the DGP does not adequately
describe the observed data, then the approximation of the moment may be unreliable. The
misspecification in the DGP can occur due to many factors, including the choice of distributions,
the specification of the dynamics used to generate the sample, or the use of imprecise
parameter estimates to simulate the data.”
10
Brooks, Chris, Introductory Econometrics for Finance (Cambridge University Press; 3rd Edition 2014)
11
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
601.1. Betty is an analyst using Monte Carlo simulation to price an exotic option. Her simulation
consists of 10,000 replications where the key random variable is a random standard normal
because the underlying process is geometric Brownian motion (GBM). For example, in Excel a
random standard normal value is achieved with an inverse transformation of a random uniform
variable by way of the nested function NORM.S.INV(RAND()). In this case, each random
standard normal, z(i) = N(0,1), is the random draw that becomes an input into the option price
function. Her simulation succeeds in producing an estimate for the option's price, but Betty is
concerned the confidence interval around her estimate is too large. If her aim is to reduce the
standard error, which of the following approaches is NEAREST to the antithetic variate
technique?
a) She simulates 5,000 pairs of random z(i) and -z(i) such that each pair has perfectly
negative covariance
b) She quadruples the number of replications which will reduce the standard error by 50%
because the sqrt(four) is equal to two
c) She imposes a condition of i.i.d. (independence and identically distributed) on the series
of z(i) which eliminates the covariance term
d) She introduces low-discrepancy sequencing with leads the Monte Carlos standard errors
to be reduced in direct proportion to the number of replications rather than in proportion
to the square root of the number of replications
602.3. Peter used a simple Monte Carlo simulation to estimate the price of an Asian option. In
his first step, he specified a geometric Brownian motion (GBM) which is the same process used
in the Black-Scholes-Merton model. His boss Sally observes, "This is nice work Peter, but the
drawback to this approach is that you've assumed underlying returns are normally distributed.
Yet we know that returns are fat-tailed in practice." How can Peter overcome this objection and
include a fat-tailed assumption in his model?
a) He could assume the errors follow a GARCH process
b) He could assume the errors are drawn from a fat-tailed distribution; e.g., student's t
c) Either he could either assume errors follow a GARCH process or that errors are drawn
from a fat-tailed distribution
d) Monte Carlo simulation cannot overcome this objection; this is a disadvantage of Monte
Carlo simulation in comparison to bootstrapping
12
Licensed to Christian Rey Magtibay at [email protected]. Downloaded August 22, 2021.
The information provided in this document is intended solely for you. Please do not freely distribute.
Answers:
601.1. A. TRUE: She simulates 5,000 pairs of random z(i) and -z(i) such that each pair has
perfectly negative covariance
Brooks: "The antithetic variate technique involves taking the complement of a set of random
numbers and running a parallel simulation on those. For example, if the driving stochastic force
is a set of T N(0,1) draws, denoted u(t), for each replication, an additional replication with errors
given by -u(t) is also used. It can be shown that the Monte Carlo standard error is reduced when
antithetic variates are used.
602.3. C. Either could either assume errors follow a GARCH process or that errors are
drawn from a fat-tailed distribution
Brooks: "13.8.1 Simulating the price of a financial option using a fat-tailed underlying process. A
fairly limiting and unrealistic assumption in the above methodology for pricing options is that the
underlying asset returns are normally distributed, whereas in practice, it is well known that asset
returns are fat-tailed. There are several ways to remove this assumption. First, one could
employ draws from a fattailed distribution, such as a Student’s, in step 2 above. Another
method, which would generate a distribution of returns with fat tails, would be to assume that
the errors and therefore the returns follow a GARCH process. To generate draws from a
GARCH process, do the steps shown in box 13.6. "
13