Sequential Calibration of Options
Sequential Calibration of Options
www.elsevier.com/locate/csda
Abstract
Robust calibration of option valuation models to quoted option prices is non-trivial but crucial for good performance. A framework
based on the state-space formulation of the option valuation model is introduced. Non-linear (Kalman) filters are needed to do
inference since the models have latent variables (e.g. volatility). The statistical framework is made adaptive by introducing stochastic
dynamics for the parameters. This allows the parameters to change over time, while treating the measurement noise in a statistically
consistent way and using all data efficiently. The performance and computational efficiency of standard and iterated extended Kalman
filters (EKF and IEKF) are investigated. These methods are compared to common calibration such as weighted least squares (WLS)
and penalized weighted least squares (PWLS). A simulation study, using the Bates model, shows that the adaptive framework is
capable of tracking time varying parameters and latent processes such as stochastic volatility processes. It is found that the filter
estimates are the most accurate, followed by the PWLS estimates. The estimates from all of the advanced methods are significantly
closer to the true parameters than the WLS estimates which overfits data. The filters are also faster than least squares methods. All
calibration methods are also applied to daily European option data on the S&P 500 index, where the Heston, Bates and NIG-CIR
models are considered. The results are similar to the simulation study and it can be seen that the overfitting is a real problem for the
WLS estimator when applied complex models.
© 2007 Elsevier B.V. All rights reserved.
1. Introduction
The predictive power of parametric option valuation models was recognized immediately after the Black and Scholes
(1973) model was introduced, cf. Campbell et al. (1997). More recent models have included jumps, see Merton (1976),
stochastic volatility, see e.g. Hull and White (1987) and Heston (1993) and state dependent diffusion terms, see e.g.
Dupire (1994) and Derman and Kani (1994) in order to improve the predictive power further. Today, models use a
combination of jumps, stochastic volatility and local volatility, see e.g. Bates (1996), Duffie et al. (2000), Carr et al.
(2003b), Chernov et al. (2004), and Carr and Wu (2004). The purpose of this paper is to outline a framework in which
these advanced models can be adaptively calibrated to data.
There are some issues when working with option data. A serious problem is the absence of a single quoted mar-
ket price. Instead, ask and bid prices are quoted and it is common practice to approximate the market price by the
0167-9473/$ - see front matter © 2007 Elsevier B.V. All rights reserved.
doi:10.1016/j.csda.2007.08.009
2878 E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891
mid-price of these. Using the mid-price increases the risk for overfitting as any parameter vector yielding option
values within the bid–ask spread is acceptable with respect to the arbitrage conditions. The overfitting is accentu-
ated when using state of the art models as complex models are more prone to overfitting than simpler, transparent
models.
Another difficulty arises from the fact that market properties change over time. This can be e.g. macro-economic
events as well as investor preferences, corresponding to time varying P and Q measures, respectively. Time varying
parameters make adaptive calibration necessary. Less complex models are easier to adapt to changing conditions and
are therefore often used.
Models can be calibrated using data mining methods if the main purpose is to interpolate or predict prices of
contracts of the same type as the calibration instruments. However, good parameter estimates are needed if the pur-
pose of the calibration is applications like hedging, risk management or pricing exotics. The most common calibra-
tion technique is some version of daily least squares estimation, which provides good in sample predictions but is
also known to be notoriously non-robust to outliers giving bad parameter estimates, cf. Cont and Tankov (2004).
A problem frequently associated with the least squares calibration methods is multimodal loss functions, making
numerical optimization difficult. Several algorithms designed to reduce this problem by adding penalties have been
suggested. However, the main purpose of these is to improve the numerical stability, not the quality of the parameter
estimates.
Adaptive estimators are rare in the academic literature but commonly used in the industry, often using some ad hoc
adjustment. The dominating approach is to estimate the parameters using rolling non-linear regression techniques,
often in combination with some penalty.
This paper suggests that standard least squares methods should be replaced by a non-linear (Kalman) filter method,
as this decomposes the residual error into changes in the latent variables (e.g. volatility) and pricing errors. The Kalman
filter will automatically use all data but assign more weight to recent data in a statistically sound way.
The method is made adaptive by introducing the parameters as latent states, and assigning dynamics to these. Filtering
the augmented latent process will estimate the original latent state and parameters simultaneously. We find that the
proposed filter method is both faster than and at least as accurate as the least squares methods on the estimation and
validation set and significantly more accurate when pricing other contracts.
2. Option valuation
It is essential that the theoretical value of the calibration instruments (i.e. the options used in the calibration) can be
calculated with high accuracy using limited computational resources. These requirements rule out many quoted options,
essentially restricting the available instruments to European call and put options. The values of these instruments can
be obtained using Fourier methods, cf. Carr and Madan (1999).
The Fourier transform based techniques are both fast and fairly accurate. The main limitation is that we need to
know the characteristic function for the log stock price in closed form. However, for a fairly large class of stock price
models e.g. Black–Scholes (Black and Scholes, 1973), Merton (Merton, 1976), Heston (Heston, 1993), Bates (Bates,
1996), Normal inverse Gaussian (NIG) (see e.g. Barndorff-Nielsen, 1997 and the references therein ), variance Gamma
(VG) (Madan and Seneta, 1990), BNS (Barndorff-Nielsen and Shepard, 2001, 2002), CGMY (Carr et al., 2002),
finite moment log-stable (FMLS) (Carr and Wu, 2003), NIG-CIR model (Carr et al., 2003b), a class of time-changed
exponential Lévy models (Carr and Wu, 2004; Huang and Wu, 2004), etc. the characteristic function is available in
closed form. We will consider some of these models in detail below (cf. Section 2.1). The Fourier transform based
methods have been used in financial mathematics for some time now (see Carr and Madan, 1999) for a good refer-
ence to start with. A long list of references to articles using Fourier transform based methods can be found in Carr
et al. (2003a).
In this paper we consider the Bates model (Bates, 1996) and a sub-model (cf. below) as well as the NIG-CIR model
(cf. below). This choice of models is based on the results in Schoutens et al. (2004).
E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891 2879
where W (S) and W (V ) are standard Brownian motions with correlation , is the mean reversion rate, is the mean
reversion level and with V0 as the initial value of V. Further, J is a compound Poisson process with intensity , having
jumps J such that log(1 + J ) ∈ N(, 2 ) and drift −(exp(2 /2 + ) − 1). This choice of drift for J forces the
discounted stock price process to be a martingale under Q. The Bates model contains several models (Black–Scholes,
Merton and Heston) as sub-models.
St = S0 exp(XIt ), (3)
t
where Xt is a NIG Lévy process, having parameters , , and where It = 0 Vs ds, Vt has dynamics according to
Eq. (2). Here the process {Vt }t 0 is assumed to be independent of the process {Xt }t 0 . It is, however, possible to
calculate the moment generating function without this assumption (see Huang and Wu, 2004; Carr and Wu, 2004).
Lee (2004) treats error bounds for a fast Fourier transform (FFT) implementation of the Fourier transform method
and list the generalized Fourier transforms for some common pay-off functions such as e.g. the European call option.
Using FFT, one can obtain prices on a regularly spaced grid of log strike levels. Options observed on the market are
usually regularly spaced in the strike level, not the log strike level. If we are to evaluate the prices for a set of European
call options we need to fit the corresponding log strike levels into a regular spaced grid. The accuracy of FFT depends
on the grid spacing used, the interval on which integrand is evaluated and the properties of the characteristic function
of the log stock price. One FFT is needed for each time to maturity. This will lead to extra work if we have several
dates of maturity each with only a few strike levels. In order to avoid using different grids for different strike levels and
maturities we will here instead consider a Gauss–Laguerre quadrature implementation for the inverse Fourier transform.
In order to price a European call option we need to calculate the inverse Fourier transform (Lee, 2004):
In order to price a Binary call option we instead need to calculate the inverse Fourier transform of (see Lee, 2004,
Eq. (5.1)):
1 ∞ e−ik − k est ( +i ) M
XT |t ( +i )
Bt (K, T − t) = p(t, T ) d . (5)
2 −∞ +i
where the last equality comes from the fact that the real part of the integrand is an even function of . This integral can
now be approximated by a Gauss–Laguerre quadrature formula. In general we have that the Gauss–Laguerre quadrature
formula approximates an exponentially weighted integral from zero to infinity as
∞ n
(n) (n)
e−x f (x) dx ≈ wj f (xj )
0 j =1
(see e.g. Abramowitz and Stegun, 1972, p. 890). This paper use n = 100, which has turned out to be sufficient to get
reasonably accurate prices compared to the accuracy of the observed market prices. We do not claim that this holds true
in general, only for the models and options considered. Some care should be taken for short maturities and deep in the
money options. The weights and abscissas are pre-calculated and stored in order to increase the computational efficiency
of the algorithm. The complexity of the Gauss–Laguerre pricing algorithm will then be of the order n × noption , i.e. the
number of weights times the number of different options.
3. Calibration
Robust calibration of option valuation models to real world option prices is non-trivial but important for good
performance. Calibration is entangled with several difficulties, rendering many standard estimators suboptimal. This
section will review standard calibration methods, discuss some of the limitations of these and suggest a new framework
leading to better use of data.
A serious data problem is the absence of a single quoted market price. Instead, ask prices ctAsk (Ki , i ) and bid prices
Bid
ct (Ki , i ) are quoted. The market price is often approximated by the mid-price defined as
The dominating (textbook) calibration method, cf. Bates (1996), Hull (2002), Cont and Tankov (2004) and Schoutens
et al. (2004) is weighted least squares (WLS), i.e. taking the parameter (vector) that minimize the weighted sum of the
E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891 2881
squared difference between the observed mid-price and the model predicted price
Ns
t
LWLS
t () = s,i (cs⋆ (Ki , i ) − csModel (Ki , i ; ))2 , (7)
s=1 i=1
ˆt = arg min LWLS (). (8)
t
∈
It is statistically optimal (minimal variance of the estimates) to choose s,i as the inverse of the variance of the
residuals, although economic argument may suggest other weights. It is common to choose s,i = 0 ∀s < t to increase
the adaptiveness of the calibration, and to choose t,i as constant or proportional to the inverse of squared bid–ask
spread, thereby relating the size of the ask–bid spread to the quality of the quoted prices
1
t,i ∝ . (9)
(ctAsk (Ki , i ) − ctBid (Ki , i ))2
The WLS estimator using these weights has one strong limitation. An implicit assumption is made that old data do
not improve the estimation, as the summation is restricted to current data. The implication of this is quite dramatic,
as it implies that the past is of limited use to predict the current prices, and hence currently available information are
of limited use for predictions of the future! Another problem with this calibration setup is that it uses a small set of
observations, easily overfitting data. It is also known to be numerically difficult to find the parameters minimizing the
loss function. The latter is discussed in e.g. Cont and Tankov (2004), where a penalty function P (, 0 ) is added to the
loss function
LPWLS
t () = LWLS
t () + P (, 0 ). (10)
Here, 0 is a reference parameter vector and P (, 0 ) is a positive, locally convex function satisfying P (, 0 ) 0 and
P (0 , 0 )=0. Assuming that P (, 0 ) is globally convex and that LWLS
t () is well defined will, for a sufficiently large
make LPWLS
t () globally convex, thus improving the performance of most numerical optimization schemes. However,
questions regarding exclusion of data remain unanswered. The penalty may also influence the loss function too much,
leading to biased parameters.
Several penalties have been used in the literature. Relative entropy is discussed in Cont and Tankov (2004). Another
approach related ridge regression or Tikhonov regularization (see e.g. Draper and Smith, 1998, Chapter 17) is a quadratic
penalty. We use the latter, using the deviation from the previously estimated parameters (0 = t−1 ) as penalty. Hence,
we get
This penalty decrease the overfitting as the paths of the estimated parameters are smoothed over time, while simultane-
ously being easy to implement. It is also computationally advantageous, as many numerical optimization routines are
exact for quadratic problems. We have used Matlabs routine for non-linear least squares minimization, lsqnonlin,
to obtain the parameters and latent state.
The penalty parameter can be chosen using different strategies. The literature on ridge regression suggests that the
parameter is chosen using the connections to Bayesian statistics, cross validation, Morozov discrepancy principle or
L-curve methods. We use the discrepancy principle discussed in Cont and Tankov (2004), choosing as
Nt
ˆ = sup : t,i (ct⋆ (Ki , i ) − ctModel (Ki , i ; ()))2 d0 d1 , (12)
i=1
Nt
d0 = t,i (ct⋆ (Ki , i ) − ctModel (Ki , i ; WLS
t ))2 , (13)
i=1
where d1 is chosen as d1 = 1.5. This improves the stability of the optimization, cf. Cont and Tankov (2004).
2882 E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891
The least squares methods presented above do not explicitly model the time variation of the parameters, and fails to
separate the effects from the measurement errors and the time varying parameters consistently. Furthermore, the least
squares estimators do not use the structure of the latent processes when estimating the current value. This is hardly
optimal use of data.
Using the state-space formulation of the model, and interpreting the market price ambiguity generated by the ask–bid
spread as measurement errors addresses some of the problem mentioned. The model is then given as
where t is a zero mean random vector having covariance matrix R = diag(t,i )/16, t,i is defined as in Section 3.1,
is a constant parameter vector and Vt is the latent factor (volatility) propagated using the exact transition kernel. Filter
algorithms use all observations (not only the most recent observation) to estimate the latent process, and will weight
them optimally. This setup can be found in e.g. Bakshi et al. (2007), where an unscented Kalman filter was used to
estimate the latent volatility and the parameters were estimated offline using the quasi-maximum likelihood estimator.
The parameters can also be estimated by augmenting the latent state Vt with the parameter vector , thereby estimating
the parameters online using non-linear filtering. This is computationally easier than maximizing the quasi-likelihood.
The online calibration model is given as
This is a dynamic Bayesian method, but does not address the problem of time varying parameters. Adaptive estimation
is achieved if the state-space model is reformulated as a self-organizing state-space model, augmenting the latent state
vector Vt with (a transformed version of) the parameter vector ˜ t = (t ), where the transformed parameter vector
is modelled as a random walk, cf. Anderson and Moore (1979) and Kitigawa (1998). This defines the self-organizing
state-space model as
Adaptively estimating the parameters and the latent volatility is thus transformed into filtering of latent processes,
a problem for which statistical methods have been developed during the last decades, Anderson and Moore (1979),
Doucet (2001) and Künsch (2001).
Updating: The mean and covariance of the filter density are given by
3.2.2. Implementation
The filter algorithm cannot be applied to data without specifying an initial distribution of the latent states. We have
initialized the filter using the asymptotics given by the WLS estimator, i.e. taking m0 = WLS1 and P0 = Cov(WLS 1 ).
This is a reasonable prior and ensures that the filter has good starting values, and is not worse off than the WLS when
comparing the calibration methods.
Many parameters and latent states are restrained to compact or semi-infinite spaces. These have been transformed
onto the real line, as the filters are based on Gaussian approximations. We denote the transformed parameters by , ˜
2884 E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891
cf. Eqs. (19)–(21). Neglecting to transform the parameters may cause the algorithm to break down. Standard, bijective
transformations have been used, e.g. taking the logarithm of variables with support on the positive real line and applying
the inverse hyperbolic tangent (arctanh) to variables with support on [−1, 1]. The transition kernel for latent volatility
is approximated by a Gaussian distribution, using the exact conditional moments. The implemented model is given by
P
= P (1 + V ), = . (37)
1 + V
Several studies estimating the risk premiums have found that their estimates are either weakly significant or even
insignificant, cf. Polson and Stroud (2003) on S&P 500 data and Carr and Wu (2007) on currency option data. We have
been unable to get statistically significant estimates on our (shorter) set of data, and have therefore approximated the
P-dynamics with the Q-dynamics (V = 0), although simultaneous estimation of Q-parameters and risk premiums is
preferred when possible.
What remains is to specify the covariance matrix of . We have for computational and statistical simplicity assumed
that increments of each parameter are pair wise independent. It is well known that stochastic volatility is clearly needed
to fit data, while the gain from introducing jumps is smaller (but still significant). We have therefore chosen to allow for
different covariances, assigning covariance Qc to parameters related to the continuous (stochastic volatility) component
and covariance QJ to parameters related to the jump component. These are different, while covariances in each group
are identical. The size of the covariances has been estimated using quasi-maximum likelihood on an independent set
of data, simulated using parameters that resemble those we are interested in estimating.
4. Simulation study
The methods are first tested in a simulation study. Data have been simulated from the Bates model, cf. Section 2.1,
as this model includes jumps and stochastic volatility. The initial Q-parameters were chosen according to Bakshi et al.
(1997) to mimic S&P 500 index as closely as possible, see Table 1.
The Bates model has seven parameters and one latent state which are grouped together in the augmented latent state
vector, x = (, , V , , Vt , , , ). Prices have been generated with constant parameters except Vt which follows a
stochastic process according to (2), and and which have been varied as depicted in Fig. 1. The purpose of this setup
is to test whether the methods can track changing parameters.
The index level and latent volatility were simulated under the P-dynamics, where we used V =−0.25, while the jump
parameters were identical under P and Q. Using different volatility parameters under P and Q, while filtering using
only Q-parameters will emulate real data, and test our approximation (cf. Section 3.2) under controlled conditions.
The total length of the data set is 200 trading days. Hundred European call options over a range of 20 strikes with
moneyness between 0.65 and 1.35 and five maturities were generated each trading day. Our options data have bid–ask
Table 1
Parameters for the Bates model used in the simulation study
Parameter S0 V V0
Value 1000 2.03 0.04 0.38 −0.7 0.0576 0.59 −0.05 0.07
E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891 2885
κ ξ σV ρ
3.5 0.06 0.5 −0.3
3 0.055 −0.4
0.45
0.05 −0.5
2.5
0.045 0.4
−0.6
2
0.04 0.35 −0.7
1.5
0.035 −0.8
1 0.3
0.03 −0.9
Vt λ µ δ
0.09 2 0.6 0.2
0.08 1.8 0.4 0.18
0.07 1.6 0.16
1.4 0.2 0.14
0.06
0.05 1.2 0 0.12
1 0.1
0.04 −0.2
0.8 0.08
0.03
0.6 −0.4 0.06
0.02 0.4 0.04
0.01 −0.6
0.2 0.02
0 0 −0.8 0
50 100 150 200 50 100 150 200 50 100 150 200 50 100 150 200
Fig. 1. Bates simulated data. True parameters (solid lines), WLS estimates (circles) and IEKF estimates (dots). and are changing over time.
spread of approximately $1, so we have added noise to the option prices, cf. Eq. (19), to mimic the mid-price uncertainty
present in the real world option prices. The noise is Gaussian with a standard deviation of one-fourth of the spread,
1
cf. R = cov() = 16 . The motivation behind this choice is that we expect at least 95% of the true option prices to be
within the spreads. Interpreting the error as a pure truncation error would give a slightly different R. The set of option
prices was split into an estimation set of 40 options per day and a validation set of 60 options per day. The options in
the different sets were selected at random each day.
Calibration can be seen as a mapping between, in our case, European calls and a set of latent processes. If the
validation is performed using the same mapping, e.g. by pricing out-of-sample calls, the interpolation capacity of the
calibrated pricing model is tested. If on the other hand the quality of the calibrated model is to be tested, a different
mapping is more suitable. A convenient choice in this simulation study is the European binary option, which can be
priced in the same framework as the European call as seen in Section 2.2. Prices of binary options are thus calculated
both for the true parameters and for the filtered states and compared using RMSE.
The time consumption has been measured by the average time consumption per step, i.e. the average time consumption
of daily recalibration. Since the time consumption is dependent of factors not specific to the actual estimation method,
e.g. the performance of the computer used, we are more interested in the ordering among the methods than the actual
numbers.
The parameters have been estimated using two different filters: EKF and IEKF. The state noise covariance matrix Q
was determined using quasi-maximum likelihood estimation as discussed in Section 3.2. We found that Qc = 5 × 10−3
and QJ = 10−4 to be optimal.
As a benchmark the parameters have also been estimated using standard and penalized WLS (PWLS), cf.
Section 3.1. The least squares estimators were initialized with the true parameter values. The PWLS used d1 = 1.5,
giving ˆ ≈ 10. We have used ˆ = 10 throughout the simulation study.
4.1. Results
The IEKF and WLS estimated parameters are depicted in Fig. 1 in addition to the true parameter values. It can be
seen that the IEKF estimates are closer to the true parameter values than the WLS estimates. We have not plotted the
2886 E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891
Table 2
Summary statistics for the simulated Bates model
Filter Set IS (%) RMSE calls RMSE parameters RMSE binary (10−5 ) Average time (s)
The RMSE for parameters are over all eight parameters. Bold face indicates the best method for that measure.
EKF and PWLS, these being intermediary in performance. WLS estimates far from the true values were moved to the
edge the subplots. This problem was mostly present in the estimates of the jump intensity and jump volatility as can
be seen in the figure. The IEKF (and EKF, PWLS) estimates are more stable over time compared to the WLS estimates,
and all the methods, possibly except WLS, estimate the latent process Vt very well. Even better, all methods are capable
of tracking time varying parameters with reasonable accuracy. This holds both for slowly varying parameters and
and for discontinuities in the parameter.
Table 2 displays the numerical results of the measured fit for the different estimators. As expected (defined) the WLS
have the best fit in sample. On the validation set the IEKF and the PWLS perform the best. The PWLS drops more in
performance on the validation set, indicating a slight overfitting, but still much better than the WLS. When measuring
the RMSE of the call prices for the different methods we again see the WLS performing best in sample and the IEKF
out of sample.
In terms of parameter RMSE, the filter methods clearly outperform the least squares methods, the WLS being several
orders of magnitude larger. This result is consistent with the overfitting we saw on the European call options and the
graphical results in Fig. 1. When considering European binary options, we again see a clear advantage for the IEKF,
while the EKF and PWLS being approximately equal. The RMSE for the WLS is much larger than for any of the other
methods.
In terms of speed, the EKF is the clear winner. It is our subjective belief from simulations using the Heston model and
Bates model that the WLS estimator suffers from the curse of dimensionality more than the filter methods. The larger
parameter space requires more function evaluation for the minimization routine, yielding increased computational time
or decreased precision. The estimation set in our simulation is quite large, and it is not unexpected that the interpolation
capacity is sufficiently tuned on the estimation set to also perform well on the validation set for the WLS. If the number
of options per day were smaller or varied, the WLS would be more prone to overfitting.
It can be observed that the filter does not seem to suffer badly (compared to the WLS and PWLS) from using only
Q-parameters instead of using Q- and P-parameters simultaneously. We are therefore using the same approximation
(V = 0) on real data.
5. Empirical study
We have used daily data on S&P 500 index options, from November 5th, 2001 to May 5th, 2003, thus excluding data
containing a short term effect following September 11th. The original data set consisted of 38 225 quotes (date, strike,
ask price, bid price, index level, time to maturity and risk free interest rate). The raw data were processed according to
the following rules:
1. All options having time to maturity less than 6 trading days were excluded. These options are illiquid and highly
sensitive to changes in any variable.
2. Options having bid–ask spread larger than 5 have been excluded, in order to eliminate illiquid options. Also op-
tions having zero or negative 0 bid–ask spread have been eliminated as these are an effect of non-synchronous
trading.
E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891 2887
3. The least liquid option when call and put options having identical strike and time to maturity is available has been
eliminated. It is well known that in-the-money options are traded infrequently compared to at-the-money or out-
of-the-money options. We have therefore excluded in-the-money options when at-the-money or out-of-the-money
options are available.
A total of 24 930 unique quotes remain after the processing. It is convenient to transform all quoted put options into
call options, using the put–call parity.
In practice r, the continuously compounded short rate (cf. Eq. (1)) is not constant and known. If we want to price
a claim at time t with maturity at time T, it is possible to replace r with the zero-coupon-bond (ZCB) yield over the
corresponding time period up to maturity (t, T ) and replace Q with the corresponding forward measure , FT instead.
A drawback with this approach is that we get different dynamics for the underlying stock for different maturities,
i.e. for each combination of t and T. This is of course due the fact that the ZCB yield varies with t and T. From a
valuation perspective this causes no serious problem, simply replacing r with the corresponding zero-coupon-yield and
set p(t, T ) equal to the ZCB price. The pricing formula Eq. (4) will look exactly the same.
Another adjustment is needed to account for dividends. It is common to approximate the dividend by a continuously
compounded dividend yield that depends on time and on time to maturity, hereafter denoted q(t, t + ). Ignoring
the dividend yield causes the transformed option prices to be discontinuous in the strike level. The dividend yield is
estimated such that the option prices are continuous in the strike level.
It is shown in e.g. Hull (2002) that dividends can be modelled by replacing the current value of the underlying equity
or index St by St e−q(t,t+) , when modelling a European call option maturing at t + . However, q(t, t + ) is not
quoted on the market, why we derive it from option quotes. It is well known from the put–call parity that
ct (Ki , ) − pt (Ki , ) = St e−q(t,t+) − Ki p(t, t + ). (38)
It is possible to estimate the corrected index level St e−q(t,t+) . Rearranging Eq. (38) and adding a noise term ei gives
St e−q(t,t+) = cKi , (t) − pKi , (t) + Ki p(t, t + ) + ei . (39)
The corrected index level can now be estimated by the sample mean, taken over overlapping mid-prices. The estimate
is stable over different Ki and reasonably stable over time.
5.1. Results
We have estimated three models on market data from options on the S&P 500 index, namely Heston, Bates and
NIG-CIR. Eighty percent of the processed data was used for estimation and the remainder for validation. The IEKF,
PWLS and WLS parameter estimates are presented in Figs. 2, 3 and 4. As in the simulation study, the axes are truncated
and some WLS estimates that obtained very large (absolute) values have been moved to the upper or lower edge of the
plots.
For the Heston model, being a sub-model of the Bates model, we used the state covariance found in the simulation
study. For the PWLS we set as given by Eq. (12) to 10, also inherited from the simulation study. A posterior numerical
calculation proved this a sensible choice. The result from the Heston estimation is depicted in Fig. 2. The WLS and
IEKF estimates (and PWLS) agree on , and Vt . However, the WLS is less robust on and V giving highly varying
estimates.
The Bates model was used in the simulation study on a set of data designed to be similar to the actual S&P 500 data.
We optimized with respect to Q and in the simulation and used the same values for the market data set. Also, was
checked posterior to have a reasonably correct value. The result from the Bates estimation is depicted in Fig. 3. For
the highly influential latent volatility process Vt , the WLS and IEKF agree, while on the other parameters, the WLS is
much more erratic, cf. the simulation study in Section 4. The PWLS and IEKF on the other hand, agree to a large extent.
It should be noted though that all parameters move a lot and are from being constant. For the jump intensity the WLS
sometimes gave extreme estimates (this has been truncated in the figure), something it counteracts by lowering both
the expected value and standard deviation . Also note the similarity with the Heston model of the volatility process
and its related parameters for the IEKF and PWLS. This, however, does not hold for the WLS.
The NIG-CIR model proved to be somewhat harder to find good parameter settings since we had no prior experience
from simulation study. We expect similar variation of the parameters in the NIG-CIR model as in the Bates model, and
2888 E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891
κ ξ σV
16 0.1 1.5
14 0.09
0.08
12
0.07 1
10 0.06
8 0.05
6 0.04
0.03 0.5
4
0.02
2 0.01
0 0 0
50 100 150 200 50 100 150 200 50 100 150 200
ρ Vt
−0.3 0.16
−0.35 0.14
−0.4
0.12
−0.45
−0.5 0.1
−0.55 0.08
−0.6 0.06
−0.65
0.04
−0.7
−0.75 0.02
−0.8 0
50 100 150 200 50 100 150 200
Fig. 2. Heston on market data. WLS (circles), PWLS (thin line) and IEKF estimates (thick line).
κ ξ σV ρ
12 0.1 1.8 −0.1
0 0 0 −1
2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr
Vt λ µ δ
0.14 3 0.15
0.12 0.05
2.5
0.1 0
2 0.1
0.08 −0.05
1.5
0.06 −0.1
1 0.05
0.04 −0.15
0.5
0.02 −0.2
0 0 −0.25 0
2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr
Fig. 3. Bates on market data. WLS (circles), PWLS (thin line) and IEKF estimates (thick line).
E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891 2889
κ ξ σ Vt
16 0.4 5 1
4.5 0.9
14 0.35
4 0.8
12 0.3
3.5 0.7
10 0.25
3 0.6
2 0.4
6 0.15
1.5 0.3
4 0.1
1 0.2
2 0.05
0.5 0.1
0 0 0 0
2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr
δ α β α/β
14 80 0 0
−2
12 70
−5 −4
60
10 −6
50 −8
−10
8
40 −10
6
−15 −12
30
−14
4
20
−20 −16
2 10 −18
0 0 −25 −20
2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr 2001−Dec 2002−May 2002−Oct 2003−Apr
Fig. 4. NIG-CIR on market data. WLS (circles), PWLS (thin line) and IEKF estimates (thick line).
Table 3
The Heston, Bates and NIG-CIR models have been calibrated to a number of S&P 500 data sets using the different estimation methods
The table contains the global fit error measurements for the calibrated model in sample (estimation set) and out of sample (validation set) for the
different estimation methods.
used the same settings. Other parameter values were also tested but no improvement was achieved. The result from
the NIG-CIR estimation is depicted in Fig. 4. Note that for the latter part of the data set, the WLS estimator has given
fairly unrealistic estimates, probably getting stuck in a local minimum. This is also seen in Table 3. Further, the ratio
/ , having a similar interpretation as the correlation in the Bates model, is fairly constant for the IEKF and PWLS,
but varies a lot for the WLS.
Global fit of the models has been measured in terms of inside spread (IS) and RMSE of the predicted prices in the
same way as in the simulation study. These results are presented in Table 3. No evaluation with respect to more exotic
contracts has been made.
2890 E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891
Since the validation set only consists of approximately 20% of the options, we expect the interpolation capacity
from the WLS estimates to be sufficiently tuned to also perform well out of sample, something that is confirmed in
Table 3. This is most evident in the simpler Heston model, while for the more complex models, evidence of overfitting
is seen. On the other hand we expect the IEKF to perform up to par with the WLS whilst still achieving stable parameter
estimates. This is confirmed by the parameter trajectories shown in Figs. 2–4 as well as the statistics in Table 3. As in
the simulation study, the IEKF performs best out of sample on all three models, indicating a well calibrated market
model and not just interpolation of the prices in the estimation set. It is worth noting that the WLS still overfits, as is
most evident in the Bates model where there is a substantial difference between in sample and out of sample IS.
6. Conclusions
The filter technique proposed in the paper has been evaluated on simulated and market option data. The filter estimates
have been compared to the standard WLS estimates as well as to estimates from a PWLS method. The overall impression
is that filter methods provide faster, more stable and more robust estimation than WLS.
When dealing with market quoted prices, the use of mid-price introduces a risk for overfitting. We find abundant
evidence of overfitting in the WLS estimates. The limitation of the standard WLS implementation is that it only uses
data from the current observation, whereas the filter methods make optimal use of past and current observations. An
intermediary approach to reduce overfitting is the PWLS, but this lacks the optimal weighting of past and current
information present in the filter.
We performed a simulation study on the Bates model to evaluate the different methods under controlled circumstances.
The WLS achieved a high in sample fit, with the PWLS and IEKF being close behind. The ordering changed out of
sample with the IEKF being better than the WLS. This indicates that the WLS overfitted to data. The IEKF on the other
hand managed to find the actual parameter values, and therefore excelling on the other metrics, especially when pricing
binary options. Even the EKF manages to track the parameters well but the inferior linearization means that it is not
capable of fine-tuning the more influential parameters such as the volatility to the extent that the IEKF does. Also, the
amount of time required by the different methods makes IEKF stand out even more.
A similar study was performed on S&P 500 option data. The pattern from the simulation study was repeated. The
filter methods managed to provide stable estimates across models while the WLS is more erratic and has a poorer
performance on the complex models. In fact we experienced some convergence issues for the WLS on the NIG-CIR
model. We can only speculate that the parameter estimates from the IEKF and PWLS represent the evolution some
actual market model, something that needs to be checked using a different mapping than European calls similar to the
binary calls in the simulation study.
These results provide evidence that the self-organizing state-space model, combined with an IEKF is a promising
method for option calibration. The IEKF is shown to be both faster and more accurate than standard methods when
applied to different market models. We believe that a similar behaviour can be expected from the filter for other models
and other data sets.
Acknowledgements
We would like to thank an anonymous referee and the guest editor for helpful comments and suggestions. Erik
Lindström gratefully acknowledges financial support from the Bank of Sweden Tercentenary Foundation under Grant
P2005-0712, Jonas Ströjby gratefully acknowledges financial support from the Swedish Research Council under Grant
2005-2799 and Magnus Wiktorsson gratefully acknowledges financial support from the Swedish Research Council
under Grant 2002-5415.
References
Abramowitz, M., Stegun, I.A. (Eds.), 1972. Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables, 9th printing.
Dover, New York.
Anderson, B.D.O., Moore, J.B., 1979. Optimal Filtering. Prentice-Hall, Englewood Cliffs, NJ.
Bakshi, G., Carr, P., Wu, L., 2007. Stochastic risk premiums: stochastic skewness in currency options, and stochastic discount factors in international
economics. J. Financial Econom., in press, doi:10.1016/j.jfineco.2006.12.001.
Bakshi, G.S., Cao, C., Chen, Z., 1997. Empirical performance of alternative option pricing models. J. Finance 52 (5), 2003–2049.
E. Lindström et al. / Computational Statistics & Data Analysis 52 (2008) 2877 – 2891 2891
Barndorff-Nielsen, O.E., 1997. Processes of Normal inverse Gaussian type. Finance and Stochastics 2, 41–68.
Barndorff-Nielsen, O.E., Shepard, N., 2001. Non-Gaussian Ornstein–Uhlenbeck-based models and some of their uses in financial economics.
J. Roy. Statist. Soc. Ser. B 63, 167–241.
Barndorff-Nielsen, O.E., Shepard, N., 2002. Econometric analysis of realized volatility and its use in estimating stochastic volatility models. J. Roy.
Statist. Soc. Ser. B 64, 253–280.
Bates, D., 1996. Jumps and stochastic volatility: the exchange rate processes implicit in Deutchemark options. Rev. Financial Stud. 9, 69–107.
Black, F., Scholes, M., 1973. The pricing of options and corporate liabilities. J. Political Economy 81, 637–659.
Campbell, J.Y., Lo, A.W., MacKinley, A.C., 1997. The Econometrics of Financial Markets. Princeton University Press, Princeton, NJ.
Carr, P., Madan, D., 1999. Option valuation using the fast Fourier transform. J. Comput. Finance 2 (4), 61–73.
Carr, P., Wu, L., 2003. The finite moment log stable process and option pricing. J. Finance 58, 753–778.
Carr, P., Wu, L., 2004. Time-changed Lévy processes and option pricing. J. Financial Econom. 71, 113–141.
Carr, P., Wu, L., 2007. Stochastic skew in currency options. J. Financial Econom., in press, doi:10.1016/j.jfineco.2006.03.010.
Carr, P., Geman, H., Madan, D., Yor, M., 2002. The fine structure of asset returns: an empirical investigation. J. Business 75, 305–332.
Carr, P., Geman, H., Madan, D., Wu, L., Yor, M., 2003a. Option pricing using integral transforms. A slideshow presentation available at:
http://www.math.nyu.edu/research/carrp/papers/pdf/integtransform.pdf .
Carr, P., Geman, H., Madan, D., Yor, M., 2003b. Stochastic volatility for Lévy processes. Math. Finance 13, 345–382.
Chernov, M., Gallant, A.R., Ghysels, E., Tauchen, G., 2004. Alternative models for stock price dynamics. J. Econometrics 116, 225–257.
Cont, R., Tankov, P., 2004. Financial Modelling with Jump Processes. Chapman & Hall, London.
Derman, E., Kani, I., 1994. Riding on a smile. Risk 7 (2), 32–39.
Doucet, A., de Freitas, N., Gordon, N. (Eds.), 2001. Sequential Monte Carlo Methods in Practice. Springer, New York.
Draper, N.R., Smith, H., 1998. Applied Regression Analysis. third ed.. Wiley, New York.
Duffie, D., Pan, J., Singleton, K., 2000. Transform analysis and asset pricing for affine jump diffusions. Econometrica 68, 1343–1376.
Dupire, B., 1994. Pricing with a smile. Risk 7 (1), 18–20.
Heston, S.L., 1993. A closed-form solution for options with stochastic volatility with applications to bond and currency options. Rev. Financial Stud.
6, 327–343.
Huang, J.-Z., Wu, L., 2004. Specification analysis of option pricing models based on time-changed Lévy Processes. J. Finance 59, 1405–1440.
Hull, J.C., 2002. Options, Futures, and Other Derivative Securities. fifth ed. Prentice-Hall, Englewood Cliffs, NJ.
Hull, J.C., White, A., 1987. The pricing of options on assets with stochastic volatility. J. Finance 42 (2), 281–300.
Kalman, R.E., 1960. A new approach to linear filtering and prediction problems. Trans. ASME Ser. D J. Basic Eng. 82, 35–45.
Kalman, R.E., Bucy, R.S., 1961. New results in linear filtering and prediction. Trans. ASME Ser. D J. Basic Eng. 83, 95–107.
Kitigawa, G., 1998. A self-organizing state-space-model. J. Amer. Statist. Assoc. 93 (443), 1203–1215.
Künsch, H.R., 2001. State space and hidden Markov models. In: Barndorff-Nielsen, O.E., Cox, D.R., Klüppelberg, C. (Eds.), Complex Stochastic
Systems, Monographs on Statistics and Applied Probability, vol. 87. Chapman & Hall, New York, pp. 109–173.
Lee, R.W., 2004. Option pricing by transform methods: extensions, unification, and error control. J. Comput. Finance 7, 51–86, an augmented version
is available at: http://www.math.uchicago.edu/∼rl/dft.pdf .
Lefebvre, T., Bruyninckx, H., De Schutter, J., 2004. Kalman filters for non-linear systems: a comparison of performance. Internat. J. Control 77 (7),
639–653.
Lindström, E., Ströjby, J., Brodén, M., Wiktorsson, M., Holst, J., 2006. Sequential calibration of options. Technical Report 21, Centre for Mathematical
Sciences, Lund University.
Madan, D.B., Seneta, E., 1990. The variance Gamma (VG) model for share market returns. J. Business 63, 511–524.
Maybeck, P., 1982. Stochastic Models, Estimation and Control, vols. 1–3. Academic Press, New York.
Merton, R.C., 1976. Option pricing when the underlying stock returns are discontinuous. J. Financial Econom. 5, 125–144.
Polson, N.G., Stroud, J.R., 2003. Bayesian inference for derivative prices. Bayesian Statist. 7, 641–650.
Schoutens, W., Simons, E., Tistaert, J., 2004. A perfect calibration! Now what? Wilmott Magazine, March.