Data Analysis and Applications 4
Data Analysis and Applications 4
Volume 6
Edited by
Andreas Makrides
Alex Karagrigoriou
Christos H. Skiadas
First published 2020 in Great Britain and the United States by ISTE Ltd and John Wiley & Sons, Inc.
Apart from any fair dealing for the purposes of research or private study, or criticism or review, as
permitted under the Copyright, Designs and Patents Act 1988, this publication may only be reproduced,
stored or transmitted, in any form or by any means, with the prior permission in writing of the publishers,
or in the case of reprographic reproduction in accordance with the terms and licenses issued by the
CLA. Enquiries concerning reproduction outside these terms should be sent to the publishers at the
undermentioned address:
www.iste.co.uk www.wiley.com
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
Preface
We wish to thank all the authors for their insights and excellent
contributions to this book. We would like to acknowledge the assistance of
all involved in the reviewing process of the book, without whose support this
could not have been successfully completed. Finally, we wish to express our
thanks to the secretariat and, of course, the publishers. It was a great pleasure
to work with them in bringing to life this collective volume.
Andreas MAKRIDES
Rouen, France
Alex KARAGRIGORIOU
Samos, Greece
Christos H. SKIADAS
Athens, Greece
January 2020
PART 1
After extensive investigation on the statistical properties of financial returns, a discrete nature
has surfaced when low price effect is present. This is rather logical since every market operates
on a specific accuracy. In order for our models to take into consideration this discrete nature of
returns, the discretization of the tail density function is applied. As a result of this discretization
process, it is now possible to improve the percentage value at risk (PVaR) and expected
percentage shortfall (EPS) estimations on which we are focusing in this work. Finally, in order to
evaluate the improvement provided by our proposed methodology, adjusted evaluation measures
are presented, capable of evaluating percentile estimations like PVaR. These adjusted evaluation
measures are not only limited to evaluating percentiles, but in any scenario where data does not
bare the same amount of information and consequently, does not all carry the same degree of
importance, like in the case of risk analysis.
1.1. Introduction
that the asset returns are preferred over prices due to their stationarity
[MEU 09, SHI 99]. Moreover, the volatility of returns is the one that can be
successfully forecasted and that is essential for risk management
[POO 03, AND 01]).
The increase in complexity of both the financial system and the nature of the
financial risk over the last decades, results in models with limited reliability.
Hence, in extreme economic events, such models become less reliable and in
some instances, fully fail to measure the underlying risk. Consequently, they
are producing inaccurate risk measurements which has a great impact on many
financial applications, such as asset allocation and portfolio management in
general, derivatives pricing, risk management, economic capital and financial
stability (based on Basel III accords).
To make thinks even worse, stock exchange markets (and in general every
financial market) operate with certain accuracy. In most European markets the
accuracy is 0.1 cents (0.001 euros), while US markets operate with 1 cent
(USD 0.01) accuracy except for securities that are priced less than USD 1.00
for which the market accuracy or minimum price variation (MPV) is USD
0.0001. Despite the readjustment for extremely low price assets, the
associated fluctuation (variation) is considerable and the corresponding
volatility is automatically increased. The phenomenon is magnified
considerably in periods of extreme economic events (economic collapses,
bankruptcies, depressions, etc.) and as a result typical market accuracy fails to
handle smoothly assets of extremely low price.
In order to answer all the above, we discuss, in this chapter, the concept of
low price effect (lpe) [FRI 36] and recommend a low price correction (lpc) for
improved forecasts. The low price effect is the increase in variation for stocks
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 5
with low price due to the existence of a minimum possible return produced
when the asset price changes by the MPV. Lpe is frequently overlooked and
the main reason for that is the lack of theoretical background related to the
reasons resulting in this phenomenon, which surfaces primarily in periods of
economic instability or extreme economic events. The pioneering of the
proposed correction is that it does not require any additional parameters and
takes into account the asset price. The proposed correction is associated with
the rationalization of the estimated asset returns, since it is rounded to the
next integer multiple of the minimum possible return. Except for the proposed
correction, in this work we also provide a mathematical reasoning for the
increase in volatility.
Inspired from the above, we came to the same conclusion as many before,
that in risk analysis the returns of an asset do not all bare the same amount of
information and do not all carry the same degree of significance
[SOK 09, ASA 17, ALH 08, ALA 10]. In the absence of a formal
mathematical notion, the term asymmetry in the importance describes
relatively satisfactory the above phenomenon which is also apparent in other
scientific areas, related to medical, epidemiological, climatological,
geophysical or meteorological phenomenon. For a mathematical
interpretation, we may consider a proper cost function that takes different
values, depending on different regions of the dataset or the entire period of
observation. For instance, in biosurveillance, the importance (and the cost)
associated with an illness incidence rate is much higher in the so-called
epidemic periods associated with an extreme rate increase. In the case of risk
analysis, risk measures like the value at risk (VaR) and expected shortfall
(ES), concentrate only on the left tail of the distribution of returns with the
attention being paid to the days of violation rather than to those of no
violation. Consequently, failures in fitting a model on the right tail are not
considered to be important. Thus, the asymmetry in the importance of
information is crucial in both choosing the most appropriate model by
assigning a proper weight to each region of the dataset and evaluating its
forecasting ability.
The rules that govern the accuracy of financial markets around the world
fail to smoothly handle assets of extremely low price resulting in a
considerable fluctuation (variation) and increased volatility. As expected, the
phenomenon is magnified considerably in periods of extreme economic
events (economic collapses, bankruptcies, depressions, etc.). Indeed, since all
possible (logarithmic) returns on a specific day are integer multiples of the
minimum possible return, the stock movement will fluctuate more nervously,
the lower the prices. As a result, violations will occur more frequently and
forecasts will turn out to be irrational in the sense that such returns cannot be
materialized. The resulting volatility increase is quite often overlooked,
primarily due to the fact that we take into account only the returns of an asset,
neglecting entirely the prices.
D EFINITION 1.1.– Let pt be the asset value at time t and c(pt ) be the
minimum price variation (market accuracy) associated with the value of the
asset at time t. Then the minimum possible return (mpr) of an asset at time t
denoted by mprt is the logarithmic return that the asset will produce if its
value changes by c(pt ) and is given by:
pt + c(pt )
mprt = log .
pt
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 7
Note that mprt is the same for both upward and downward movements due
to the symmetry of logarithmic returns. For the special case, that a market has
a constant accuracy, say c, irrespectively of the stock price, Definition 1 can be
simplified to mprt = log ((pt + c)/pt ) (see [SIO 17]).
E XAMPLE 1.1.– Let us assume that the value of an asset is equal to e0.19
and the market operates under e0.001 accuracy. Then the minimum possible
return for the asset is 0.5%. Also, all possible (logarithmic) returns for the
asset in this specific day are the integer multiples of the minimum possible
return. This will have as a result, the stock movement to become even more
nervous and models’ failures to increase. Consequently, PVaR violations will
occur more frequently and our model will almost always derive forecasts that
are irrational since the stock cannot produce such returns.
Now that we have defined the mpr, we can provide a strict definition of lpe
as well as the mathematical reasoning behind this.
D EFINITION 1.2.– Low price effect (lpe) is the inevitable increase of variance
in stocks with low prices due to the existence of a minimum possible return.
(k+1)·mpr t
+∞
ˆ
= (rt − μ)2 f (rt ) drt
k=−∞ k·mprt
(k+1)·mpr t
+∞
ˆ
≤ sup (rt − μ)2 f (rt ) drt
k=−∞ rt ∈[k·mprt ,(k+1)·mprt ]
k·mprt
(k+1)·mpr t
+∞
ˆ
2
= sup (rt − μ) f (rt ) drt
k=−∞ rt ∈[k·mprt ,(k+1)·mprt ] k·mprt
8 Data Analysis and Applications 4
where Rt is the random variable of returns at time t and f (·), μ are the density
and the mean of returns, respectively.
R EMARK 1.1.– It should be noted that in JSE, the lpe was present for shares
priced below 30 cents, but the performance was not equally good for “super-
low priced shares” (0–19 cents). For WSE, the analysis was based on the 30%
of stocks with the lowest prices.
D EFINITION 1.3.– Low price effect area is the range of prices for which the
mpr is greater than a pre-specified threshold Θ.
It must be noted that the minimum possible return is a direct result of the
combination of a low price with the minimum possible variation for the stock.
E XAMPLE 1.2.– We present below two examples for the low price effect area,
which is easily defined by applying Definition 1 for specific values of the
accuracy and the threshold.
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 9
Thus, for 1 ≤ pt ≤ 10 and for pt ≤ 0.100 the low price effect is present,
and proper actions should be taken to minimize it. While in the range of
1–10 dollars, numerous real-life examples can be found, the same is not true
for prices under 0.1 dollars, in which case, the stock has already been halted
from the stock market, and hence, lpe makes only theoretical sense and
becomes a theoretical concept.
For the resulting low price effect area, the models considered should be
appropriately adapted. This adaptation is done through the so-called low price
correction of both the estimation of percentage value at risk denoted by PVaR
and the expected percentage shortfall denoted by EPS to be introduced
respectively in the following two sections.
The percentage value at risk which is a risk measure for estimating the
possible percentage losses from trading assets, within a set time period, is
defined as follows:
The probability of a PVaRt (p) violation, namely, the p-th percentile of the
distribution, is given by
ˆ −P V aRt (p)
p = Pr[Rt ≤ −P V aRt (p)] = f (x)dx.
−∞
where Rt is the random variable of returns at time t, and f (·) is the probability
density function of returns.
Note that for the evaluation of PVaRt (p), we may consider any
econometric model and then apply an estimation technique for the model
parameters. Consider, for instance, the general Asymmetric Power ARCH
(APARCH) model [DIN 93]:
Rt = σt εt
with
p
q
δ
σtδ = a0 + ai (|εt−i | − γi εt−i ) + δ
bj σt−j ,
i=1 j=1
P V aRt = qp (F )σ̂t ,
For the distribution F , usual choices are the normal, the Student-t and the
skewed Student-t distribution [LAM 01]. In case where the standardized
returns are generated from a Student-t distribution with ν degrees of freedom,
the variance is equal to ν/(ν − 2); hence, it is never equal to 1. If that sample
variance is used in the calculation of PVaR, the PVaR would be
overestimated. Volatility effectively shows up twice, both in F −1 (p) and in
the estimation σ̂ of σ, which is obtained by numerical maximization of the
ν
log-likelihood. Hence, we need to scale volatility as follows: σ 2 ≡ ν−2 σ̃ 2 ,
where σ̃ 2 is the variance in excess of that implied by the standard Student-t.
We conclude this section with the definition of the low price correction
for the percentage value at risk. A PVaRt (p) estimate can take any real value
not necessarily equal to integer multiples of mprt . Under the low price effect,
asset movements become inevitably more nervous, and any continuous model
used would produce forecasts that are irrational in the sense that assets cannot
produce such returns, more often than it should. To resolve this “inconsistency”
we propose the low price correction by rounding the PVaRt (p) estimate to the
closest legitimate value, namely, the next integer multiple of the mprt .
12 Data Analysis and Applications 4
D EFINITION 1.5.– Let P V aRt (p) be the estimation of the PVaR on day t for
a specific asset. The low price correction of the estimation denoted by
PV aRt (p) is given by:
P V aR (p)
t
+ 1 · mprt , if mprt ≥ Θ
P VaRt (p) = mprt
[1.1]
P V aRt (p), if mprt < Θ
Note that we prefer to deal with the percentage value at risk for comparative
reasons between assets of the same portfolio with different allocations. We
observe that under the low price correction, the market’s accuracy is passed on
to the evaluation of the percentage value at risk, resulting in a more reasonable
number of violations.
After we have obtained the PVaRt (p) in the previous section, we now
calculate the conditional expectation under PVaRt (p), which is given by:
We observe that the area under f (·) in the interval [−∞, −P V aRt (p)] is
less than 1, implying that f (·) is not a proper density function any more. This
can be resolved by defining the tail density (right-truncated) function
fP V aR (·), obtained by truncation on the right, so that the area below this
density becomes exactly equal to 1. Thus:
1 −P V aRt (p)
ˆ ˆ −P V aRt (p)
1= f (x)dx = fP V aR (x)dx.
p −∞ −∞
R EMARK 1.2.– Even though, on the previous definition we call DEP St the
discrete approximation of EP St , the truth is that the nature of f and by
extension fP V aR , is discrete, since there always exists a minimum possible
return. We may treat returns as a continuous random variable when mpr is
extremely small, but the discrete nature of returns still exists.
x = −PV
aRt (p) − mprt , −P V aRt (p) − 2mprt , . . . , −P V aRt (p) −
mini≤t Ri
mprt + 1 · mprt , where Ri are the realizations of returns.
For the evaluation of the performance of the competing models, we will use
statistical methods, as well as backtesting. Popular evaluation measures used
in the literature include the mean square error (MSE), the mean absolute error
(MAE), and the mean absolute percent error (MAPE). Since the main interest
lies in returns, which mostly (except for a very few extreme cases) take values
in (−0.5, 0.5), we prefer MAE and MAPE, because the square in MSE will
decrease further the errors. These measures should be appropriately adapted in
order to capture the needs of the problem at hand.
T
Let Rt=1 be a sample of a time series corresponding to daily logarithmic
losses on a trading portfolio, T the length of the time series and PVaRt the
estimation of PVaR on day t. If on a particular day, the logarithmic loss exceeds
the PVaR forecast, then the PVaR limit is said to have been violated. For a given
PVaRt we define the indicator ηt as follows:
1 if Rt > P V aRt
ηt = .
0 if Rt ≤ P V aRt
Under the above setting, it is easily seen, that the MSE of violation days is
defined as follows:
1
M SE = (Rt − P V aRt (p))2 .
η
t≤T t t≤T &ηt =1
We can easily observe that the above is a special weighted mean squared
error expression. Indeed, we note that the above can be written as
ηt
M SE = · (Rt − P V aRt (p))2
v(T )
t≤T
The comparison between the observed frequency and the expected number
of violations provides the primary tool for backtesting, which is known as the
violation ratio [CAM 06]. VR can be used to evaluate the forecast ability of the
model, PVaR estimations, and normalized shortfall (NS) for backtesting EPS
estimations.
Note that if the corrected version of PVaR is used, then PVaRt should be
replaced by P V aRt in (1.2.3). A PVaR violation is said to have occurred
whenever the indicator ηt is equal to 1.
where EP St is the observed EPS on day t. From the definition of EPS, the
expected return, given that PVaR is violated, is:
With EPS, we are testing whether the mean of returns on days when PVaR is
violated is the same as the expected EPS on these days. As it is clear, it is much
harder to create a formal test in order to ascertain, whether normalized EPS
equals to 1 or not. Such a test would have to simultaneously test the accuracy of
PVaR and the expectation beyond PVaR. This means that the reliability of any
EPS backtest procedure is likely to be much lower than that of PVaR backtest
procedures.
R EMARK 1.6.– The importance of the proposed technique lies on the fact
that under the low price correction, fewer violations are expected to occur.
Note that a VaR estimate can take any real value not necessarily equal to
integer multiples of mpr since it is not controlled by the market’s accuracy.
This observation simply implies that any model almost always derives
forecasts that are irrational in the sense that stocks cannot produce such
returns. Thus, due to this “inconsistency” between PVaR forecasts and stock
movements, PVaR violations occur more frequently than they should.
Indeed, assume mprt is greater that the threshold Θ. Then, the PVaR
estimate is plausible, as the asset’s (logarithmic) return, as long as it coincides
with an integer multiple of mpr. Otherwise, if k · mprt < P V aRt <
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 17
1.3. Application
R EMARK 1.7.– For less extreme examples, the reader may refer to examples
for the Athens Exchange (ATHEX) and the American Stock Exchange
(NYSE MKT) analyzed and discussed in <http://actuarweb.aegean.gr/
labstada/publications.html> [SIO 19a, SIO 19b].
Warrants are in many ways similar to options, but with a few key
differences distinguishing them. In our case, the warrants of Alpha Bank were
issued in 2013 and they could be exercised biannually over the next 4.5 years.
The exercise price of each warrant was set to be equal to the offer price of a
18 Data Analysis and Applications 4
new share, with an interest rate of 3% plus a spread increasing by 100 bps per
annum, due to the first recapitalization of the Greek banking system, and were
traded on the Athens Exchange (ATHEX). Since later on Alpha’s stock price
collapsed, warrants’ price also collapsed (see Figures 1.1 and 1.2).
For the distribution of returns, the Q-Q plots for the Normal and Student-t
with 3 and 5 degrees of freedom (df) are also furnished (Figures 1.3–1.5). The
plots clearly show that the underlying distribution is not normal, since the tails
do not fit.
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 19
We first implement below the low price correction (lpc) on the percentage
value at risk (PVaR) methodology given in [1.1]. Then, the backtesting results
for various models considered (EWMA, normal, historical simulations (HS)
20 Data Analysis and Applications 4
Table 1.1 gives the violation ratios (VRs) for PVaR estimations for four
different models, with and without low price correction implemented, as well
as, the PVaR volatilities. Based on the results on PVaR estimations of
Table 1.1 and the fact that the acceptable values of VR (according to Basel III
accords) lie in (0.8,1.2), all models fail with HS failing the most among all
models considered. The estimation of all models systematically underestimate
the underlying risk. Observe that the proposed methodology of the low price
correction significantly improves the violation ratio in all models, which is
also evident from Figures 1.6 and 1.7. These specific figures show the
improvement of the proposed methodology and are provided for a visual
understanding of the contribution of this work. We must note here that this
specific example was chosen for its extreme behavior. Our methodology
succeed to provide an improved result in all cases, even in this extreme case.
The predictive ability of all models used are presented in Table 1.2, which
contains evaluation measures only in the final days of violations (the days of
violations of the corrected model).
The most important results by far are the ones from the adjusted evaluation
measure, which completely captures the genuine improvement in our
estimations with the help of lpc, in Table 1.2. In this table, the improvement is
genuine for all models for all measures, as it was supposed to be by the
definition of lpc. The adjusted evaluation measure of Table 1.2 not only
completely agrees with the violation ratios of Table 1.1, but also quantifies
something that from the definition of lpc was clear, that the improvement for
the risk analysis point of view is genuine.
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 21
Figure 1.6. Backtesting results of PVaR without low price correction for
Alpha warrants. For a color version of this figure, see
www.iste.co.uk/makrides/data4.zip
22 Data Analysis and Applications 4
Figure 1.7. Backtesting results of PVaR with low price correction for
Alpha warrants. For a color version of this figure, see
www.iste.co.uk/makrides/data4.zip
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 23
Table 1.2. Evaluation measures results for PVaR with and without low
price correction of Alpha warrants
R EMARK 1.8.– Note that in addition to the Violation ratio, which is equivalent
to the well-known Kupiec test [KUP 95], other backtesting procedures could
also be applied (see for example [CHR 98]).
Decreasing the error on these days, translates to decreasing the loss at days
that loss exceeds what we are prepared for. Hence, this clear improvement is
incredible important. According to these results, we conclude that low price
correction significantly improves the estimation of risk for all models. This
decrease was expected, since the methodology of low price correction always
provides improved estimations as well as “logical” ones, for every model.
Finally, we have backtested expected percentage shortfall (EPS) with the
method of normalized shortfall (NS), i.e. with and without lpe taken into
account. For comparative purposes, Table 1.3 contains the VRs of the PVaR
estimations, as well as, NSs of the EPS estimations. Backtesting results
clearly show the improvement by the implementation of the low price
correction. Although the number of violations has decreased significantly, as
shown in the decrease of the VRs of the PVaR estimations, the NS of EPS
seems to be slightly increased. This is due to the fact that the violations that
still occurred after the low price correction were the most extreme ones, and
hence, the increase in EPS is expected. This increase is minor and as a result,
we succeeded in retaining the NS of EPS in an acceptable range around 1,
while significantly improving the VR of PVaR.
24 Data Analysis and Applications 4
VR without lpc VR with lpc PVaR vol without lpc PVaR vol with lpc
EWMA 0.7308161 0.6090134 0.02700116 0.02697531
Normal 0.5115713 0.4750305 0.006328121 0.006624875
HS 0.8282582 0.7673569 0.005578643 0.006021879
GARCH(1,1) 0.8038977 0.7429963 0.02289559 0.02289522
GARCH(2,2) 0.8404385 0.7795371 0.02284879 0.02285167
APARCH(1,1) 0.8891596 0.7795371 0.02066832 0.02072566
As before, the results are the same. We have a genuine improvement in all
cases. Even though in this case we have an overestimation of risk and as a
result getting further away from values in the (0.8,1.2) range, these values are
more defensive ones, a fact which in combination with the results of Table 1.5
will confirm the usefulness of the proposed methodology. The performance of
the lpc methodology according to accuracy measures is presented in Table 1.5,
which contains evaluation measures only in the final days of violations (the
days of violations of the corrected model).
Table 1.5. Standard evaluation measures results for PVaR with and
without low price correction of ARTX
The results in Table 1.6 verify the genuine improvement in all cases.
Without lpc With lpc
HS PVaR 1.061477 1.092811
HS EPS 1.061477 1.011258
1.4. Conclusion
are achieved via improved volatility estimations that were derived by taking
into consideration the asset prices. The method turns out to be both simple
and straightforward. EPS and PVaR estimations after the implementation of
the low price correction either remain unaffected or are improved.
Our methodology that pays attention not only to the asset return but also to
the asset price provides sufficient evidence that prices could contain important
information which could, if taken under consideration, result in improved
forecasts of risk estimation. Last but not least, we also provide the solution to
the problem. The asset under investigation must be reverse split (also known
as a stock merge) if possible in order to correct the above-mentioned
problems. Apart from any theoretical considerations, the practical
implications that the low price correction entails are the key issues in the
contributions of the present work.
easily extended to other scientific areas where data do not necessarily carry
the same amount of information, such as in physical (earthquakes, floods,
etc.) or climatological (heat or cold waves) phenomena.
1.5. Acknowledgements
This work was completed as part of the research activities of the Laboratory
of Statistics and Data Analysis of the University of the Aegean.
1.6. References
[ALA 10] A LADAG C.H., E GRIOGLU E., G UNAY S. et al., “Improving weighted information
criterion by using optimization”, Journal of Computational and Applied Mathematics,
vol. 233, no. 10, pp. 2683–2687, Elsevier, 2010.
[ALH 08] A L -H AWAMDEH S., Knowledge Management: Competencies and Professionalism,
vol. 7, World Scientific, 2008.
[AND 01] A NDERSEN T.G., B OLLERSLEV T., D IEBOLD F.X. et al., “The distribution of
realized stock return volatility”, Journal of Financial Economics, vol. 61, no. 1, pp. 43–76,
Elsevier, 2001.
Forecasting Methods in Extreme Scenarios and Advanced Data Analytics 31
[ASA 17] A SADABADI M.R., S ABERI M., C HANG E., “Targets of unequal importance using
the concept of stratification in a big data environment”, International Journal of Fuzzy
Systems, pp. 1–12, Springer, 2017.
[BOL 86] B OLLERSLEV T., “Generalized autoregressive conditional heteroskedasticity”,
Journal of Econometrics, vol. 31, no. 3, pp. 307–327, Elsevier, 1986.
[BRA 16] B RAIOME M., S CHOLTES N.K., “Forecasting value-at-risk under different
distributional assumptions”, Econometrics, vol. 4, no. 1, pp. 1–27, MDPI, 2016.
[BRO 11] B RODA S.A., PAOLELLA M.S., “Expected shortfall for distributions in finance”,
Statistical Tools for Finance and Insurance, pp. 57–99, Springer, 2011.
[CAM 06] C AMPBELL S.D., “A review of backtesting and backtesting procedures”, The
Journal of Risk, vol. 9, no. 2, Page 1, Incisive Media Limited, 2006.
[CHR 82] C HRISTIE A.A., “The stochastic behavior of common stock variances: Value,
leverage and interest rate effects”, Journal of Financial Economics, vol. 10, pp. 407–432,
Elsevier, 1982.
[CHR 98] C HRISTOFFERSEN P.F., “Evaluating interval forecasts”, International Economic
Review, pp. 841–862, JSTOR, 1998.
[CLE 51] C LENTERIN J.C., “Quality versus price as factors influencing common stock price
fluctuations”, Journal of Finance, vol. 6, no. 4, pp. 398–405, Elsevier, 1951.
[DAN 11] DANIELSSON J., Financial Risk Forecasting, Wiley, 2011.
[DES 97] D ESAI H., JAIN P.C., “Long-run common stock returns following stock splits and
reverse splits”, Journal of Business, vol. 70, pp. 409–433, Elsevier, 1997.
[DIN 93] D ING Z., G RANGER C.W.J., E NGLE R.F., “A long memory property of stock
market returns and a new model”, Journal of Empirical Finance, vol. 1, pp. 83–106, Taylor
& Francis, 1993.
[FRI 36] F RITZMEIER N.H., “Relative price fluctuations of industrial stocks in different price
groups”, Journal of Business, vol. 9, no. 2, pp. 133–154, Elsevier, 1936.
[GIL 82] G ILBERTSON R.A.C., A FFECK -G RAVES J.F., M ONEY A.H., “Trading in low
priced shares: An empirical investigation 1968–1979”, Investment Analysts Journal,
vol. 11, no. 19, pp. 21–29, Elsevier, 1982.
[HWA 08] H WANG S., L U C., “Is share price relevant?. Available at: http://papers.ssrn.com/
sol3/papers.cfm?abstract_id=1341790”, Elsevier, 2008.
[KUP 95] K UPIEC P., “Techniques for verifying the accuracy of risk measurement models”,
The Journal of Derivatives, vol. 3, no. 2, pp. 73–84, 1995.
[LAM 01] L AMBERT P., L AURENT S., Modelling financial time series using GARCH-type
models with a skewed student distribution for the innovations (Discussion Paper No. 0125),
Institut de Statistique, Université Catholique de Louvain, Louvain-la-Neuve, Belgium,
Taylor & Francis, 2001.
[MEU 09] M EUCCI A., Risk and Asset Allocation, Springer Science & Business Media, 2009.
32 Data Analysis and Applications 4
[POO 03] P OON S.-H., G RANGER C.W., “Forecasting volatility in financial markets: A
review”, Journal of Economic Literature, vol. 41, no. 2, pp. 478–539, American Economic
Association, 2003.
[SHI 99] S HIRYAEV A.N., Essentials of Stochastic Finance: Facts, Models, Theory, vol. 3,
World Scientific, 1999.
[SIO 17] S IOURIS G.-J., K ARAGRIGORIOU A., “A low price correction for improved
volatility estimation and forecasting”, Risks, vol. 5, no. 3, Page 45, Multidisciplinary Digital
Publishing Institute, 2017.
[SIO 19a] S IOURIS G.-J., S KILOGIANNI D., K ARAGRIGORIOU A., “Adjusted evaluation
measures for asymmetrically important data”, Econometric Research in Finance, vol. 4,
no. 1, pp. 41–66, 2019.
[SIO 19b] S IOURIS G.-J., S KILOGIANNI D., K ARAGRIGORIOU A., “Post model correction
in value at risk and expected shortfall”, International Journal of Mathematics, Engineering
and Management Sciences, vol. 4, no. 3, pp. 542–566, 2019.
[SOK 09] S OKOLOVA M., L APALME G., “A systematic analysis of performance measures for
classification tasks”, Information Processing & Management, vol. 45, no. 4, pp. 427–437,
Elsevier, 2009.
[WAE 97] WAELKENS K., WARD M., “The low price effect on the Johannesburg Stock
Exchange”, Investment Analysts Journal, vol. 26, no. 45, pp. 35–48, Taylor & Francis,
1997.
[ZAR 14] Z AREMBA A., Ż MUDZI ŃSKI R., “Low price effect on the Polish Market”,
e-Finanse: Financial Internet Quarterly, vol. 10, no. 1, pp. 69–85, 2014.
2
To measure credit risks using a structural modeling approach, one-factor Merton models are
often employed. These assume a common single risk factor for all counterparties. We extend
these models so that non-Gaussian and serially correlated asset returns are used for both the
single common risk and the obligor’s idiosyncratic factors. By using a standard Edgeworth
expansion, we derive the approximate default rate distributions. Maximum likelihood estimators
are introduced to estimate model parameters, and some credit risk measures are evaluated. Our
empirical results illustrate that the proposed non-Gaussian models offer better credit risk
evaluations compared with those obtained by Vasicek models. Applications of this model to CDO
pricing are also discussed.
2.1. Introduction
Since the U.S. subprime mortgage crisis and the collapse of Lehman
Brothers in 2008, financial institutions have paid more attention to credit risk
management. It has become increasingly essential for banks and financial
institutions to have an accurate credit risk model to manage and measure the
credit risk of their portfolio loans. As noted in Hull (2009) and Demyanyk
and van Hemert (2009), one of the reasons for the credit crush in 2008 was
the lack of easily updatable models for evaluating the credit risk of loan
portfolios.
There are two main types of models for quantifying credit risk: one is the
structural model and the other is the reduced-form model. The former is based
on Merton (1974) and is known as the balance sheet approach. This approach
defines a default event as occurring when the obliger’s asset value process
declines beyond a certain threshold value.
The Vasicek single-factor model assumes that the default events are driven
by a latent common factor, which is assumed to follow the Gaussian
distribution. The one-factor Gaussian copula model by Li (2000) is the
market standard tool for pricing CDOs. In this approach, the Gaussian copula
function is used to model the dependency of asset returns. However, it is well
known that the Gaussian assumption for the asset return process is not valid,
as many empirical studies suggest. It is often reported that the Gaussian
Credit Portfolio Risk Evaluation 35
copula model cannot capture the estimates of the implied CDO tranche
correlation and cannot model the external events caused by the dependency of
the tail behavior. These shortcomings are reported in Brigo et al. (2010). In
addition, various types of copula functions are proposed in the literature (see,
for example, Hofert and Scherer (2011) and references therein).
where Vi,t denotes the asset value at time t of obligor i, and µ is the drift,
which means the average growth rate of Vi,t , and σ is the volatility parameter.
We assume that these parameters are common to every obligor. To model the
obligor’s asset correlation, the Brownian motion Wi,t is divided into two
independent Brownian motions such that
√
dWi,t = ρdX0,t + 1 − ρdXi,t ,
A SSUMPTION 2.2.– The kth order cumulants cXi (u1 , . . . , uk−1 ) of {Xi }, for
k = 2, 3, 4 and i ∈ {0, . . . , N } satisfy
∞
|cXi (u1 , . . . , uk−1 )| < ∞.
u1 ,...,uk−1 =−∞
We define Y0,n = √1n nj=1 X0,j and Yi,n = √1n nj=1 Xi,j . Then, we
obtain the following cumulants for the processes Yi,n . Recall that we have
assumed that the random variables Yi,n and Yj,n are independent for all i, j =
0, . . . , N with i = j, and all the joint cumulants of Yi,n and Yj,n are equal to
zero.
L EMMA 2.1.– Under Assumptions 2.1 and 2.2, the cumulants of {Yi,n } for
i ∈ {0, . . . , N } are evaluated as follows:
1) E[Yi,n ] = 0;
2 + o(n−1 );
2) cum(Yi,n , Yi,n ) = σi,n
(n)
3) cum(Yi,n , Yi,n , Yi,n ) = n−1/2 CYi ,3 + o(n−1 );
(n)
4) cum(Yi,n , Yi,n , Yi,n , Yi,n ) = n−1 CYi ,4 + o(n−1 );
2 , C (n) (n)
where σi,n Yijk ,3 and CYijkl ,4 are bounded for n.
√ √
To have the process ρY0,n + 1 − ρYi,n possess the unit variance and
correlation coefficient ρ for different obligors, we consider the following
38 Data Analysis and Applications 4
√ −1 √ −1
standardized process. Let Wi,n = ρσ0,n Y0,n + 1 − ρσi,n Yi,n . Then, the
asset value process Vi,t is expressed by
√
Vi,t = Vi,nΔ = Vi,0 exp{µt + tσWi,n }
√ √ −1 −1
= Vi,0 exp µt + tσ ρσ0,n Y0,n + 1 − ρσi,n Yi,n . [2.2]
L EMMA 2.2.– Under Assumptions 2.1 and 2.2, the cumulants of {WN,n } are
evaluated as follows:
1) E[WN,n ] = 0N ;
2) Var(WN,n ) = ΣN,n + o(1), where the (i, j)th elements of ΣN,n are ρ
for i, j = 1, . . . , N and the diagonal elements are 1;
(n)
3) For i, j, k ∈ {1, . . . , N }, cum(Wi,n , Wj,n , Wk,n ) = CW,ijk + o(1),
where
(n)
(n) n−1/2 ρ3/2 CY0 ,3 (i = j = k),
CW,ijk = (n) (n)
n−1/2 (ρ3/2 CY0 ,3 + (1 − ρ)3/2 CYi ,3 ) (i = j = k).
(n)
4) For i, j, k, l ∈ {1, . . . , N }, cum(Wi,n , Wj,n , Wk,n , Wl,n ) = CW,ijkl +
o(1), where
(n)
(n) n−1 ρ2 CY0 ,4 (i = j = k = l),
CW,ijkl = −1 2 (n) 2 (n)
n (ρ CY0 ,4 + (1 − ρ) CYi ,4 ) (i = j = k = l).
Here, we use the following rules for the cumulants with the sum of
independent random variables, namely the kth-order joint cumulant of X + Y
is given by
k
(k) k
cum (X + Y ) = cum(k) (X, . . . , X , Y, . . . , Y ).
j
j=0 j k−j
Credit Portfolio Risk Evaluation 39
We then arrive at the following theorem. For more details and proofs, we
refer to Taniguchi and Kakizawa (2000).
where φ(·) is the standard normal density function and Hk (·) is the kth-order
Hermite polynomial.
1 (n) ijkl 1 (n) (n) ijki j k
+ CW,ijkl H (y) + CW,ijk CW,i j k H (y) + o(n−1 ),
24n N,n 72n N,n
where φΣN,n (y) is the N -dimensional normal density function with mean 0N
and covariance matrix ΣN,n .
To confirm that the process defined by equation [2.2] has expected return
µ, we define
3
√ 3/2 (n) 3/2 (n)
σ 2 σ t(ρ CY0 ,3 + (1 − ρ) CYi ,3 )
m= + √
2 6 n
(n) (n)
σ 4 t(ρ2 CY0 ,4 + (1 − ρ)2 CYi ,4 )
+ .
24n
With N = 1 and √
Theorem 2.2, we obtain the expected value of Ṽi,t =
Vi,0 exp (µ − m)t + tσWi,n as follows
ˆ ∞ √
(μ−m)t
E0 [Ṽi,t ] = Vi,0 e eσ ty
gW1 (y)dy = Vi,0 eμt .
−∞
√ Hereafter,
we consider the process Ṽi,t = Vi,0 exp {(µ − m)t+
tσWi,n .
Let PDi be the probability of default and WCDRi be the worst-case default
rate for obligor i. In general, the worst-case default rate is measured with a
99% occurrence of probability. The IRB approach of Basel II focuses on credit
risk measures
nsuch as expected loss (EL) and nvalue at risk (VaR), which are
defined by i=1 PDi × LGDi × EADi and i=1 WCDRi × LGDi × EADi ,
Credit Portfolio Risk Evaluation 41
respectively. Here, LGDi is obligor i’s loss given default and EADi is obligor
i’s exposure at default. The economic capital (EC) is defined as VaR − EL.
Recall that we have assumed that all loans have the same cumulative
probability distribution for the time to default, which is given in Theorem 2.2,
and we denote this distribution as GW1 . For notational convenience, we write
the representative idiosyncratic random factor as Y1,n . A representative
obligor defaults when its asset value, Ṽt , falls below the default trigger Kt .
Hence, the probability of default is stated in the following theorem.
C OROLLARY 2.1.– Under the same conditions as Theorem 2.3, the conditional
default rate is obtained for Y0,n = y0 :
where
√ −1
(1) Kt − ρσ0,n Y0,n
pt (Y0 ) = Φ(Ct (Y0 )), Ct (Y0 ) = √ −1 ,
1 − ρσ1,n
(2) 1 (n)
pt (Y0 ) = − √ CY1 ,3 (Ct (Y0 )2 − 1)φ(Ct (Y0 ) ,
6 n
(3) 1 (n)
pt (Y0 ) = − CY1 ,4 (Ct (Y0 )3 + 3Ct (Y0 ))φ(Ct (Y0 ) ,
24n
(4) 1 (n)
pt (Y0 ) = (CY1 ,3 )2 (−Ct (Y0 )5
72n
+10Ct (Y0 )3 − 15Ct (Y0 ))φ(Ct (Y0 ) .
Kt = G−1
W1 (P D)
(n) (n)
ρ3/2 CY0 ,3 + (1 − ρ)3/2 CY1 ,3 −1 2
= Φ−1 (P D) + √ Φ (P D) − 1
6 n
(n) (n)
ρ2 CY0 ,4 + (1 − ρ)2 CY1 ,4 −1 3
+ Φ (P D) − 3Φ−1 (P D)
24n
(n) (n) (n) (n)
ρ3 (CY0 ,3 )2 + (1 − ρ)3 (CY1 ,3 )2 + 2(ρ(1 − ρ))3/2 CY0 ,3 CY1 ,3
−
36n
−1 2
× 2 Φ (P D) − 5Φ−1 (P D) + o(1).
Here, the Edgeworth density functions gY0 (x) and gY1 (x), and their
corresponding distribution function GY1 (x), are given in Theorem 2.1. Under
44 Data Analysis and Applications 4
the Gaussian assumptions on the asset returns, the above density function
reduces to that of a Gaussian–Merton model. That is,
! "
1−ρ 1
f (DR) = exp (N −1 (DR))2
ρ 2
√ #$
1 − ρN −1 (DR) − N −1 (P D) 2
− √ .
ρ
We use default rate data from the U.S. default history record analyzed in
Hull (2012). We formulate a parameter estimation based on the maximum
likelihood method below. The model parameter vector θ is denoted as
(n) (n) (n) (n) T
θ = P D, ρ, C̃Y0 ,3 , C̃Y0 ,4 , C̃Y1 ,3 , C̃Y1 ,4 . The observed default records are
denoted as DRt (t = 1, . . . , M ). Then, the maximum likelihood estimator
(MLE) of θ is defined as
M
θ̂ (M L) = argmaxθ log hDR (DRt ; θ).
t=1
(n)
for CYi ,3 ≥ 0,
√ (n)
(n) 24 nH3 (Fn,max
−1
(CYi ,4 ))
CYi ,3 − (n) (n) (n)
> 0,
(3H3 (Fn,max
−1
(CYi ,4 )) · H2 (Fn,max
−1
(CYi ,4 )) − 4H3 (Fn,max
−1
(CYi ,4 ))2 )
(n)
for CYi ,3 < 0, together with 0 ≤ ρ ≤ 1 and 0 ≤ P D ≤ 1. Here, function Fn
is defined as
−1 −1
and the values Fn,min (x) and Fn,max (x) are the minimum and maximum roots
of Fn (y) = x, respectively.
Figure 2.1. Histogram of U.S. default rates with estimated default rate
distributions together with their credit risk measures. For a color version
of this figure, see www.iste.co.uk/makrides/data4.zip
46 Data Analysis and Applications 4
is made. Assuming that the short-term interest rate is (r(t))t≥0 , the initial value
of all default payments up to time T is given by
"ˆ T ˆ t #
def {AP,DP }
V{AP,DP } =E exp − r(s)ds dLt .
0 0
To keep our analysis simple, the value of the default payments can be
expressed with the expected value of the loss process by the partial
integration below.
ˆ T
def {AP,DP }
V{AP,DP } = exp − r(s)ds E LT
0
ˆ T ˆ t
{AP,DP }
+ r(t) exp − r(s)ds E Lt dt.
0 0
Credit Portfolio Risk Evaluation 47
Tn
ˆ tn
prem
V{AP,DP } (x) =x (tn − tn−1 ) exp − r(s)ds
n=1 0
% &
{AP,DP }
× (DP − AP )N − E Ltn .
The fair tranche spread, x{AP,DP } , is then determined by equating the value
of the default and premium payments,
In addition, we assume that default can only occur on the dates t1 < . . .
< tN and that Li,t = PDi × LGDi × EADi . All we have to do to evaluate
the expected value of Lt is to focus on the default rate distribution. Then, both
sides of the equation can be expressed as functions of
{AP,DP }
E Lt = E ((Lt − AP · N )+ − (Lt − DP · N )+ )
´ DP
Lt · h(Lt )dLt
AP
= .
P ((AP < Lt ) ∩ (Lt < DP ))
We present the numerical results of CDO pricing in Table 2.3. Table 2.3
shows that the CDO spread with a Gaussian model is higher than Y1 : the
non-Gaussian model for all tranches. In addition, we find that for Y0
non-Gaussian cases, the CDO spreads for the [6,9] and [12,22] tranches have
larger spreads than that of the other models.
48 Data Analysis and Applications 4
2.5. Conclusion
2.6. Acknowledgements
2.7. References
[BRI 10] B RIGO D., PALLAVICINI A., T ORRESETTI R., Credit Models and the Crisis: A
Journey into CDOs, Copulas, Correlations and Dynamic Models, John Wiley & Sons, New
York, 2010.
[BUT 07] B UTLER R.W., Saddlepoint Approximations with Application, Cambridge
University Press, New York, 2007.
[DEM 09] D EMYANYK Y., VAN H EMERT O., “Understanding the subprime mortgage crisis”,
The Review of Financial Studies, vol. 24, no. 6, pp. 1848–1880, 2009.
[GIG 16] G IGLIO S., K ELLY B., P RUITT S., “Systemic risk and the macroeconomy: An
empirical evaluation”, Journal of Financial Economics, vol. 119, no. 3, pp. 457–471, 2016.
[GOR 02] G ORDY M., “Saddlepoint approximation of credit risk”, Journal of Banking and
Finance, vol. 26, pp. 1335–1353, 2002.
[HOF 11] H OFERT M., S CHERER M., “CDO pricing with nested Archimedean copulas”,
Quantitative Finance, vol. 11, no. 5, pp. 775–787, 2011.
Credit Portfolio Risk Evaluation 49
[HUA 07] H UANG X., O OSTERLEE C.W., VAN DER W EIDE J.A.M., “Higher-order
saddlepoint approximations in the Vasicek portfolio credit loss model”, Journal of
Computational Finance, vol. 11, pp. 93–113, 2007.
[HUA 11] H UANG X., O OSTERLEE C.W., “Saddlepoint approximations for expectations and
application to CDO pricing”, SIAM Journal on Financial Mathematics, vol. 2, no. 1,
pp. 692–714, 2011.
[HUL 04a] H ULL J., N ELKEN I., W HITE A., “Merton’s model, credit risk, and volatility
skews”, Journal of Credit Risk Volume, vol. 1, no. 1, pp. 1–27, 2004.
[HUL 04b] H ULL J., W HITE A., “Valuation of a CDO and an nth to default CDS without
Monte Carlo simulation”, Journal of Derivatives, vol. 12, no. 2, pp. 8–23, 2004.
[HUL 09] H ULL J.C., “The credit crunch of 2007: What went wrong? Why? What lessons
can be learnt?”, Journal of Credit Risk, vol. 5, no. 2, pp. 3–18, 2009.
[HUL 12] H ULL J.C., Risk Management and Financial Institutions, 3rd edition, Wiley,
New York, 2012.
[JAR 95] JARROW R., T URNBULL S., “Pricing derivatives on financial securities subject to
credit risk”, Journal of Finance, vol. 50, pp. 53–85, 1995.
[KAW 16] K AWADA A., S HIOHAMA T., “Structural credit risks with non-Gaussian and
serially correlated innovations”, American Journal of Mathematical and Management
Sciences, vol. 35, no. 2, pp. 143–158, 2016.
[LI 00] L I D.X., “On default correlation: A copula function approach”, The Journal of Fixed
Income, vol. 9, no. 4, pp. 43–54, 2000.
[TAN 00] TANIGUCHI M., K AKIZAWA Y., Asymptotic Theory of Statistical Inference for Time
Series, Springer, New York, 2000.
[VAS 02] VASICEK O., “The distribution of loan portfolio value”, Risk, vol. 15, no. 12,
pp. 160–162, 2002.
3
During the development of credit risk assessment models, it is very important to identify variables
that allow us to evaluate a company’s credit risk accurately, as the classification results depend
on the appropriate characteristics for a selected data set. Many studies have been focused on the
characteristics that should be used in credit scoring applications. The data that has been used
in most of these studies is either financial data or credit behavior data. However, there are other
sources which can also provide useful information and which have not been explored accordingly.
Our main objective is to explore these alternative sources of information. To that end, we introduce
alternative data to a predictive model which uses only traditional credit behavior data, in order to
see if the former contribute to the model’s performance. In this chapter, a new credit risk model,
tested on real data, which evaluates the credit risk of Greek hotels, is introduced. This model
uses a combination of credit behavior data and alternative data. The credit risk model introduced
in this chapter does have some important additional advantages: a) it contains a relatively small
number of variables, b) its stability is tested on samples after the time period of data selection and
for different populations and c) the characterization of “good” and “bad” credit behavior is strictly
defined.
3.1. Introduction
Credit risk is one of the major threats that financial institutions face. Credit
scoring is concerned with assessing credit risk and providing for informed
There are two periods that are studied during the model’s creation: the
observation period and the performance period. In this chapter, a period of
12 months (01/01/2016 to 12/31/2016) will be used as a performance period
and 24 months (01/01/2014 to 12/31/2015) as an observation period, as often
occurs in creating similar models (e.g. in Siddiqi [SID 06]). These models are
intended to discriminate the “bad” from “good” behavior in the performance
period. First of all, we have to specify what we mean by “bad” and “good”
credit behavior of a company:
1) Companies with “good” behavior are companies with no delinquency
or companies with maximum delinquency in the last 12 months of from 0 to
29 days past due or credit limit utilization over 102% from 0 to 29 days,
concerning SME overdrafts.
2) Companies with “bad” behavior are companies showing “severe
delinquency”, which means:
- SME (small and medium-sized enterprises) contracts, not overdrafts,
with maximum delinquency in the last 12 months of greater than or equal to
90 days past due or
- SME overdrafts with maximum delinquency in the last 12 months of
greater than or equal to 90 days past due or credit limit utilization over 102%
for a time period greater than or equal to 90 days with the overlimit amount
greater than 100 euros.
In the case where there are some guarantors for the company, “bad” is the
company which has the following credit behavior:
– SME contracts, not overdrafts, with maximum delinquency in the last
12 months of greater than or equal to 150 days past due or
54 Data Analysis and Applications 4
A company is also included in the ones with “bad” credit behavior when
there is a new DFO (Default Financial Obligation) (loan denunciation) within
the performance period. The term “utilization” refers to the ratio current
balance of the company or its credit limit.
and, finally, data from mortgages. The set of hotels having credit transactions
with banks (all 678 of them) was used throughout the analysis. The data set
includes several independent variables to create a credit scorecard. These
variables are associated with information from the application form (e.g. a
loan amount), the status of the credit (e.g. the current balance) and the credit
behavior of company (e.g. bankruptcy). In this data set, we added the
“alternative” variables that we created using information from social media
and customer reviews in order to analyze them together with the already
existing variables. The alternative variables that took part in the analysis are:
a hotel’s registration on Facebook, Twitter, Instagram, LinkedIn or YouTube;
the number of hotel awards; the hotel’s rating on TripAdvisor; the number of
votes on TripAdvisor; the hotel’s rating on Booking.com and the number of
votes on Booking.com. Using the above alternative variables, we created two
two-dimensional variables: the first is the combination of a hotel’s
registration on Twitter and Instagram, and the second is the average rating of
TripAdvisor and Booking.com combined with the sum of votes on
TripAdvisor and Booking.com. This data set also includes a binary response
variable that indicates whether or not a default event was observed in a given
period of time.
where
Table 3.1 is the classification table, which shows that the addition of
independent variables increases the proportion of cases (from the 50-50 case)
of the dependent variable that are correctly predicted by the model. In this
case, the model correctly predicts 92.9% (accuracy) of the observations. This
percentage is higher than the previous model’s accuracy, which was 91.4%
(see in section 3.5).
Table 3.2 contains K-S and Gini index values, which are 77.0% and 0.90,
respectively, and they are used in order to verify if the model is capable of
distinguishing two populations. We note that both K-S and Gini index values
are higher than they were in the previous model (74.8% and 0.88, respectively;
see section 3.5).
In Table 3.3, it appears that K-S (79.0%) is better than before (see Table 3.2)
and Gini index remains the same (0.90), so we can conclude that the model is
still efficient and stable. The model’s stability is also confirmed once again in
Table 3.4, as its stability value is 0.00.
In Table 3.5, it is observed that K-S (77.6%) and Gini index (0.90) remain
high, proving that the model is also suitable for other populations.
Towards an Improved Credit Scoring System with Alternative Data 59
3.7. Conclusion
3.8. References
[BEL 09] B ELLOTTI T., C ROOK J., “Credit scoring with macroeconomic variables using
survival analysis”, Journal of the Operational Research Society, vol. 60, pp. 1699–1707,
2009.
[BEL 12] B ELLOTTI T., C ROOK J., “Loss given default models incorporating macroeconomic
variables for credit cards”, International Journal of Forecasting, vol. 28, no. 1, pp. 171–182,
2012.
60 Data Analysis and Applications 4
[DES 96] D ESAI V.S, C ROOK J.N., OVERSTREET G.A., “A comparison of neural networks
and linear scoring models in the credit union enviroment”, European Journal of Operational
Research, vol. 95, no. 1, pp. 24–37, 1996.
[FAL 10] FALANGIS K., G LEN J.J., “Heuristics for feature selection in mathematical
programming discriminant analysis models”, Journal of the Operational Research Society,
vol. 61, no. 5, pp. 804–812, 2010.
[FLE 93] F LETCHER D., G OSS E., “Forecasting with neural networks: An application using
bankruptcy data”, Information and Management, vol. 24, no. 3, pp. 159–167, 1993.
[FLO 10] F LOREZ -L OPEZ R., “Effects of missing data in credit risk scoring. A comparative
analysis of methods to achieve robustness in the absence of sufficient data”, Journal of the
Operational Research Society, vol. 61, no. 3, pp. 486–501, 2010.
[HAN 98] H AND D.J., JACKA S.D., Statistics in Finance, Edward Arnold Publishers Ltd.,
London, 1998.
[JO 97] J O H., H AN I., L EE H., “Bankruptcy prediction using case-based reasoning, neural
networks and discriminant analysis”, Expert Systems with Applications, vol. 13, no. 2,
pp. 97–108, 1997.
[LES 96] L ESHNO M., S PECTOR Y., “Neural networks prediction analysis: The bankruptcy
case”, Neurocomputing, vol. 10, no. 2, pp. 125–147, 1996.
[LIU 05] L IU Y., S CHUMANN M., “Data mining feature selection for credit scoring models”,
Journal of the Operational Research Society, vol. 56, no. 9, pp. 1099–1108, 2005.
[LOT 12] L OTERMAN G., B ROWN I., M ARTENS D. et al., “Benchmarking regression
algorithms for loss given default modelling”, International Journal of Forecasting, vol. 28,
no. 1, pp. 161–170, 2012.
[PEN 05] P ENDHARKAR P.C., “A threshold-varying artificial networks approach for
classification and its application to bankruptcy prediction problems”, Computers and
Operations Research, vol. 32, no. 10, pp. 2561–2582, 2005.
[SAL 92] S ALCHENBERGER L.M., C INAR E.M., L ASH N.A., “Neural networks: A new tool
for predicting thrift failures”, Decision Sciences, vol. 23, no. 4, pp. 899–916, 1992.
[SID 06] S IDDIQI N., Credit Risk Scorecard: Developing and Implementing Intelligent Credit
Scoring, John Wiley & Sons, New York, 2006.
[SOM 07] S OMERS M., W HITTAKER J., “Quantile regression for modelling distribution of
profit and loss”, European Journal of Operational Research, vol. 183, no. 3, pp. 1477–
1487, 2007.
[STE 02] S TEPANOVA M., T HOMAS L., “Survival analysis methods for personal loan data”,
Operational Research, vol. 50, no. 2, pp. 277–289, 2002.
[TAM 92] TAM K.Y., K IANG M.Y., “Managerial applications of neural networks: The case
of bank failure predictions”, Management Science, vol. 38, no. 7, pp. 926–947, 1992.
[TON 12] T ONG E.N.C., M UES C., T HOMAS L.C., “Mixture sure models in credit scoring:
If and when borrowers default”, European Journal of Operational Research, vol. 218,
no. 1, pp. 132–139, 2012.
4
Research of α-stable distributions is especially important nowadays, because they often occur in
the analysis of financial data and information flows along computer networks. It has been found
that financial data are often leptokurtic with heavy-tailed distributions. Many authors have proved
that the most commonly used distribution is not the most suitable way to analyze economic
indicators and suggested to replace it with more general, for example, stable distributions.
However, a problem arises how to estimate a multidimensional stable data. Moreover,
one-dimensional α-stable distributions are efficiently estimated by the maximum likelihood
method. Therefore, a maximum likelihood method for the estimation of multivariate α-stable
distributions by using an expectation–maximization algorithm is presented in this work. Integrals
included in the expressions of the estimates have been calculated using the Gaussian and
Gauss–Laguerre quadrature formulas. The constructed model can be used in stock market data
analysis.
4.1. Introduction
Weron [JAN 93]; Rachev and Mittnik [RAC 00]; Samorodnitsky and Taqqu
[SAM 94]; etc.). Therefore, a well-known normal distribution does not
always fit, for example returns of stocks or risk factors are badly fitted by the
normal distribution (Kabasinskas et al. [KAB 09]; Belovas, Kabasinskas and
Sakalauskas [BEL 06]). In this case, normal distributions are replaced with
more general, for example, stable distributions, which make it possible to
model both leptokurtic and asymmetric (Fielitz and Smith [FIE 72]; Rachev
and Mittnik [RAC 00]; Kabasinskas et al. [KAB 12]; Sakalauskas et al.
[SAK 13]). Therefore, stable distributions are the most often used in business
and economics data analysis. According to some experts, the α-stable
distribution offers a reasonable improvement – if not the best choice – among
the alternative distributions that have been proposed in the literature over the
past four decades (e.g. Bertocchi et al. [BER 05]; Hoechstoetter, Rachev and
Fabozzi [HOE 05]). However, a practical application of stable distributions is
limited by the fact that their distribution and density functions are not
expressed through elementary functions, except for a few special cases
(Janicki and Weron [JAN 93]; Rachev and Mittnik [RAC 00]; Belovas,
Kabasinskas and Sakalauskas [BEL 06]). By the way, stable distributions
have infinite variance (except for the normal case). In this work, the stable
multivariate variables expression through normal multivariate vector with
random variance, changing by a particular stable law, are used for the
simulation.
√
In one-dimensional case, it is known that s = s1 · s2 , where:
where, s ≥ 0, and
(α/(2 − α))
f ( x| μ, Ω, α) = ×
2 · (2π)d/2 · |Ω|1/2
Uα (y) α/(2−α)
ˆ ∞ˆ 1 −1
1 T Ω
× exp − (x − μ) (x − μ) − ×
0 −1 2 s s
α/2−α
Uα (y)
× d/2+2/(2−α)
dy ds. [4.5]
s
K
(α/(2 − α))K
f X i μ, Ω, α =
L̃ (X, μ, Ω, α) =
i=1 2K · (2π)K·d/2 · |Ω|K/2
K ˆ ∞ˆ 1
1 i T Ω−1 i α/(α−2)
× exp − X −μ X − μ − si · Uα (yi ) ×
0 −1 2 si
i=1
Uα (yi )
× d/2+2/(2−α)
dyi dsi . [4.6]
si
EM Algorithm for Estimating the Parameters of the Multivariate Stable Distribution 65
Denote
α/(α−2)
zi = si · Uα (yi ) . [4.7]
K ˆ ∞ ˆ 1
B X i , yi , zi , μ, Ω, α dyi dzi ,
=− ln exp {−zi } [4.8]
i=1 0 −1
where
1 (d(2−α))/(2α)
B X i , yi , zi , μ, Ω, α =
1/2
· zi ×
2· (2π)d/2 · |Ω| · Uα (yi )
T
X i − μ Ω−1 X i − μ
(2−α)/α
× exp − · zi . [4.9]
2 · Uα (yi )
K
⎧
∂f X i μ, Ω, α
⎪ ∂L (X, μ, Ω, α) 1
⎪
⎪ =− · i
= 0,
∂μ ∂μ f ( X | μ, Ω, α)
⎪
⎪
i=1
⎨
[4.10]
K
Xi
∂L (X, μ, Ω, α) ∂f μ, Ω, α 1
⎪
⎪
⎪
⎪
⎪ =− · i
= 0.
⎩ ∂Ω ∂Ω f ( X | μ, Ω, α)
i=1
Ω−1 X i − μ
(2−α)/α
· B X i , yi , zi , μ, Ω, α ,
= · zi [4.11]
Uα (yi )
∂B X i , yi , zi , μ, Ω, α
∂Ω
1 (d+2)(2−α)/(2α)
= 1/2
· zi ×
d/2
2 · (2π) · |Ω| · Uα (yi )
i T
X − μ Ω−1 X i − μ
(2−α)/α
× exp − · zi =
2 · Uα (yi )
×B X i , yi , zi , μ, Ω, α .
[4.12]
K K
X i · gi gi
h (X, μ, Ω, α) = , [4.13]
fi fi
i=1 i=1
K T
X i − μ̂ X i − μ̂ gi
w (X, μ, Ω, α) = . [4.14]
fi
i=1
K
∂L gi
= (h (X, μ, Ω, α) − μ) · , [4.15]
∂μ fi
i=1
∂L
= −K · Ω−1 + Ω−1 · w (X, μ, Ω, α) · Ω−1 , [4.16]
∂Ω
EM Algorithm for Estimating the Parameters of the Multivariate Stable Distribution 67
where
ˆ
1
∞ B X i , y, z, μ, Ω, α
ˆ
g (X, μ, Ω, α) = dy z (2−α)/α e−z dz,[4.17]
0 −1 Uα (y)
ˆ ∞ ˆ 1
B X , y, z, μ, Ω, α dy e−z dz.
i
f (X, μ, Ω, α) = [4.18]
0 −1
K K
X i · gi gi
μ̂ = , [4.19]
fi fi
i=1 i=1
K T
X i − μ̂ X i − μ̂
1 gi
Ω̂ = . [4.20]
K fi
i=1
Figure 4.3. Likelihood function test. For a color version of this figure,
see www.iste.co.uk/makrides/data4
4.5. Conclusion
1) ML method for the multivariate α-stable distribution was created in this
work, which makes it possible to estimate parameters of this distributions using
the EM algorithm.
2) The α-stable distribution parameter estimators obtained by the numerical
simulation method are statistically adequate, because after a certain number
of iterations, values of likelihood function and parameters converge to the ML
values.
3) It was shown that this method realizes the log-likelihood function golden
section search, implementing it with EM algorithm.
4) This algorithm was applied to create the model of balance data of US
companies. It can be used to create financial models in stock market data
analysis. In addition, it can be used to test the systems of stochastic type and
to solve other statistical tasks.
4.6. References
[BEL 06] B ELOVAS I., K ABASINSKAS A., S AKALAUSKAS L., “A study of stable models of
stock markets”, Information Technology and Control, vol. 35, no. 1, pp. 34–56, 2006.
[BER 05] B ERTOCCHI M., G IACOMETTI R., O RTOBELLI S. et al., “The impact of different
distributional hypothesis on returns in asset allocation”, Finance Letters, vol. 3, no. 1,
pp. 17–27, 2005.
72 Data Analysis and Applications 4
[BOG 09] B OGDAN K., B YCZKOWSKI T., K ULCZYCKI T. et al., Potential Analysis of Stable
Processes and its Extensions, Springer Science and Business Media, New York, 2009.
[CAS 11] C ASIO C OMPUTER C O “Nodes and weights of Gauss-Laguerre Calculator”.
Available at: http://keisan.casio.com/exec/system/1281279441, 2011.
[DAV 99] DAVYDOV Y., PAULAUSKAS V., “On the estimation of the parameters of
multivariate stable distributions”, Acta Applicandae Mathematica, vol. 58, no. 1, pp. 107–
124, 1999.
[EHR 02] E HRICH S., “On stratified extensions of Gauss-Laguerre and Gauss-Hermite
quadrature formulas”, Journal of Computational and Applied Mathematics, vol. 140,
nos 1–2, pp. 291–299, 2002.
[EON 10] EON: E NHANCED O NLINE N EWS, “Audit Integrity’s “AGR” rating outperforms
leading academic accounting risk measures”, independent study finds, 2010. Available at:
http://www.businesswire.com/news/home/20100315006057/en/.
[FIE 72] F IELITZ B.D., S MITH E.W., “Asymmetric stable distributions of stock price
changes”, Journal of American Statistical Association, vol. 67, no. 340, pp. 331–338, 1972.
[HOE 05] H OECHSTOETTER M., R ACHEV S., FABOZZI F.J., “Distributional analysis of the
stocks comprising the DAX 30”, Probability and Mathematical Statistics, vol. 25, no. 1,
pp. 363–383, 2005.
[JAN 93] JANICKI A., W ERON A., Simulation and Chaotic Behavior of - Stable Stochastic
Processes, Marcel Dekker, New York, 1993.
[KAB 09] K ABASINSKAS A., R ACHEV S., S AKALAUSKAS L. et al., “Stable paradigm in
financial markets”, Journal of Computational Analysis and Applications, vol. 11, no. 3,
pp. 642–688, 2009.
[KAB 12] K ABASINSKAS A., S AKALAUSKAS L., S UN E.W. et al., “Mixed-stable models
for analyzing high-frequency financial data”, Journal of Computational Analysis and
Applications, vol. 14, no. 7, pp. 1210–1226, 2012.
[KOV 12] KOVVALI N., Theory and Applications of Gaussian Quadrature Methods, Morgan
and Claypool Publishers, 2012.
[KRI 09] K RING S., R ACHEV S.T., H OCHSTOTTER M. et al., “Estimation of α-stable sub-
Gaussian distributions for asset returns”, Risk Assessment: Decisions in Banking and
Finance, pp. 111–152, 2009.
[NOL 98] N OLAN J.P., “Multivariate stable distributions: Approximation, estimation,
simulation and identification”, A Practical Guide to Heavy Tails, pp. 509–525, 1998.
[NOL 07] N OLAN J.P., Stable Distributions – Models for Heavy Tailed Data, Birkhauser,
Boston, 2007.
[OGA 13] O GATA H., “Estimation for multivariate stable distributions with generalized
empirical likelihood”, Journal of Econometrics, vol. 172, no. 2, pp. 248–254, 2013.
[PRI 11] P RICE R.A., S HARP N.Y., W OOD D.A., “Detecting and predicting accounting
irregularities: A comparison of commercial and academic risk measures”, Accounting
Horizons, vol. 25, no. 4, pp. 755–780, 2011.
EM Algorithm for Estimating the Parameters of the Multivariate Stable Distribution 73
[PRE 72] P RESS S.J., “Estimation in univariate and multivariate stable distributions”, Journal
of the American Statistical Association, vol. 67, no. 340, pp. 842–846, 1972.
[RAC 93] R ACHEV S.T., M ITTNIK S., “Modeling asset returns with alternative stable
distributions”, Econometric Reviews, vol. 12, no. 3, pp. 261–330, 1993.
[RAC 00] R ACHEV S.T., M ITTNIK S., Stable Paretian Models in Finance, Wiley, New York,
2000.
[RAC 93] R ACHEV S.T., X IN H., “Test for association of random variables in the domain
of attraction of multivariate stable law”, Probability and Mathematical Statistics, vol. 14,
no. 1, pp. 125–141, 1993.
[RAV 99] R AVISHANKER N., Q IOU Z., “Monte Carlo EM estimation for multivariate stable
distributions”, Stat. Prob. Lett., vol. 45, no. 4, pp. 335–340, 1999.
[SAK 10] S AKALAUSKAS L., “On the empirical Bayesian approach for the Poisson-Gaussian
model”, Methodology and Computing in Applied Probability, vol. 12, no. 2, pp. 247–259,
2010.
[SAK 13] S AKALAUSKAS L., K ALSYTE Z., VAICIULYTE I. et al., “The application of stable
and skew t-distributions in predicting the change in accounting and governance risk ratings”,
Proceedings of the 8th International Conference Electrical and Control Technologies,
pp. 53–58, 2013.
[SAK 14] S AKALAUSKAS L., VAIIULYT I., “Sub-gausinio vektoriaus skirstinio parametr
vertinimas Monte-Karlo Markovo grandins metodu”, Jaunj mokslinink darbai, vol. 41,
no. 1, pp. 104–107, 2014.
[SAM 94] S AMORODNITSKY G., TAQQU M.S., Stable Non-Gaussian random processes.
Stochastic Models with Infinite Variance, Chapman and Hall, New York, 1994.
[STO 02] S TOER J., B ULIRSCH R., Introduction to Numerical Analysis, Springer Science and
Business Media, New York, 2002.
PART 2
In the study of complex systems (for example a human body), a large number of characteristics
is required to assess the current state and forecast its development. It is usually impossible to
measure all of them because of the lack of time and equipment. So, it becomes necessary to
assess the state and short-term dynamics of its change using signals that can be obtained in
real time.
It is assumed that the behavior of the system can be described by a characteristic RT, which is
nearly periodic. In this way, investigating a complex system is reduced to analyzing the time
series. This paper provides an overview of the existing methods of such analysis. Particular
attention is paid to the methods of nonlinear dynamics and chaotic behavior of systems.
5.1. Introduction
The numerical approach proposes the transition from the time series
(signal) itself to a certain object formed in the phase space of finite
dimensionality which is called as a restored attractor. Various measures can
be used for its characterization, of which the most fundamental appears to be
its fractal dimension (Gudkov 2008).
Figures 5.1 and 5.2 give examples illustrating the concept of HRV.
Statistical methods are used to directly quantify the signal in the time
interval under study (dispersion, coefficient of variation, etc.).
Its disadvantage is that the time when the frequency components occur is
unknown.
There are many widely used signal transformations other than the Fourier
transform. To name a few, there is the Hilbert transform, windowed PF,
Wigner distribution, Walsh transform, wavelet transform and so on. For each
transformation, we can specify the most suitable area of application, its
merits and disadvantages.
as .
The idea here is to weight the probability of the most often visited cubes
according to the order of the dimension. Embedding process has to precede
any estimation of fractals from a data series. The main practical issue lies in
the choice of the embedding variable and embedding delay. If the
measurement variable at time t is x(t), then an (n+1)-dimensional embedding
is defined as [x(t), x(t+τ),…,x(t+nτ)]. The right embedding delay can be
estimated by the first zero crossing of the autocorrelation function or, better,
by the first local minimum of the mutual information. However, the
probability of finding a point on the attractor is necessary. Usually, the
information dimension and related informational entropy are used for this
purpose; the correlation dimension and correlation entropy are also used.
software used for rapid diagnosis of a body state. Some results for the
attractor’s correlation dimension are shown in Figures 5.7 and 5.8.
x −b
∞
1
Wψ ( a, b) =
a
−∞
f ( x) ψ
a
dx
where a is the scale parameter and b is the coordinate or time. The initial
signal f(x) is divided using the function ψ(x) generated from the soliton-like
one with special features by its scale measurements and shifts (Mallat and
Hwang 1992). In the simplest version (Holder’s exponent h), the scaling of
one of the lines (e.g. the maximum one) is studied:
Wψ ( ti ,s ) ~ s ( i )
ht
q
Pq ( s ) = Wψ ( t ,s ) ,
l =⊂ L( t )
where L(t) is the set of all lines (l) on which the wavelet coefficients
reach the modules maximum on the scale t. The graph of the function
k ( q ) : Pq ( s ) =~ s ( ) is given in Figure 5.9.
k q
Methods for Assessing Critical States of Complex Systems 87
5.4. Conclusion
5.5. References
Alligood K.T., Sauer T., Yorke J.A. (1997). Chaos: an introduction to dynamical
systems. Springer-Verlag.
Antonov V., Zagainov A. and Kovalenko A. (2016). “Stochastic Models in Society.
Fractal Analysis of Biological Signals in a Real Time Mode. Global and
Stochastic Analysis”, GSA. 3(2), 75–84.
Antonov V., Zagaynov A. (2015). “Software Package for Calculating the Fractal and
Cross Spectral Parameters of Cerebral Hemodynamic. In a Real Time Mode.
New Trends in Stochastic Modeling and Data Analysis. Demography and
Related Applications”, ISAST, 440, 339–345.
Daubechies I. (1992). Ten Lectures on Wavelets. CBMS-NSF Regional Conference
Series in Applied Mathematics, Philadelphia, Pa.
Gudkov G.V. (2008). “The role of deterministic chaos in the structure of the fetal
heart rate variability”, Modern Problems of Science and Education, Moscow.,
Krasnodar, 1, 413–423.
Malik M., Bigger J.T., Camm A.J., Kleiger R.E., Malliani A., Moss A.J. and
Schwartz P.J. (1996). “Heart rate variability. Standards of measurement,
physiological interpretation and clinical use”, European Heart Journal. 17.
354–381.
Mallat S. and Hwang W.L. (1992). “Singularity Detection and Processing with
Wavelets”, IEEE Transactions on Information Theory, 38(2), 617–643.
Mandelbrot B. (1977). Fractals: Form, Chance, Dimension. Freeman, San-Francisco.
Parlov A. N. and Anis-chenko B.C. (2007). “Multifractal analysis of complex
signals”, Successes of Physical Science, 177(8), 859–876.
Takens F. (1980). “Detecting strange attractors in turbulence”. In: Dynamical
Systems and Turbulence. Lecture Notes in Mathematics, (eds) D.A.R. and L.S.
Young. Heidelberg: Springer-Verlag. 366–381.
6
Extreme value theory (EVT) deals essentially with the estimation of parameters of extreme or
rare events. Extreme events are usually described as observations that exceed a high threshold.
In many environmental and financial applications, clusters of exceedances of that high threshold
are of practical concern. One important parameter in EVT, which measures the amount of
clustering in the extremes of a stationary sequence, is the extremal index, θ. It needs to be
adequately estimated, not only by itself but also due to its influence on other parameters such as
a high quantile, return period or expected shortfall. Some classical estimators of θ and their
asymptotic properties are revisited. The challenges that appear for finite samples are illustrated.
A resampling procedure that has shown to give good results in extreme value theory, the
generalized jackknife methodology, is discussed and applied to improve the extremal index
estimation. An extensive simulation study was performed, and some results are shown. Finally, a
heuristic procedure, based on a stability criterion, is applied to some simulated samples to
estimate θ.
[BUI 89] and Carter and Challenor [CAR 81]), wind speeds (Walshaw and
Anderson [WAL 00]), air temperatures (Smith et al. [SMI 97] and Coles et al.
[COL 94]), precipitation levels (Coles and Tawn [COL 96]), burned areas
(Díaz-Delgado et al. [DÍA 04] and Schoenberg et al. [SCH 03]) and
earthquake thermodynamics (Lavenda and Cipollone [LAV 00]).
θ is the key parameter for extending extreme value theory from independent
and identically distributed random variables to stationary processes.
– Down-crossing characterization (O’Brien [OBR 87])
An alternative characterization of θ, in terms of down-crossings, is given
by O’Brien [OBR 87], for sequences that satisfy a weak mixing condition that
locally restricts the occurrence of clusters
Two classical methods to define clusters are the blocks method and the runs
method, [HSI 91, HSI 93]. The blocks estimator is derived by dividing the data
into approximately kn blocks of length rn , where n ≈ kn ×rn , i.e. considering
kn = [n/rn ]. Each block is treated as one cluster, and the number of blocks
in which there is at least one exceedance of the threshold un is counted. The
B (un ), is then defined as:
blocks estimator, Θ n
kn
B i=1 I max X(i−1)rn +1 , · · · , Xirn > un
Θn (un ) := n . [6.9]
i=1 I (Xi > un )
For many dependent structures, the bias of Θ U C (k) has two dominant
components of orders k/n and 1/k; see Gomes et al. [GOM 08],
U C (k)] = ϕ1 (θ) k 1 k 1
Bias[Θ + ϕ2 (θ) +o +o , [6.12]
n k n k
The generalized jackknife methodology states that if the bias has two main
terms that we would like to reduce, we need to have access to three estimators,
with the same type of bias.
Using the information obtained from equation [6.12] and based on the
estimator Θ U C computed at the three levels, k, [δk] + 1 and [δ 2 k] + 1, where
[x] denotes, as usual, the integer part of x, Gomes et al. [GOM 08] proposed a
class of generalized jackknife estimators, depending on a tuning parameter δ,
0 < δ < 1, Θ GJ(δ) ≡ Θ GJ(δ) (k), defined as:
2 UC UC 2 UC
GJ(δ) := (δ + 1)Θ ([δk] + 1) − δ Θ ([δ k] + 1) + Θ (k) . [6.13]
Θ
(1 − δ)2
A Monte Carlo simulation for the mean value and the mean square error
(MSE) of the up-crossing estimator, Θ U C , and the generalized jackknife
estimators, Θ GJ(δ) , for some values of δ (δ = 0.05 (0.05) 0.95) and several
values of θ was performed. Some of the results obtained are shown in
Figure 6.1.
For this model and also for other models studied, the best values for mean
values and mean square error depend on the value of δ. Among those values of
δ producing the trajectories plotted in Figure 6.1, δ = 0.1 and δ = 0.25 seem
to be the favorites.
A path stability algorithm (see Gomes et al. [GOM 13] and Neves et al.
[NEV 15]) has revealed quite agreeable results for extreme value parameters
estimation and can now be adapted to the choice of δ, followed by the choice
of kopt (in a given sense) for estimating θ.
GJ(δ) estimators:
Let us see the description of the algorithm, for Θ
1) Given an observed sample (x1 , . . . , xn ), compute, for k = 1, . . . , n − 1,
GJ(δ) for a range of values of δ, 0 < δ < 1.
the observed values of Θ
2) Obtain the rounded values, to 0 decimal places, of the estimates in the
GJ(δ) GJ(δ) , 0), k = 1, 2, . . . , n − 1, the
previous step. Define aΘ
k (0) = round(Θ
rounded values of Θ GJ(δ) (k) to 0 decimal places.
3) For each value of δ, consider the sets of k values associated with equal
GJ(δ) GJ(δ) GJ(δ)
consecutive values of aΘk
Θ
(0), obtained in step 2. Set kmin Θ
and kmax
the minimum and maximum values, respectively, of the set with the largest
GJ(δ)
Θ GJ(δ)
Θ
range. The largest run size is then lΘ GJ(δ) := kmax − kmin .
4) Choose the δ value, δ0 , which correspond to the largest value of lΘ
GJ(δ) .
θ GJ(δ0 ) k̂Θ
δ 0 lΘ GJ(δ0 )
GJ(δ0 ) θ
Figure 6.3. The “optimal” choice of δ and k for the sample paths of the
estimates obtained from Θ GJ(δ) , with δ = 0.05, 0.1, 0.25, 0.5, for three
samples of size n = 1, 000 generated from the max-autoregressive
process with θ = 0.1 (top left), θ = 0.5 (top right) and θ = 0.9 (bottom)
6.5. Acknowledgements
6.6. References
[ALP 89] A LPUIM T., “An extremal markovian sequence”, Journal of Applied Probability,
vol. 26, pp. 219–232, 1989.
[BUI 89] B UISHAND A.,“Statistics of extremes in climatology”, Statistica Neerlandica,
vol. 43, pp. 1–30, 1989.
[CAR 81] C ARTER T., C HALLENOR G., “Estimating return values of environmental
parameters”, Quarterly Journal of the Royal Meteorological Society, vol. 107, pp. 259–266,
1981.
[COL 94] C OLES S., TAWN J., S MITH R., “A sazonal Markov model for extremely low
temperatures”, Environmetrics, vol. 5, pp. 221–339, 1994.
[COL 96] C OLES S., TAWN J., “A Bayesian analysis of extreme rainfall data”, Applied
Statistics, vol. 45, pp. 463–478, 1996.
[DAV 11] DAVIDSON A., “Statistics of extremes” Courses 2011–2012, École Polytechnique
Fédérale de Lausanne (EPFL), 2011.
[DÍA 04] D ÍAZ -D ELGADO R., L LORET F., P ONS X., “Spatial patterns of fire occurrence in
Catalonia”, Landscape Ecology, vol. 19, no. 7, pp. 731–745, Spain, 2004.
[GOM 90] G OMES M.I., “Statistical inference in an extremal markovian model”, CompStat,
pp. 257–262, 1990.
[GOM 93] G OMES M.I., “On the estimation of parameters of rare events in environmental
time series”, Statistics for the Environment, pp. 226–241, 1993.
[GOM 08] G OMES M.I., H ALL A., M IRANDA C., “Subsampling techniques and the
Jackknife methodology in the estimation of the extremal index”, Computational Statistics
and Data Analysis, vol. 52, no. 4, pp. 2022–2041, 2008.
100 Data Analysis and Applications 4
[GOM 13] G OMES M.I., H ENRIQUES -RODRIGUES L., F RAGA A LVES M.I. et al., “Adaptive
PORT-MVRB estimation: An empirical comparison of two heuristic algorithms”, Journal
of Statistical Computation and Simulation, vol. 83, no. 6, pp. 1129–1144, 2013.
[GRA 72] G RAY H., S CHUCANY W., The Generalized Jackknife Statistic, Marcel Dekker,
New York, 1972.
[GUM 58] G UMBEL E., Statistics of Extremes, Columbia University Press, New York, 1958.
[HSI 91] H SING T., “Estimating the parameters of rare events”, Stochastic Processes and their
Applications, vol. 37, pp. 117–39, 1991.
[HSI 93] H SING T., “Extremal index estimation for a weekly dependent stationary sequence”,
Annals of Statistics, vol. 21, pp. 2043–2071, 1993.
[HSI 88] H SING T., H USLER J., L EADBETTER M., “On exceedance point process for a
stationary sequence”, Probability Theory and Related Fields, vol. 78, pp. 97–112, 1988.
[LAV 00] L AVANDA B., C IPOLLONE E., “Extreme value statistics and thermodynamics of
earthquakes: Aftershock sequences”, Annals of Geophysics, vol. 43, no. 5, pp. 967–982,
2000.
[LEA 83] L EADBETTER M., “Extremes and local dependence in stationary sequences”, Z.
Wahrsch. Verw. Gebiete, vol. 65, no. 2, pp. 291–306, 1983.
[LEA 83] L EADBETTER M., L INDGREN G., ROOTZÉN H., Extremes and Related Properties
of Random Sequences and Processes, Springer-Verlag, New York, 1983.
[LEA 89] L EADBETTER M., NANDAGOPALAN L., “On exceedance point process for
stationary sequences under mild oscillation restrictions”, in Extreme Value Theory:
Proceedings, Oberwolfach 1987, H ÜSLER J., R EISS R.D. (eds), Lecture Notes in Statistics,
vol. 51, pp. 69–80, Springer-Verlag, Berlin, 1989.
[NAN 90] NANDAGOPALAN S., Multivariate Extremes and Estimation of the Extremal Index,
PhD Thesis, University of North Carolina, Chapel Hill, 1990.
[NEV 15] N EVES M., G OMES M. I., F IGUEIREDO F. et al., “Modeling Extreme Events:
Sample Fraction Adaptive Choice in Parameter Estimation”, Journal of Statistical Theory
and Practice, vol. 9, no. 1, pp. 184–199, 2015.
[OBR 87] O’B RIEN G., “Extreme values for stationary and Markov sequences”, Annals of
Probability, vol. 15, pp. 281–291, 1987.
[PRA 15] P RATA G OMES D., N EVES M., “Bootstrap and other resampling methodologies in
statistics of extremes”, Communications in Statistics – Simulation and Computation, vol.
44, no. 10, pp. 2592–2607, 2015.
[PRA 18] P RATA G OMES D., N EVES M., “Revisiting resampling methods in the extremal
index estimation: Improving risk assessment”, in Recent Studies on Risk Analysis and
Statistical Modeling, O LIVEIRA T.A., K ITSOS C., O LIVEIRA A. et al. (eds), Springer
International Publishing, 2018.
[REI 97] R EISS D., T HOMAS M., Statistical Analysis of Extreme Values: With Applications
to Insurance, Finance, Hydrology and Other Fields, Birkhaüser Verlag, 1997.
Resampling Procedures for a More Reliable Extremal Index Estimation 101
[SCH 03] S CHOENBERG F., P ENG R., H UANG Z. et al., “Detection of nonlinearities in the
dependence of burn area on fuel age and climatic variables”, International Journal of
Wildland, vol. 12, no. 1, pp. 1–10, 2003.
[SMI 86] S MITH R., “Extreme value theory based on the r largest annual events”, Journal of
Hydrology, vol. 86, pp. 27–43, 1986.
[SMI 94] S MITH R., W EISSMAN I., “Estimating the extremal index”, J. R. Statist. Soc. B,
vol. 56, pp. 515–528, 1994.
[SMI 97] S MITH M., M ALATHY D EVI V., B ENNER D. et al., “Temperature dependence of
air-broadening and shift coefficients of O3 lines in the ν1 band”, Journal of Molecular
Spectroscopy, vol. 182, pp. 239–259, 1997.
[TAW 88] TAWN J., “An extreme value theory model for dependent observations”, Journal of
Hydrology, vol. 101, pp. 227–250, 1988.
[WAL 00] WALSHAW D., A NDERSON C.W., “A model for extreme wind gusts”, Journal of
the Royal Statistical Society, Series C (Applied Statistics), vol. 49, pp. 499–508, 2000.
[WEI 98] W EISSMAN I., N OVAK S., “On blocks and runs estimators of the extremal index”,
Journal of Statistical Planning and Inference, vol. 66, pp. 281–288, 1998.
7
The stochastic process theory provides concepts and theorems that enable us to build
probabilistic models concerning accidents. A crucial role in the construction of the models plays
a Poisson process and its generalizations. The non-homogeneous Poisson process and
corresponding non-homogeneous compound Poisson process can be applied for modeling the
number of road, sea and railway accidents in the given time intervals. Those stochastic
processes are used for modeling the road accident number, and number of injured people and
fatalities. To estimate model parameters, data coming from the annual reports of the Polish
police are used.
7.1. Introduction
years because of the kick by a horse. A random variable, say , denoting the
number of solders killed accidentally by the horse kick per year, turned out to
have a Poisson distribution:
( )
( )= ( = )= , ∈ .
!
Since then, the Poisson distribution and its associated Poisson random
process have found applications in various fields of science and technology.
A Poisson process and its extensions are used in safety and reliability
problems. They enable us to construct the number of road, sea and railway
accidents in given time intervals. A non-homogeneous Poisson process
(NPP) in modeling accident number in Baltic Sea and Seaports was
presented in the coference Summer Safety and Reliability Seminars 2017
and is published in the Journal of Polish Safety and Reliability Association
(Grabski F., 2017). The non-homogeneous compound Poisson process
(NCPP) enables us to anticipate the number of injured people and fatalities.
In the elaboration of the theoretical part of the article, books from Fisz M.
(1969), Grabski F. (2015), Limnios N. and Oprisan N.G. (2001), Shiryayev
A.N. (1984) and papers from Di Crescenzo A., Martinucci B. and Zacks S.
(2015), Fisz M. (1969), Grabski F. (2017), Grabski F. (2018), Zinchenko N.
(2017) were used. The statistical data come from the annual police reports
(Symon E. 2018, Symon E. 2019).
( ) ( )
( ( )= )= , = 0,1,2, … [7.3]
!
The expectation and variance of NPP are the functions given by the rules
Λ( ) = [ ( )] = ( ) , [7.4]
V(t) = [ ( )] = ( ) , ≥ 0. [7.5]
D(t) = [ ( )] = ( ) , ≥ 0. [7.6]
Δ( ; ℎ) = ( ( + ℎ) − ( )) = ( ) . [7.7]
D( ; ℎ) = ( ( + ℎ) − ( )) = ( ) [7.8]
( )
( ( )= )= , = 0,1,2; ≥ 0.
!
Table 7.1 shows the number of road accidents and their consequences in
Poland in 2007–2018. The data come from the annual police reports Symon
E. (2018), Symon E. (2019).
= + [7.9]
( , )= ∑ [ −( + )] → [7.10]
= , = − , [7.11]
̅= = ∑ , = = ∑ ,
= ∑ , = − ,
= ∑ , = − .
60,000
50,000
40,000
30,000
20,000
10,000
0
0.5 1.5 2.5 3.5 4.5 5.5 6.5 7.5 8.5 9.5 10.5 11.5
= –1617.7; =48090.42.
Λ( ) = (−1617.7 + 48090.42) .
Hence, we obtain
( ) ( )
( ( )= )= , = 0,1,2, …. [7.14]
!
[ ( ) ( )] [ ( ) ( )]
( ( + ℎ) − ( ) = ) = . [7.15]
!
( + ℎ) − ( ) = Δ( ; ℎ) = ℎ + + . [7.16]
Generalizations of Poisson Process in the Modeling of Random Processes 109
D( ; ℎ) = ℎ( + + ). [7.17]
EXAMPLE 7.1.–
It means that the average predicted number of the road accidents between
June 1, 2019 and September 30, 2019 will be about 9,812 accidents with a
standard deviation of about 99. For example, the probability that the number
of accidents in this time interval will be not greater than d = 10010 and not
less than c = 9611 is
= (9611 ≤ ( + ℎ) − ( ) ≤ 10010) ≅
, ,
≅∑ .
!
( )= + + ⋯+ ( ), ≥0 [7.19]
Let { ( ): ≥ 0} be an NCPP.
If ( ) < ∞, then
1) [ ( )] = Λ( ) ( ) [7.20]
2) [ ( )] = Λ( ) ( ), [7.21]
where
Λ( ) = [ ( )] = ( ) .
[ ( )] = [ ( ( )| ( ))]
we have
[ ( ( )| ( ))] = ( + + ⋯+ ( )) ( ) =
∑∞ (( + + ⋯+ ) ( ( )=
) = ∑∞ ( ) ( ( )= )
= ( ) ( ) = Λ( ) ( )
[ ( )] = [ ( ( )| ( ))] + [ ( ( )| ( ))]
Generalizations of Poisson Process in the Modeling of Random Processes 111
we get
[ ( ( )| ( ))] = ( + + ⋯+ ( )) ( ) =
= ∑∞ + + ⋯+ ( ) ( )= ( ( )= )=
= ∑∞ ( + +⋯+ ) ( ( )= )=
= ∑∞ ( ) ( ( )= )= ( ) ( ) = ( )Λ( ),
( ( )| ( ) ] = ( + +⋯+ ( )) ( ) =
= ( ( ) ( )) = ( ( )) ( ) =( ( )) ( ).
Therefore,
[ ( )] = ( )Λ( ) + ( ( )) ( ) = ( )[ ( )−
( ( )) + ( ( )) ] =
= Λ( ) ( ).
COROLLARY 7.1.–
If ( ) < ∞, then
[ ( + ℎ) − ( )] = Δ( ; ℎ) ( ), [7.22]
[ ( + ℎ) − ( )] = Δ( ; ℎ) ( ), [7.23]
where
Δ( ; ℎ) = ( ) . [7.24]
112 Data Analysis and Applications 4
PROPOSITION 7.2.–
( ) ( )
( , )= [ , )( ) +∑ ( ; ) ( ), [7.25]
where
( )
( ) denotes the k-fold convolution of the CDF of the random
variables , i = 1,2,… and
( ( )) ( )
( ; )= , ≥ 0, = 0,1, …, [7.26]
!
Λ( ) = [ ( )] = ( ) . [7.27]
Proof: Using the total probability low, we obtain the CDF of the NCPP:
( , ) = ( ( ) ≤ )= + +⋯+ ( ) ≤ =
=∑ +⋯+ ( ) ≤ | ( )= ( ( )= )=
( ) ( ) ( )
=∑ ( ; ) ( )= [ , )( ) + ∑ ( ; ) ( ).
COROLLARY 7.2.–
It should be noted that the results presented above are known for
homogeneous Poisson processes – equations [7.25] and [7.28] are presented
in paper Di Crescenzo A., Martinucci B. and Zacks S. (2015).
Generalizations of Poisson Process in the Modeling of Random Processes 113
0.12
0.1
0.08
0.06
0.04
0.02
0
0.5 1.5 2.5 3.5 4.5 5.5 6.5 7.5 8.5 9.5 10.5 11.5
Figure 7.2. Frequency of fatalities with respect to the road accident number
1.3
1.28
1.26
1.24
1.22
1.2
1.18
1.16
1.14
1.12
0.5 1.5 2.5 3.5 4.5 5.5 6.5 7.5 8.5 9.5 10.5 11.5
Figure 7.3. Frequency of injured people with respect to the road accident number
= – 0,002301748, = 0,111402156
Hence,
[ ( + ℎ) − ( )] = Δ( ; ℎ) [7.31]
[ ( + ℎ) − ( )] = Δ( ; ℎ) ( + ) [7.32]
[ ( + ℎ) − ( )] = Δ( ; ℎ) ( + ) [7.33]
where
Δ( ; ℎ) = ℎ + + . [7.34]
EXAMPLE 7.2.–
Δ( ; ℎ) = 9812,201695, D( ; ℎ) = 99,05655806.
Generalizations of Poisson Process in the Modeling of Random Processes 115
= −1617,7,
= 48090,42, Δ( ; ℎ) = 9812,202, D( ; ℎ) = 99,06.
= 0,073555264
= ( ) + [ ( )] = +
for = , , we get
= 0,078965641
[ ( + ℎ) − ( )] = Δ( ; ℎ) ( + ) = 27.84
EFN = 721.74.
DFN = 27.84.
In the same way, we calculate the parameters of the model that enables us
to anticipate the number of injured people in the road accidents in Poland in
this period.
116 Data Analysis and Applications 4
0.4
0.3
0.2
0.1
0
0 1 2 3 4 5 6 7
Finally, we get
7.7. Conclusion
The random process theory delivers concepts and theorems that enable us
to construct stochastic models releted to the road accidents. The processes
with independent increaments are the most appropriate for modeling the
number of accident in a specified time interval. A crucial role in the model
construction plays a non-homogeneous Poisson process. The presented
models enable us to anticipate the number of accidents in a certain period of
time and their consequences. The identification of almost real parameters
was possible thanks to the statistical data derived from the police report
(Symon E. 2018, Symon E. 2019). The non-homogeneous compound
Poisson processes as the models of the road accident consequences enable us
Generalizations of Poisson Process in the Modeling of Random Processes 117
7.8. References
8.1. Introduction
Over the last decades, the design of large-scale and complex technological
systems has become important mainly due to the rapid development of
technology and the increasing demand of various critical applications. The
reliability and the availability of these systems are of great importance since
their deterioration and/or failure may lead to serious economic and social
losses [AGH 16]. Thus, in order to improve the operation of such a system
and increase its availability and reliability, redundancy can be introduced. One
of the most commonly used types of redundancy is the standby redundancy.
In a standby redundant system, apart from the operational units, there is a
number of standby units as backups, in order to replace any component after a
failure. The switching process from the failed unit to the standby unit is
usually assigned to an automated mechanism. It is assumed that this
mechanism can fail with a positive probability [WAN 06, GOE 85, PLA 09].
The automated restoration mechanism usually fails due to frequent use, which
results in aging and degradation, or due to the long period within which it is
in standby mode. In the latter case, usually called imperfect switch, the
switching process is operated manually [DUG 89, TRI 01]. The manual
switch restores the system to an operational state again. However, more time
is needed for manual than for automatic switch. After an automated
mechanism failure, either maintenance or replacement actions are initiated.
E O = {O, D1 , D2 , D3 , I0 , I1 , I2 , I3 , m, M, F }.
As far as the standby unit is concerned, its operational states can be denoted
by:
– OS : if the standby unit is in its perfect state;
– DiS : if the standby unit is in the ith deterioration level (i = 1, 2, 3).
In order to denote the state of the system, a pair (i, j) is used, where
i, j ∈ E denote the condition of the primary and the supplementary unit,
respectively. Initially, the system starts to operate in its perfect state, i.e. the
primary and the standby units are both in their fully operational state
(O, OS ). From the perfect state, the system may enter one of the next
deterioration levels (Di , OS ), i = 1, 2, 3 or state (F, O) if system control is
switched automatically from the failed unit to the standby unit, or even state
(F, M S F ) if the automated switch mechanism experiences a failure and the
switching process is performed manually. Moreover, from state (O, OS ), the
system may enter the inspection state (I0 , OS ). From each of the deterioration
states D(i, j), i = 1, 2, 3, j ∈ {OS , D1S , D2S , D3S , m, M, F }, the operational
unit may enter an inspection state (Ii , j), respectively, due to the assumption
that during the inspection of the operating unit, the state of the standby unit
does not change. In case unit deterioration occurs prior to inspection, the unit
enters one of the next deterioration states (depending on the level of
deterioration) or even the failure state in case of a sudden failure, which can
occur mainly due to external factors. If after the inspection the operating unit
is detected in its perfect state, or in the first deterioration state (D1 ), no
maintenance action takes place. However, if the unit is either in the second
(D2 ) or in the third (D3 ) deterioration level, minimal or major maintenance
action is triggered, respectively, and the system control is switched to the
standby unit, if this is allowed. Either minimal or major maintenance can be
perfect, imperfect or failed. A perfect minimal maintenance restores the unit
to its previous deterioration level (from state D2 to state D1 ), though an
imperfect minimal maintenance restores the unit to the same deterioration
level. Failed minimal maintenance leads the unit to a worse deterioration level
or even to a total failure (from state D2 to state D3 or F ). Perfect major
maintenance restores the unit to its perfect state (as good as new), though
imperfect major maintenance can restore the system to any of the previous
deterioration states (D1 or D2 ) or even to the same deterioration level (D3 ),
124 Data Analysis and Applications 4
with different rates. Failed major maintenance leads the unit to the total
failure state (F ). The resulting system consists of 145 states in total. All
states of the system as well as all possible transitions are presented in Table
A.5 in the Appendix (section 8.7).
Exponential distributions are assumed for all the sojourn times. Note that
the failure rates for the deterioration process are state-dependent, i.e. the higher
the deterioration level, the more probable it is for the system to enter a worse
deterioration level [AGH 16]. Additionally, it is assumed that the inspection
duration is negligible compared to any other sojourn time; thus, there is not
enough time for any other transition to occur during inspection [XIE 05]. The
transition rates among unit possible states are presented in Table 8.1. Thus,
the evolution of the system in time is described by a Markov process Z =
{Z(t), t ≥ 0}.
Rate Description Rate Description
λ1 Deterioration rate from O to D1 μI Inspection response rate
λ12 Deterioration rate from O to D2 λm Minimal maintenance rate
λ13 Deterioration rate from O to D3 λIm Imperfect minimal maintenance rate
λf 1 Sudden failure rate from O to F λf m Failed minimal maintenance rate
λ2 Deterioration rate from D1 to D2 λF m Failure rate due to minimal maintenance failure
λ22 Deterioration rate from D1 to D3 λM Major maintenance rate
λf 2 Sudden failure rate from D1 to F λIM 1 Imperfect major maintenance rate of level 1
λ3 Deterioration rate from D2 to D3 λIM 2 Imperfect major maintenance rate of level 2
λf 3 Sudden failure rate from D2 to F λIM 3 Imperfect major maintenance rate of level 3
λ4 Deterioration rate from D3 to F λF M Failure rate due to major maintenance failure
λIN Inspection rate λR Repair rate
Initially, the transient phase of the system is studied. However, since such
systems are designed to operate continuously in time, its asymptotic behavior,
in terms of dependability and performance, is also studied.
where CD is the cost per unit of downtime time that occurs when the system
is in a non-operational state. Thereafter, wD (i, j) is used for evaluating the
total expected downtime cost per unit time. In particular, let g(Z(t)) be the
downtime reward rate at time t:
g(Z(t)) = wD (i, j)I{Z(t)=(i,j)} [8.5]
(i,j)∈E ′
= wD (i, j)P r(Z(t) = (i, j))
(i,j)∈E ′
= wD (i, j) · π(i.j) (t) [8.6]
(i,j)∈E ′
where π(i.j) (t) is the probability of state (i, j) ∈ E ′ at time t and can be
obtained using [8.2].
wA (i, j) =
⎧
⎪
⎪ CI if in state (i, j), a unit is under inspection
⎪ C if in state (i, j), a unit is under minimal maintenance
⎪
⎪
⎪ m
CM if in state (i, j), a unit is under major maintenance
⎨
[8.8]
⎪
⎪ CR if in state (i, j), a unit is under repair
C if in state (i, j), the switched mechanism should be replaced
⎪
S
⎪
⎪
⎪
0 else
where CI is the cost per unit time for inspecting a unit, Cm is the cost per unit
time for implementing minimal maintenance, CM is the cost per unit time for
implementing major maintenance, CR is the cost per unit time for unit
repair/replacement, and while CS is the cost for replacing the switch
mechanism after a failure. Depending on their nature, let
128 Data Analysis and Applications 4
and then the total expected action cost can be defined similarly to equation
[8.6]. However, we will alternatively use equation [8.10] to define and evaluate
this cost:
T EOC = wD (i, j)π(i,j) + wA (i, j)π(i,j) [8.15]
(i,j)∈E ′ (i,j)∈E ′
problems for the dependability and performance measure that use the
expressions derived in equations [8.14] and [8.15] as objective functions are
formulated. However, once the maintenance schedule that optimizes
asymptotic availability and/or performance is derived, it would be
implemented in the considered system from the beginning of its lifetime.
Thus, the effects of the aforementioned optimal maintenance schedules on the
system’s transient behavior could also be observed.
max AV [8.16]
s.t. λLB UB
IN ≤ λIN ≤ λIN
Note that any inspection policy selected to be adopted for the system
would have totally different effects on availability and overall cost. More
specifically, performing inspections very often benefits the total expected
operational cost, since by inspection and thereafter maintenance, the system is
prevented from failures which incur a considerably high cost. However, since
inspection states are considered as down states, such an inspection policy will
provide lower availability. Contrarily, when inspection is performed less
often, system availability is benefited because the system does not enter down
states often. However, in the latter case, an increased overall cost would be
achieved since, when delaying inspection and hence maintenance, a total
failure that incurs high cost is more probable for the system. Consequently,
the system designer should take into account these features in order to
schedule the optimal maintenance policy that simultaneously optimizes both
measures. This is usually achieved by formulating and solving multi-objective
optimization problems. In this chapter, the multi-objective optimization
problems are solved as an optimization problem with constraints. More
specifically, the most important measure is considered as the objective
function and the remaining measure participates in the optimization problem
as a constraint.
The proposed model can be further examined and illustrated through some
numerical results based on the experimental data provided in Table 8.2.
Initially, let us set the probability of perfect switch among units in the case
of maintenance or in the case of a total unit failure to be c = 0.95. Transient
availability and total operational cost behavior for a time horizon of 10,000
hours are presented in Figures 8.1 and 8.2, respectively. As expected, system
availability decreases with time although the overall cost increases, since, in
the long run, the system will eventually experience failures. Additionally, the
convergence of AV (t) and T EOC(t) to AV and T EOC, respectively,
indicating how fast the system reaches steady state is also depicted in Figures
8.1 and 8.2. As it can be observed, the system needs about 4,000 hours to
reach steady state. This time interval is large enough, highlighting thus the
necessity of transient analysis.
Dependability and Performance Analysis 133
Figure 8.5. Transient availability with respect to the inspection rate λIN
Since, one of the innovative aspects of the proposed model for the system
under consideration is the imperfect switch among its units, it is also
interesting to examine how the time needed to switch system control
manually, in the case of a failure on the automated mechanism, affects the
dependability and performance measures. This can be examined through the
manual switch rate. In Figure 8.7, we observe that the increase of the manual
switch rate results in increasing system availability, as expected. This is
obvious, since in this case, the system spends less time in the non-operational
state of manual switch. However, b does not seem to significantly affect the
T EOC(t) indicator. This is more or less reasonable because entering the
manual switch state affects the cost (by triggering automation mechanism
replacement cost) but the sojourn time in this state has no effect on the cost.
Nevertheless, a slight cost improvement can be obtained when b increases.
138 Data Analysis and Applications 4
The effect of the time needed for a unit to be restored to its previous
deterioration level through minimal maintenance on dependability and
performance is also examined in this section, through the minimal
maintenance rate λm , which represents the rate of minimal maintenance
completion. As expected, when minimal maintenance lasts a short time,
transient availability increases since the system is restored faster to an
operational state. However, similarly to the previous case, the total expected
cost does not seem to be greatly affected by the duration of minimal
maintenance rather than the fact that the system enters minimal maintenance
state, which incurs cost, although the reduced minimal maintenance duration
slightly improves the overall cost.
The effects of the time needed for a unit to be restored to its fully
operational state through major maintenance on dependability and
performance is also examined through the major maintenance rate λM .
Similarly to the minimal maintenance rate and due to exactly the same reason,
transient system availability increases with the increase of λM , as shown in
Dependability and Performance Analysis 139
success probability c. The explanation lies on the fact that frequent inspection
triggers maintenance actions, hence avoiding or delaying operating unit
failures that incur high cost. Finally, for the multi-objective optimization
problem (see Table 8.4), we observe that the optimal inspection policy λ∗
mostly depends on the objective function. Hence, λ∗ is closely related to the
optimal solution of problem [8.18]. However, there are instances of problem
[8.19], in particular when higher levels of availability are desired
(AV0 = 0.999), for which the optimal inspection policy is slightly different
from the optimal inspection schedule for problem [8.18].
c λ∗ (h−1 ) AV ∗ λ∗ (h−1 ) T EOC ∗
0.900 0.001 0.999650 0.1 20.0075
0.950 0.001 0.999681 0.1 19.9430
0.990 0.001 0.999706 0.1 19.9108
0.995 0.001 0.999709 0.1 19.8850
145
policies from the multi-objective problem
146
Data Analysis and Applications 4
Figure 8.17. Transient total operational cost for inspection
policies from the multi-objective problem
Dependability and Performance Analysis 147
8.7. Appendix
Table 8.5. States and possible transitions for the proposed model
152 Data Analysis and Applications 4
8.8. References
[AGH 16] AGHEZZAF E.-H., K HATAB A., L E TAM P., “Optimizing production and imperfect
preventive maintenance plannings integration in failure-prone manufacturing systems”,
Reliability Engineering & System Safety, vol. 145, pp. 190–198, 2016.
[AMA 06] A MARI S.V., M C L AUGHLIN L., P HAM H., “Cost-effective condition-based
maintenance using Markov decision processes”, RAMS’06: Annual Reliability and
Maintainability Symposium, pp. 464–469, 23–26 January, 2006.
[CHE 02] C HEN D., T RIVEDI K.S., “Closed-form analytical results for condition-based
maintenance”, Reliability Engineering and System Safety, vol. 76, no. 1, pp. 43–51, 2002.
[CHE 03] C HEN D., C AO Y., T RIVEDI K.S. et al., “Preventive maintenance of multi-
state system with phase-type failure time distribution and non-zero inspection time”,
International Journal Reliability of Quality and Safety Engineering, vol. 10, no. 3,
pp. 323–344, 2003.
[DIN 15] D ING S.H., K AMARUDDIN S., “Maintenance policy optimization-literature review
and directions”, The International Journal of Advanced Manufacturing Technology, vol. 76,
no. 5, pp. 1263–1283, 2015.
[DUG 89] D UGAN J.B., T RIVEDI K.S., “Coverage modelling for dependability analysis of
fault-tolerant systems”, IEEE Transactions on Computers, vol. 38, no. 6, pp. 775–787,
1989.
[GOE 85] G OEL L.R., G UPTA R., S INKH S.K., “Cost analysis of a two-unit priority standby
system with imperfect switch and arbitrary distributions”, Microelectronics and Reliability,
vol. 25, no. 1, pp. 65–69, 1985.
[KIJ 97] K IJIMA M., Markov Processes for Stochastic Modeling, CRC Press, Boca Raton, FL,
USA, 1997.
[KOU 17] KOUTRAS V.P., M ALEFAKI S., P LATIS A.N., “Optimization of the dependability
and performance measures of a generic model for multi-state deteriorating systems under
maintenance”, Reliability Engineering and System Safety, vol. 166, pp. 73–86, 2017.
[LAP 06] L APA C.M.F., P EREIRA C.M.N.A, DE BARROS M.P., “A model for preventive
maintenance planning by genetic algorithms based in cost and reliability”, Reliability
Engineering and System Safety, vol. 91, no. 2, pp. 233–240, 2006.
[LIS 07] L ISNIANSKI A., D ING Y., F RENKEL I. et al., “Maintenance optimization for
multi-state aging systems”, Proceedings of the 5th International Conference on
Mathematical Methods in Reliability, Methodology and Practice, Glasgow, United
Kingdom, 2007.
[LIS 10] L ISNIANSKI A., F RENKEL I., D ING Y., Multi-state System Reliability Analysis and
Optimization for Engineers and Industrial Managers, Springer, London, 2010.
[MAL 17] M ALEFAKI S., KOUTRAS V.P., P LATIS A.N., “Optimizing availability and
performance of a two-unit redundant multi-state deteriorating system”, Recent Advances
in Multi-state Reliability, pp. 71–105, Springer, Berlin, 2017.
Dependability and Performance Analysis 153
[NAG 06] NAGA P., R AO S., NAIKAN V.N.A., “A condition-based preventive maintenance
policy for Markov deteriorating systems”, International Journal of Performability
Modeling, vol. 2, no. 2, pp. 175–189, 2006.
[NAT 11] NATVIG B., “Probabilistic modeling of monitoring and maintenance”, in Multi-state
Systems Reliability Theory with Applications, John Wiley & Sons, Inc., Chichester, UK,
2011.
[PLA 09] P LATIS A., D ROSAKIS E., “Coverage modeling and optimal maintenance frequency
of an automated restoration mechanism”, IEEE Transactions on Reliability, vol. 58, no. 3,
pp. 470–475, 2009.
[SAD 05] S ADEK A., L IMNIOS N., “Nonparametric estimation of reliability and survival
function for continuous-time finite Markov processes”, Journal of Statistical Planning and
Inference, vol. 133, no. 1, pp. 1–21, 2005.
[SHE 15] S HEU S.-H., C HANG C.-C., C HEN Y.-L. et al., “Optimal preventive maintenance
and repair policies for multi-state systems”, Reliability Engineering and System Safety,
vol. 140, pp. 78–87, 2015.
[THE 12] T HEIN S., C HANG Y.S., M AKATSORIS C., “A study of condition based preventive
maintenance model for repairable multi-stage deteriorating system”, International Journal
of Advanced Logistics, vol. 1, no. 1, pp. 83–102, 2012.
[TRI 01] T RIVEDI K.S., Probability and Statistics with Reliability, Queuing, and Computer
Science Applications, John Wiley & Sons, New York, USA, 2001.
[WAN 06] WANG K.-H., D ONG W.-L., K E J.-B., “Comparison of reliability and the
availability between four systems with warm standby components and standby switching
failures”, Applied Mathematics and Computation, vol. 183, no. 2, pp. 1310–1322, 2006.
[XIE 05] X IE X., Y IGUANG H., T RIVEDI K.S., “Analysis of a two level software
rejuvenation policy”, Reliability Engineering and System Safety, vol. 87, no. 1, pp. 13–22,
2005.
9
Economical time series usually include trends, and it is important to capture trends adequately.
In many cases, trends show repeated up-and-down behavior. In this study, estimation problems
on local maximum and minimum values included in such time series are considered by assuming
that the trend is piecewise linear. However, it is not clear whether estimation is meaningful or not,
unless the property of time series is clarified. The first purpose of this chapter is to propose two
kinds of models for a time series whose trend is piecewise linear. One is a trend stationary model,
and the other is a random walk type model. Proposed models provide a basis for discussion
about appropriateness of estimation and prediction of local maximum and minimum values in the
trend. Simulation studies suggest that estimation and prediction is meaningless in some cases
when a time series includes a stochastic trend even though the mean value function has local
maximum and minimum values. The second purpose of this chapter is to propose a procedure for
estimating the piecewise linear trend. Local maximum and minimum values are obtained directly
from the estimated trend. The proposed method is based on a piecewise linear regression and an
application of a fuzzy trend model.
9.1. Introduction
a trend. Estimation and prediction of local maximum and minimum values are
important topics. However, it is not clear whether these topics are meaningful
or not, unless the properties or models of a time series are clarified. The first
purpose of this chapter is to propose two kinds of models for a time series
whose trend is continuous piecewise linear. One is a trend stationary model,
and the other is a random walk type model. Problems on the piecewise linear
trend are discussed by [KIM 09] and [TIB 14], among others. The piecewise
linear trend can be used for approximation in many cases, although it has a
restricted form. One advantage of the assumption of the piecewise linear trend
is that peaks and troughs are obtained directly from the estimated trend.
9.2. Models
Let {xn |n = 1, ..., N } be an observed time series, whose trend has local
maximum and minimum values. We consider two models for {xn }. The first is
a mean or trend stationary model, and the second is a random walk type model.
9.2.1. Model 1
xn = µn + v n , [9.1]
Models for Time Series Whose Trend Has Local Maximum and Minimum Values 157
where {vn } is a zero mean stationary process, and {µn } is the mean value
function of {xn } given by the recursion:
dn = u(s1n ). [9.5]
The series {dn } is stochastic, but we consider this under the condition that
{dn } is given. The process {u(k)} can also be stochastic, as shown in an
example in section 9.3. In this model, the mean value function µn is the trend,
and the trend is continuous piecewise linear. The assumption [9.3] means that
the local maximum and minimum values appear by turns. In other words, the
trend shows up-and-down behavior. Note that the assumption [9.3] can be
weakened or removed.
The first component of the state variable, s1n , means the number of local
linear trends. That is, µn is on the s1n -th straight line. The second component
s2n is the time point when the s1n -th local linear trend begins. The time point
of the local maximum or minimum value is given by
and we have T0 = 0.
158 Data Analysis and Applications 4
9.2.2. Model 2
where dn is the series of constants defined in model 1, and {en } is a zero mean
i.i.d. process with the variance σ 2 . We assume that the initial value x0 is zero
for simplicity.
Figure 9.1 for b = 1, 2, ..., 9 and m = 10, 15, 20, ..., 40. The furthest left
curve is for m = 10 and the furthest right is for m = 40 in each graph.
Models for Time Series Whose Trend Has Local Maximum and Minimum Values 159
0 0 0
20 40 60 20 40 60 20 40 60
0 0 0
20 40 60 20 40 60 20 40 60
0 0 0
20 40 60 20 40 60 20 40 60
9.3. Simulation
In model 2, we assume that the white noise {en } is Gaussian with the
variance 0.42 . An example of a time series obtained by model 2 is shown in
Figure 9.3.
160 Data Analysis and Applications 4
The mean value function shown by the bold piecewise linear line in
Figure 9.3 is the same as the trend in Figure 9.2.
5
-1
0 50 100 150 200
On the other hand, validity of estimation for random walk type models
is weak, since there is usually no direct relationship between up-and-down
behavior of a time series and local maximum and minimum values of the mean
value function, as shown in Figure 9.3. The estimated peaks and troughs have
no meaning, unless the variance of the white noise is relatively small compared
with the range of fluctuation of dn . As a result, it is difficult to show validity
of prediction methods for peaks and troughs in some cases when a stochastic
trend is included in a time series. Another problem is whether a peak detection
method for a time series from model 2 is useful or not in a practical analysis
apart from the theoretical aspect.
Models for Time Series Whose Trend Has Local Maximum and Minimum Values 161
-1
0 50 100 150 200
Identification procedure
The fuzzy trend model used here is a one-parameter model for a scalar time
series based on the fuzzy if-then rule:
Rℓ : If n is Aℓ , then µn (ℓ) = αℓ ,
Let {α̂ℓ |ℓ = 1, ..., L} be the estimate of the latent process {αℓ } in the fuzzy
trend model. In step 2, peaks and troughs are detected by checking the change
of α̂ℓ . We judge that the change occurs if |α̂ℓ − α̂ℓ−1 | > tα Sα , where tα is a
given positive constant and Sα is the sample standard deviation of {α̂ℓ }. Then,
the time points of nodes in the piecewise linear function are pre-estimated.
The piecewise linear regression is easy, if the time points of nodes in the
piecewise linear function are given (see [DRA 98], [GAL 73]).
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
20 40 60 80 100
1 Estimated trend
-1
0 50 100 150 200
4.5
3.5
2.5
1.5
Pre-estimated trend
0.5
True trend
-0.5
0 50 100 150 200
9.5. Conclusion
We proposed two models whose mean value functions are piecewise linear.
We can discuss the validity of estimation methods for peaks and troughs based
on the proposed models. As a result, it is found that validity is doubtful in some
cases when stochastic trends are included. Thus, it is important to investigate
the property of an observed time series in practical analysis. This is a problem
of a statistical hypothesis test or model selection.
After peaks and troughs are estimated, inference on {dn } and {u(k)}, from
which a trend is generated, becomes possible. For example, prediction of the
time point and height of the next peak can be considered.
Models for Time Series Whose Trend Has Local Maximum and Minimum Values 165
9.6. References
[DRA 98] D RAPER N.R., S MITH H., Applied Regression Analysis, 3rd edition, Wiley, New
York, 1998.
[GAL 73] G ALLANT A.R., F ULLER W.A., “Fitting segmented polynomial regression models
whose join points have to be estimated”, JASA, vol. 68, no. 341, pp. 144–147, 1973.
[KIM 09] K IM S.-J., KOH K., B OYD S., G ORINEVSKY D., “1 trend filtering”, SIAM Review,
vol. 51, no. 2, pp. 339–360, 2009.
[ŞEN 17] Ş EN Z., “Temporal trend analysis”, in Innovative Trend Methodologies in Science
and Engineering, Ş EN Z. (ed.), pp. 133–174, Springer, Berlin, 2017.
[TIB 14] T IBSHIRANI R.J., “Adaptive piecewise polynomial estimation via trend filtering”,
Annals of Statistics, vol. 42, no. 1, pp. 285–323, 2014.
[WAT 15] WATANABE E., WATANABE N., “Weighted multivariate fuzzy trend model for
seasonal time series”, in Stochastic Modeling, Data Analysis and Statistical Applications,
F ILUS L. et al. (eds), pp. 443–450, ISAST, 2015.
10
The basic Kriging model assumes a Gaussian distribution with stationary mean and stationary
variance. In such a setting, the joint distribution of the spatial process is characterized by the
common variance and the correlation matrix or, equivalently, by the common variance and the
variogram matrix. In this chapter, we will discuss in detail the option to actually use the variogram
as a parameterization.
10.1. Introduction
Γ = σ 2 (11′ − R).
1
0≤ Var (Yi − Yj ) = γij = γ(d(xi , xj )) = σ 2 (1 − ρij ) ≤ σ 2 .
2
The increasing function γ : R+ → R+ is assumed to be continuous for
every possible value. As it is bounded at +∞, the general shape is as shown in
Figure 10.1.
2) The parameters in the Krige’s universal model are unrestricted values of
μ ∈ R, σ 2 > 0 and restricted values of R that are usually estimated over a
suitable parametric model.
3) Krige’s idea is to predict the value Y0 = Y (x0 ) at an untried location x0
with the conditional expectation based on a plug-in estimate of the parameters.
If I = {1, . . . , n} and the locations in the model are x0 , x1 , . . . , xn , the
regression value is
−1 ΣI,I ΣI,0
Y0 − μ = Σ0,I ΣI,I (YI − μ1I ), with Σ = .
Σ0,I σ02
170 Data Analysis and Applications 4
The set of data that give the same prediction is an affine plane in Rn . The
variance of the prediction is σ02 − Σ0,I Σ−1
I,I ΣI,0 .
We note that Γ = 0 if, and only if, R = 11′ , such an extreme case being
always excluded in the following. In fact, in most cases, we will assume
det R = 0.
The difference between the covariance matrix and the variogram matrix is
a matrix of rank 1, Γ + Σ = σ 2 11′ . Moreover, let us remark that
1 ′ 1
11 = (Γ + Σ) [10.2]
n nσ 2
is the orthogonal projector on the space of constant vectors span [10.1].
n2 σ 2 α2 = x′ Γx + x′ Σx .
because of condition 3.
(σ 2 , Γ) ↔ (σ 2 , R) ↔ Σ
and
2) The inverse is
2
(σ , Γ) Γ ∈ V, sup x′ Γx x′ 1 = 1 ≤ σ 2 ∋ (σ 2 , Γ)
→ σ 2 11′ − Γ = Σ ∈ S=
How to Model the Covariance Structure in a Spatial Framework 173
The two equations presented above, both based on the definition of the
variogram matrix as Γ = σ 2 (11′ − R), provide a simple connection between
the parameterization based on the covariance matrix Σ and the
parameterization based on the couple σ 2 and Γ. However, we want to spell
out the computation of another key statistical parameter, namely the
concentration matrix Σ−1 . We begin by recalling a well-known equation in
matrix algebra [PRE 96]. We review the result in detail as we need an exact
statement of the conditions under which it is true in our case.
and the statement about the determinant follows. The inversion formula is
directly checked.
P ROOF.–
1) n = Tr R = nj=1 λj . From det R = n
j=1 λj , as the arithmetic mean
is larger than the geometric mean:
⎛ ⎞1
n n
n
j=1 λj 1
1= ≥⎝ λj ⎠ = (det R) n ,
n
j=1
with equality if, and only if, the λj ’s are all equal, hence equal to 1, which
happens if R = In .
2) The geometric mean is larger or equal than the harmonic mean, hence
⎛ ⎞1 ⎛ ⎞−1
n
n n
1
1 ≥ (det R) = ⎝ n λj ⎠ ≥ n ⎝ λ−1
j
⎠ ,
j=1 j=1
n n
1=1R ′ −1
1= λ−1 ′
j (1 uj )
2
=n 2
(λj )−1 θj ,
j=1 j=1
How to Model the Covariance Structure in a Spatial Framework 175
n
where θj = (1′ u)2 /n2 ≥ 0 and j=1 θj = 1. From the convexity of λ →
λ−1 , we obtain
⎛ ⎞−1
n n
1 = n2 (λj )−1 θj ≥ n2 ⎝ λ j θj ⎠ ,
j=1 j=1
and
σ −2 − 1′ Σ−1 1 = σ −2 (1 − 1′ R1) = 0 ,
and
dH (Γ → y ′ (11′ − σ −2 Γ)−1 y)
= σ 2 Tr (σ 2 11′ − Γ)−1 yy ′ (σ 2 11′ − Γ)−1 H .
How to Model the Covariance Structure in a Spatial Framework 177
We now change our point of view to consider the same problem from a
different angle suggested by the observation that the variogram does not
change if we change the general mean μ. In fact, we can associate the
variogram with the state space description of the Gaussian vector.
P ROOF.–
1) If Γ = σ 2 (11′ − R) is the variogram matrix of Σ = σ 2 R, then from
equation [10.5] we have
′
1 1 ′
Σ0 = σ 2
I − 11′ R I − 11 ,
n n
178 Data Analysis and Applications 4
which is indeed positive definite. Let us show that the diagonal elements of Σ0
are constant:
2 ′ 1 ′ ′ 1 ′
(Σ0 )ii = σ ei I − 11 R I − 11 ei
n n
′
2 1 1
= σ ei − 1 R ei − 1
n n
2 ′ 2 1 ′
= σ ei Rei − ei R1 + 2 1 R1
n n
1 ′
= σ2 1 R1 − 1
n2
P ROOF.–
1 ′ ′ 1 ′ 1 ′ ′ 1 ′
− I − 11 Γ I − 11 = I − 11 Σ I − 11 .
n n n n
In the same way, we could suggest the simulation of a random variable with
variogram matrix Γ by the generation of Nn (0, Σ0 ) data in span (1)⊥ .
10.6. Elliptope
1 − x2 − y 2 + 2cxy ≥ c2 .
It is the same for other sections (see Figures 10.2 and 10.3).
Going forward with the discussion of our example, the volume is easily
computed and the uniform a priori is defined. Simulation is feasible for
example by the rejection method. Another option is to write R = A′ A, where
the columns of A are unit vectors. This gives another possible a priori starting
from independent unit vectors. Simulation is feasible, for example starting
with independent standard Gaussians.
such that
⎡ ⎤
t211 t11 t12 t11 t13
A = T ′ T = [ti · tj ]ij = ⎣t11 t12 t212 + t222 t12 t13 + t22 t23 ⎦ .
t11 t13 t12 t13 + t22 t23 t213 + t223 + t233
Moreover, t11 t22 t33 = 0 ⇔ T is unique and invertible if, and only if, A is
invertible. It is an identifiable parameterization for non-singular matrices.
182 Data Analysis and Applications 4
It follows:
⎡ ⎤
1 1 − t212 − t213 t12 1 − t212 − t213 t13
⎢
R=⎢ 2 2
⎣ 1 − t12 − t13 t12 1 t12 t13 + 1 − t223 t23 ⎦
1 − t212 − t213 t13 t12 t13 + 1 − t223 t23 1
and
10.7. Conclusion
Over the last few decades, Kriging models have been recommended not
only for the original application, but also for spatial noisy data in general. The
accuracy of this model strongly depends on the detection of the correlation
structure of the responses. In the Bayesian approach, where the posterior
distribution of a prediction Krige’s Y0 given the training set (Y1 , . . . , Yn )
requires less uncertainty as possible on the correlation function, the use of the
variogram as a parameter should be preferred because it does not demand a
parametric approach as the correlation estimation does. In a previous paper
[PIS 16], the authors proved the equivalence between the variogram and
spatial correlation function for stationary and intrinsically stationary
processes. This study has been devoted to the characterization of matrices
which are admissible variograms in the case of first-order stationarity.
10.8. Acknowledgements
This paper was presented at the 4th Stochastic Modeling Techniques and
Data Analysis International Conference, June 1–4, 2016, University of Malta,
Valletta, Malta, with the title Bayes and Krige: Generalities. We thank both
G. Kon Kam King (CCA, Torino) and L. Malagò (RIST, Cluj-Napoca) for
How to Model the Covariance Structure in a Spatial Framework 183
10.9. References
[BAR 00] BARNARD J., M CCULLOCH R., M ENG X.-L., “Modeling covariance matrices in
terms of standard deviations and correlations, with application to shrinkage”, Statist. Sinica,
vol. 10, no. 4, pp. 1281–1311, 2000.
[CHI 12] C HILÈS J.-P., D ELFINER P., Geostatistics: Modeling Spatial Uncertainty, 2nd
edition, John Wiley & Sons Inc., Hoboken, NJ, USA, 2012.
[CRE 93] C RESSIE N.A.C., Statistics for Spatial Data, John Wiley & Sons Inc., Hoboken,
NJ, USA, 1993.
[GAE 10] G AETAN C., G UYON X., Spatial Statistics and Modeling, Translated by Kevin
Bleakley, Springer, New York, USA, 2010.
[GNE 01] G NEITING T., S ASVÁRI Z., S CHLATHER M., “Analogies and correspondences
between variograms and covariance functions”, Adv. in Appl. Probab., vol. 33, no. 3,
pp. 617–630, 2001.
[MAT 62] M ATHERON G., Traité de Géostatistique Appliqué, Éditions Technip, Paris, France,
1962.
[PIS 16] P ISTONE G., V ICARIO G., “A note on semivariogram”, in D I BATTISTA T.,
M ORENO E., R ACUGNO W. (eds), Topics on Methodological and Applied Statistical
Inference, Springer, Cham, Switzerland, 2016.
[PRE 96] P RESS W.H., T EUKOLSKY S.A., V ETTERLING W.T. et al., Numerical Recipes:
The Art of Scientific Computing, Cambridge University Press, Cambridge, NY, USA, 1996.
[ROU 94] ROUSSEEUW P.J., M OLENBERGHS G., “The shape of correlation matrices”, Amer.
Statist., vol. 48, no. 4, pp. 276–279, 1994.
[RAP 07] R APISARDA F., B RIGO D., M ERCURIO F., “Parameterizing correlations: A
geometric interpretation”, IMA J. Manag. Math., vol. 18, no. 1, pp. 55–73, 2007.
[SAS 94] S ASVÁRI Z., Positive Definite and Definitizable Functions, vol. 2, Akademie Verlag,
Berlin, Germany, 1994.
[VIC 16] V ICARIO G., C RAPAROTTA G., P ISTONE G., “Meta-models in computer
experiments: Kriging vs artificial neural networks”, Quality and Reliability Engineering
International, vol. 32, pp. 2055–2065, 2016.
11
This chapter explores a distance that enables us to compare Markovian processes. It shows the
relationship of this distance to the Kullback–Leibler divergence and reveals its stochastic behavior
in terms of the chi-square distribution. The distance enables us to decide whether there is any
discrepancy between two samples of stochastic processes. When a discrepancy exists, the use
of this distance allows us to find the strings where the discrepancy is manifested. We apply the
distance to written texts of European Portuguese coming from two authors: Vieira, 1608 and
Garrett, 1799. In the application, the distance reveals the linguistic configurations that expose
discrepancies between written texts of different genres from the same author. This type of result
could characterize linguistic genres and varieties in the same language.
11.1. Introduction
Chapter written by Jesús Enrique G ARCÍA, Ramin G HOLIZADEH and Verónica Andrea
G ONZÁLEZ -L ÓPEZ.
11.2. Preliminaries
Let (Xt ) be a discrete time (order o < ∞) Markov chain on a finite alphabet
A. Let us call S = Ao the state space and denote the string am am+1 . . . an by
anm , where ai ∈ A, m ≤ i ≤ n. For each a ∈ A and s ∈ S, P (a|s) =
t−1
Prob(Xt = a|Xt−M = s). In a given sample xn1 , coming from the stochastic
process, the number of occurrences of s in the sample xn1 is denoted by Nn (s)
and the number of occurrences of s followed by a in the sample xn1 is denoted
by Nn (s, a). In this way, NNnn(s,a)
(s) is the estimator of P (a|s).
D EFINITION 11.1.– Consider two Markov chains (X1,t ) and (X2,t ), of order o,
with finite alphabet A and state space S = Ao . With sample xnk,1
k
, for k = 1, 2
respectively. For any s ∈ S,
α
Nn1 (s, a)
ds (xn1,1
1
, xn2,1
2
) = Nn1 (s, a) ln
(|A| − 1) ln(n1 + n2 ) Nn1 (s)
a∈A
Nn2 (s, a)
+Nn2 (s, a) ln
Nn2 (s)
Nn1 +n2 (s, a)
−Nn1 +n2 (s, a) ln
Nn1 +n2 (s)
with Nn1 +n2 (s, a) = Nn1 (s, a) + Nn2 (s, a), Nn1 +n2 (s) = Nn1 (s) + Nn2 (s),
where Nn1 and Nn2 are given as usual, computed from the samples xn1,1 1
and
n2
x2,1 respectively, and α being a real and positive value.
Comparison of Stochastic Processes 187
ds (xn1,1
1
, xn2,1
2
) ≤ ds (xn1,1
1
, xn3,1
3
) + ds (xn3,1
3
, xn2,1
2
).
ii) Local behavior of processes laws. If the stochastic laws of (X1,t ) and
(X2,t ) are the same in s, then ds (xn1,1
1
, xn2,1
2
) −→ 0.
min(n1 ,n2 )→∞
Otherwise, ds (xn1,1
1
, xn2,1
2
) −→ ∞.
min(n1 ,n2 )→∞
In the following result, we show the relationship between this distance and
the Kullback–Leibler divergence D(P ||Q), a concept commonly used in this
topic, but that does not constitute a distance. We also show the asymptotic
behavior of the distance. We will use the following notations:
P (a) 2
a∈A P (a) ln( Q(a) ) and χ (P (·), Q(·)) =
D(P (·)||Q(·)) = a∈A
(P (a)−Q(a))2
Q(a) ,for two distributions P and Q defined in the alphabet A, with
Q(a) = 0, a ∈ A. First, we will see how the quantity D(P (·)||Q(·)) behaves
under certain conditions on P and Q. Consider the function f (x) = x ln(x),
near to x = 1, by the Taylor’s expansion we have f (x) = (x − 1)+
(x−1)2
2 + δ(x)(x − 1)2 , where δ(x) = − (x−1)
6t2
for some value t ∈ (x, 1)
(Lagrange’s form). We note that when x → 1, δ(x) → 0. Thus, for two
probability distributions P and Q in A,
P (a) P (a)
P (a) ln = Q(a)f
Q(a) Q(a)
1 (P (a) − Q(a))2
= P (a) − Q(a) +
2 Q(a)
P (a) (P (a) − Q(a))2
+δ ,
Q(a) Q(a)
188 Data Analysis and Applications 4
for a ∈ A,
and
P (a) (P (a)−Q(a))2
D(P (·)||Q(·)) 1 a∈A δ Q(a) Q(a)
2
= + . [11.2]
χ (P (·), Q(·)) 2 χ2 (P (·), Q(·))
P (a) P (a)
If Q(a) → 1, given is positive and small enough, |δ Q(a) | < and
P (a) (P (a)−Q(a))2
a∈A δ Q(a) Q(a) D(P (·)||Q(·)) 1
< , so → .
χ2 (P (·), Q(·)) 2
χ (P (·), Q(·)) 2
P̂ (a)
and when Q(a) → 1,
D(P̂ (·)||Q(·)) 1
→ .
χ2,k (P̂ (·), Q(·)) 2k
If we have two samples of sizes k1 and k2 generated from the law W, with
empirical distribution P̂ (a) = X(a) Y (a)
k1 and Q̂(a) = k2 respectively. We obtain
Comparison of Stochastic Processes 189
(equation [11.1])
P̂ (a) W (a)
+ 1 + 2δ (P̂ (a) − W (a)) −1 . [11.4]
a∈A
Q̂(a) Q̂(a)
W (a) P̂ (a)
Hence, when → 1 and → 1,
Q̂(a) Q̂(a)
D(P̂ (·)||Q̂(·))
1 1
→ 1. [11.5]
2,k1 (P̂ (·), W (·)) 2,k2 (Q̂(·), W (·))
2k1 χ + 2k2 χ
T HEOREM 11.1.– Let (Xk,t ) be a Markov chain of order o, with finite alphabet
A, state space S = Ao and xnk,1
k
a sample of the process for k = 1, 2. Consider
Nnk (s,·) Nn1 +n2 (s,·)
also s ∈ S. If D Nn (s) Nn +n (s) < ∞, for k = 1, 2, then
k 1 2
α
ds (xn1,1
1
, xn2,1
2
)=
(|A| − 1) ln(n1 + n2 )
N (s, ·) Nn1 +n2 (s, ·)
nk
Nnk (s)D .
Nnk (s) Nn1 +n2 (s)
k=1,2
(|A| − 1)
2 ln(n1 + n2 ) ds (xn1,1
1
, xn2,1
2
) ∼d
α
N (s, ·) N
nk n1 +n2 (s, ·)
χ2,Nnk (s) , W (·) + χ2,Nn1 +n2 (s) , W (·) ,
Nnk (s) Nn1 +n2 (s)
k=1,2
190 Data Analysis and Applications 4
Nn1 (s, a)
Nn2 (s, a)
= Nn1 (s, a) ln + Nn2 (s, a) ln
Nn1 (s) Nn2 (s)
a∈A
Nn1 +n2 (s, a)
−(Nn1 (s, a) + Nn2 (s, a)) ln .
Nn1 +n2 (s)
N (s, a)
n1
Nn1 +n2 (s, a)
= Nn1 (s, a) ln − ln
Nn1 (s) Nn1 +n2 (s)
a∈A
Nn2 (s, a) Nn1 +n2 (s, a)
+ Nn2 (s, a) ln − ln
Nn2 (s) Nn1 +n2 (s)
a∈A
Nn (s, a) Nn (s, a) Nn +n (s, a)
= Nnk (s) k
ln k
/ 1 2
Nnk (s) Nnk (s) Nn1 +n2 (s)
k=1,2 a∈A
N (s, ·) Nn1 +n2 (s, ·)
nk
= Nnk (s)D .
Nnk (s) Nn1 +n2 (s)
k=1,2
Nn (s,·)
Nn (s) χ2,Nnk (s) Nnk (s) , W (·)
k k
2 Nnk (s)
k=1,2
Nn1 +n2 (s,·)
χ2,Nn1 +n2 (s) Nn1 +n2 (s) , W (·)
+ .
Nn1 +n2 (s)
Then,
(|A| − 1)
2 ln(n1 + n2 ) ds (xn1,1
1
, xn2,1
2
) ∼d
α
Comparison of Stochastic Processes 191
N (s, ·) N
nk n1 +n2 (s, ·)
χ2,Nnk (s) , W (·) + χ2,Nn1 +n2 (s) , W (·) .
Nnk (s) Nn1 +n2 (s)
k=1,2
We can define
and
As seen in definition 11.1, if the stochastic laws of (X1,t ) and (X2,t ) are
the same in s, then
ds (xn1,1
1
, xn2,1
2
) −→ 0.
min(n1 ,n2 )→∞
In the same way, if the local laws for s are different, then
ds (xn1,1
1
, xn2,1
2
) −→ ∞.
min(n1 ,n2 )→∞
We can see that if dmax is large, smax is exactly the string we want to
recognize, as being relevant in terms of discrepancy, but all the strings with a
large relative value of d will reveal changes on the local laws of the processes
relative to the string. In this application, a value larger than 1 will be considered
significant.
Comparison of Stochastic Processes 193
We note that the comparison is made between the different texts of the same
author. The memory o used in this application is equal to 2.
Other studies in the area show that the strings 2-4, 7-2 and 2-7 (see
Tables 11.3 and 11.4) are volatile configurations of European Portuguese
(from the 16th Century to the 19th Century) – see García et al. (2017)
[GAR 17a]. We can see that this characteristic persists when analyzing the
variability of different written texts by the same author, whether that author is
Vieira or Garrett.
ds (1608c, 1608d) s ds (1608c, 1608s) s ds (1608d, 1608s) s
1.02591 7-6 1.18101 4-7 1.07770 1-7
1.11191 1-6 1.28567 2-3 1.07883 4-4
1.13048 3-6 1.98674 2-7 1.33124 4-7
2.14046 7-2 3.86756 2-4 1.67395 2-4
1.74245 2-7
String Meaning
2–4 A monosyllable without stress followed by a dissyllable with stress on the first syllable
2–7 A monosyllable without stress followed by a paroxytone word
7–2 A paroxytone word followed by a monosyllable without stress
between the written texts in relation to the prosodic construction, for example
P (2|7 − 2) is 0.34777 in the text 1799t (theater) and it goes to 0.15735 in
the written text 1799n (narrative), both texts from Garrett. Moreover, the most
probable choice for the second text, since the string 7-2 has been observed, is
7 (P (7|7 − 2) = 0.3241).
We can define three groups of strings: (i) strings that show discrepancies
between Vieira’s texts but not in the case of Garrett’s texts; (ii) strings that
show discrepancies between Garrett’s texts and not in the case of Vieira’s texts;
and (iii) strings that show discrepancies between texts for each of these authors
(see the detailed description of each group in Table 11.6).
Table 11.6. Strings (see Table 11.3) and meaning of the linguistic
compositions that characterize the variability between the texts of the
same author. We also list the strings (with d > 1) that are common
among the authors; the constructions listed in Table 11.4 are excluded
11.4. Conclusion
11.5. References
[FRO 12] F ROTA S., G ALVES C., V IGÁRIO M. et al., “The phonology of rhythm from
Classical to Modern Portuguese”, Journal of Historical Linguistics, vol. 2.2, pp. 173–207,
2012.
[GAL 10] G ALVES C., DE A NDRADE A.L., FARIA P., Tycho Brahe
Parsed Corpus of Historical Portuguese, December 2010. Available at:
http://www.tycho.iel.unicamp.br/∼tycho/corpus/texts/psd.zip.
[GAL 12] G ALVES A., G ALVES C., G ARCIA J.E. et al., “Context tree selection and linguistic
rhythm retrieval from written texts”, The Annals of Applied Statistics, vol. 6, no. 1, pp. 186–
209, 2012.
[GAR 17a] G ARCÍA J.E., G HOLIZADEH R., G ONZÁLEZ -L ÓPEZ V.A., “Linguistic
compositions highly volatile in Portuguese”, Cadernos de Estudos Lingüísticos, vol. 59,
no. 3, pp. 617–630, 2017.
[GAR 17b] G ARCÍA J.E., G ONZÁLEZ -L ÓPEZ V.A., “Consistent estimation of partition
Markov models”, Entropy, vol. 19, no. 4, p. 160, 2017.
[GAR 15] G ARCÍA J.E., G ONZÁLEZ -L ÓPEZ V.A., “Detecting Regime Changes in Markov
Models”, in M ANCA R., M C C LEAN C., S KIADAS C.H. (eds), New Trends in Stochastic
Modeling and Data Analysis, pp. 103–109, ISAST, Lisbon, 2015.
PART 3
This chapter discusses annual salary enhancements for newly hired employees in companies
taking on new graduates. The decision process for selecting candidates with the best skill sets
seems to be one of the most difficult obstacles to overcome. In particular, the analysis is based on
the Education-for-Labor Elicitation from Companies’ Attitudes towards University Studies project
(ELECTUS), involving 471 enterprises, with 15 or more employees, operating in Lombardy. The
recruiters’ preference analysis was carried out using conjoint analysis (CA). Starting from CA part-
worth utilities, a coefficient of economic revaluation was used to compare different salary profiles
for five job vacancies.
12.1. Introduction
creating a mismatch between education and the labor market. In general, the
mismatch occurs when:
The mismatch in Italy has been well documented by The Italian National
Institute of Statistics (ISTAT), using data from a sample survey on university
graduates’ vocational integration1. In fact, in 2015, the transition into the
labor market was rather difficult for Humanities graduates (with 61.7% of the
bachelor’s degree graduates and 73.4% of the master’s degree graduates being
employed) and Earth Sciences graduates (with 58.6% of the bachelor’s degree
graduates and 76.5% of the master’s degree graduates being employed). On
the other hand, the master’s degree graduates in Defence and Security,
Medicine and Engineering had the highest employment levels (99.4%, 96.5%
and 93.9%, respectively). Moreover, 52.8% of the bachelor’s graduates and
41.9% of the master’s graduates found employment in “non-stable” jobs. Of
course, this information has to be considered in light of the worldwide
economic crisis, which significantly impacted the employment situation in the
labor market in general. Looking at the Italian situation in 2015, while the
overall unemployment rate was at 11.9%, the unemployment rate among
young people (15–24 years old) increased dramatically to 40.3%. Moreover,
unemployment in the service sector hit 16%2.
This chapter concerns labor market inclusion policies for new graduates
and the relationships between enterprises and universities. The study is based
on the multi-center research project, Education-for-Labor Elicitation from
Companies’ Attitudes towards University Studies (ELECTUS) [FAB 15],
which involved several Italian universities. This work has three main
objectives. First, it focuses on the identification of an ideal graduate profile
for several job positions. Second, it seeks recommendations of some broad
skills, universally recognized as “best practice” for graduates. Finally, the
analysis attempts to give a comparative view of the differences between, and
the assessment of, wages and skill sets for new graduates.
1 See: htpp://www.istat.it/en/archive/190700
2 See: http://ec.europa.eu/eurostat/data/database
Conjoint Analysis of Gross Annual Salary Re-evaluation 201
12.2. Methodology
Conjoint analysis (CA) is among the methods that are mostly used to
analyze consumer choices and to assign consumers’ utility, drawn from the
properties of single characteristics of goods, services or, as in this application,
jobs being offered on the market.
Different models that have been used to measure the economic value
derived from CA are described in the literature. In [BUS 04], the monetary
value of a utility unit is computed as the ratio of the difference between its
maximum price and its minimum price, compared to related utilities.
In this work, the conjoint rating response format is used to gather and use
additional information about respondents’ preferences. This preference model
uses a part-worth utility linear function. Part-worth utilities are also assumed
for each level of the various attributes, estimated by using ordinary least
squares (OLS) multiple regression. In this formulation, attention is focused on
a rating scale, opting for a very general preference model used in traditional
CA.
In fact, of all the attribute levels that describe the graduate, the information
contained in the conjoint rating format is exploited by regressing individual
202 Data Analysis and Applications 4
n
Uk = βi xik [12.1]
i=0
where xi is equal to 1 and n is the number of all levels of the attributes which
define the combination of a given good. Each xik variable is a dichotomous
variable, which refers to a specific attribute level. This variable is equal to 1 if
the corresponding attribute level is present in the combination of attributes
that describes the alternative k. Otherwise, that variable will be 0. As a result,
the utility associated with the alternative k (Uk ) is obtained by summing the
terms βi xik over all attribute levels, where βi is the partial change in Uk for
the presence of the attribute level i, with all other variables remaining
constant. In this work, we refer to this piecewise linear function as a
part-worth function model that gives a specific utility value for each level of
the considered attributes, usually referred to as part-worth utility.
Consequently, the number of parameters estimated by assuming the
part-worth specification is greater than what is required by alternative
preference model specifications, such as the vector model form and the ideal
model.
Having chosen the preference model (and the rating scale), a coefficient
of economic valuation is developed for a hypothetical change that may occur
in the combination of attribute levels, as described by Mariani and Mussini
[MAR 13].
Conjoint Analysis of Gross Annual Salary Re-evaluation 203
Ui − Ub
Mi = [12.2]
Ub
where Ui denotes the sum of the utility scores associated with alternative
profile i and Ub (assumed to be different from 0) denotes the sum of the
part-worth utilities associated with the status quo b of the job. Equation [12.2]
indicates whether the status quo b modification gives a loss or a gain. If
Mi = 0, there is no loss or gain in terms of total utility. However, the utility
change arising from an attribute-level modification can be considered more or
less important by respondents. Hence, this change may have a more important
economic impact with respect to a changed utility, which has a similar
intensity but involves a less relevant attribute. As a solution, the relative
importance of the modified attribute is used as a weighting [GAR 17].
The range of the utility values for each attribute from highest to lowest
provides an indicator of how important the attribute is compared to the others.
The larger the utility ranges, the more important is the role that the attributes
play. This applies in the same way to smaller ranges. For any attribute j, the
relative importance can be computed by dividing its utility range by the sum
of all utility ranges as follows:
M Iij = Mi ∗ Ij . [12.4]
204 Data Analysis and Applications 4
Assuming a change in the status quo profile, formula [12.1] will be used to
estimate the variation of the total revenue generated. Given the gross annual
salary (GAS) associated with the status quo profile, the coefficient of
economic valuation (CEV) is expressed as follows:
where Vij denotes the amount of the salary variation. The variation Vij is
obtained by supposing that the job’s monetary attribute will vary in
proportion to the change in total utility. This assumption may seem restrictive.
However, it is possible to argue that the monetary amount asked of an
employer for a job reflects how that user values the combination of attributes
of the job in terms of its utility. Under this hypothesis, it is credible to assess
the economic value of a change in the combination of attributes as a function
of the utility and importance of the modified attribute. In addition, CA serves
to approximate the real structure of preferences, given that only a partial
knowledge of preferences can be known. Therefore, it is possible to use the
CEV as a monetary indicator that approximates the impact of a given utility
change in monetary terms.
The proposed coefficient was then applied to the ELECTUS survey. First,
an ideal profile was obtained by maximizing part-worth utilities. Then,
economic variations on the proposed gross annual salary for graduates were
computed by using the coefficient.
The questionnaire contained two sections: the first concerned the conjoint
experiment for the five job positions and the second contained general
information about the company (demographic data). The five job positions
considered for the new graduates were administration clerk, HR assistant,
3 See: www.sawtoothsoftwaver.com
Conjoint Analysis of Gross Annual Salary Re-evaluation 205
marketing assistant, ICT professional and CRM assistant. Six attributes were
used to specify the candidates’ profile:
– Field of Study with 10 levels (philosophy and literature, educational
sciences, political sciences/sociology, economics, law, statistics, industrial
engineering, mathematics/computer sciences, psychology, foreign languages);
– Degree Mark with 3 levels (low, 66–90, medium, 91–100, high, 101–
110+);
– Degree Level with 2 levels (bachelor’s, master’s);
– English Knowledge with 2 levels (suitable for communication with non-
native Italian speakers, inadequate for communication with non-native Italian
speakers);
– Relevant Work Experience with 4 levels (no experience at all, internship
during or after completion of university studies, intermittent or occasional
employment during university studies, one year or more of regular
employment);
– Willingness to Travel on Business with 3 levels (unwilling to travel on
business, willing to travel on business only for short periods, willing to travel
on business, even for long periods).
After having rated the selected profile and chosen the best one, employers
were asked to propose a gross annual salary for the chosen profile, to measure
WTP [BRE 06].
It is important to recall that, since the definition of the sum of utilities for all
levels of an attribute will be equal to 0, less desirable attributes may generate
negative utilities.
In the application, the part-worth utilities seemed to be similar for all the
attributes, except for Field of Study. This means that other competencies have
some levels that are universally identified as “best practice” for graduates.
Utility scores for the Degree Level variable were very close to 0 for each
position. This means that for the respondents, there was no significant
difference between a bachelor’s and a master’s degree. This was due to the
208 Data Analysis and Applications 4
fact that all the positions analyzed were entry-level roles, not requiring
specialized skills.
Table 12.3 shows the ideal profiles for each job vacancy. As can be noted,
these profiles were similar to one another, except for Field of Study. This
confirms the theory of the existence of some transferable or specialized
competencies.
Table 12.3. Competencies attributes and ideal levels for job vacancies
Economics was the preferred Field of Study for three positions (AC, MKT
and CRM) out of the five. A degree in psychology was desirable for the HR
assistant role, while for a more technical position, such as ICT professional,
the Field of Study with the largest part-worth utility was computer
sciences/mathematics. For this reason, according to the previous definition,
Field of Study is a specialized competence.
The Relevant Work Experience and English Knowledge attributes show that
the best level perceived does not depend on the duties to be performed, which
means they could be considered cross-competencies. After all, it is easy to
imagine that companies would prefer to employ a candidate with one year
or more of regular work experience and who is capable of communicating in
another language.
Two levels are recognized as “best practice” for the attributes Degree Mark
and Willingness to Travel on Business, which could be defined as being nearly
cross-competencies.
Conjoint Analysis of Gross Annual Salary Re-evaluation 209
Finally, since part-worth utilities for the Degree Level variable were very
close to 0 and there was no difference between these levels, this could be
defined as a non-binding attribute.
Table 12.4 shows the percentage of the importance index for the five CAs
computed using equation [12.3]. It is important to remember that higher
values for the index corresponds with a more important competence for the
respondents.
Competence AC HR ICT MKT CRM
Field of Study 53.35% 59.39% 80.54% 54.27% 42.98%
Degree Level 1.75% 1.04% 1.91% 0.54% 3.19%
Degree Mark 11.09% 7.67% 4.98% 7.63% 10.07%
English Knowledge 16.67% 10.55% 3.82% 18.39% 16.32%
Relevant Work Experience 13.75% 9.12% 4.89% 14.86% 9.62%
Willingness to Travel 3.37% 12.23% 3.61% 4.30% 17.83%
For each profile, the attribute with the highest values for the index was Field
of Study. This percentage reached a maximum of 80.54% for ICT professional.
The only other position where this value was not over 50% was CRM assistant.
Another high percentage for this role was Willingness to Travel, which was
nearly 20%.
Third, on the same level, there were Degree Mark, Relevant Work
Experience and Willingness to Travel with an average importance of about
10%. Finally, because all part-worth utilities were close to 0, the index of
importance for Degree Level was very low for each job position.
The last step of the analysis done for this research consisted of the
computation of the CEVs Mij . To obtain these figures, the part-worth utilities
in Table 12.2 were used, combined with the importance indices in Table 12.4.
changes in utility for a unique attribute. In this case study, the attribute field of
study was considered as not fixed. Therefore, the utilities were computed with
other attributes remaining constant. In particular, in this application, the status
quo profile was the worst profile because it minimized total utility. Therefore,
Ub is the sum of the lowest part-worth utilities (plus an intercept) for each
attribute j. This means that all M Iij coefficients and all variations Vij are
positive.
Table 12.5 shows the values of (Mij + 1) × 100 for Field of Study, as
expected each Mij ≥ 0 and Mij = 0 only in correspondence of the Field of
Study with the minimum of part-worth utilities. Comparing the job positions,
ICT professional showed higher variations. This is due to the fact that Ij was
very high: a degree in mathematics or computer sciences was fully
specialized for this position. The biggest Mij is for a degree in mathematics
or computer sciences for ICT professional, in comparison to a graduate in
foreign languages, the former earns twice the gross annual salary.
Finally, for CRM assistant, all wages were very close one another, in
comparison to the worst (philosophy and literature), with the only role that
went over 10% being economics.
With respect to the CEV coefficients, first, the minimum GAS was
computed: in the questionnaire, the entrepreneurs were asked to determine a
possible GAS for the five job vacancies. The CEV shows very interesting
results: for example, looking for economics graduates, the value of GAS
increases to 4809.75 euros for AC, 3920 euros for HR, 4111.2 euros for ICT,
4003.3 euros for MKT and 2995.2 euros for CRM (with respect to the Field of
Study with the minimum part-worth utility for each of the five job vacancies).
12.4. Conclusion
Future research aims to focus attention in two directions. First, the results
of stratified CA based on socio-demographic features of companies
responding in the ELECTUS project could be discussed. Second, it would
clearly be advantageous to extend the Mariani–Mussini economic
re-evaluation coefficient to more than one attribute, in order to jointly analyze
more competencies and possible interactions.
12.5. References
[BAK 12] BAK A., BARTLOMOWICZ T., “Conjoint analysis method and its implementation
in conjoint R package”, Data Analysis Methods and its Applications, pp. 239–248, 2012.
[BRE 06] B REIDERT C., Estimation of Willingness-to-Pay: Theory, Measurement,
Application, Deutscher Universitats-Verlag, Wiesbaden, 2006.
[BUS 04] B USACCA B., C OSTABILE M., A NCARANI F., Prezzo e valore per il cliente.
Tecniche di misurazione e applicazioni manageriali, Etas, Perugia, 2004.
[CAR 78] C ARMONE F., G REEN P.E., JAIN A.K., “Robustness of conjoint analysis: Some
Monte Carlo results”, Journal of Marketing Research, vol. 15, pp. 300–303, 1978.
[CAT 82] C ATTIN P., W ITTINK D., “Commercial use of conjoint analysis: A survey”, Journal
of Marketing, vol. 46, pp. 44–53, 1982.
[DAR 08] DARBY K., BATTE M.T., E RNST S. et al., “Decomposing local: A conjoint
analysis of locally produced foods”, American Journal of Agricultural Economics, vol. 90,
no. 2, pp. 476–486, 2008.
[EUR 08] E UROPEAN C OMMISSION, Commission Staff Working Paper accompanying New
Skills for New Jobs. Anticipating and Matching Labour Market and Skills Needs, Brussels,
p. 868, COM, 2008.
[FAB 15] FABBRIS L., S CIONI M., “Dimensionality of scores obtained with a paired-
comparison tournament system of questionnaire item”, M EERMAN A., K LIEWE T. (eds),
Academic Proceedings of the 2015 University-Industry Interaction Conference: Challenges
and Solutions for Fostering Entrepreneurial Universities and Collaborative Innovation.
Amsterdam, 2015.
[GAR 17] G ARAVAGLIA C., M ARIANI P., How much do consumers value protected
designation of origin certifications? Estimates of willingness to pay for PDO dry-cured
ham in Italy, Agribusiness, 2017.
Conjoint Analysis of Gross Annual Salary Re-evaluation 213
[HU 12] H U W., BATTE M.T., W OODS T. et al., “Consumer preferences for local production
and other value-added label claims for a processed food product”, European Review of
Agricultural Economics, vol. 39, no. 3, pp. 489–510, 2012.
[LOU 00] L OUVIERE J.J., H ENSHER D.A., S WAIT J.D., Stated Choice Methods: Analysis
and Application, Cambridge University Press, 2000.
[MAR 13] M ARIANI P., M USSINI M., “A new coefficient of economic valuation based on
utility scores”, Argumenta Oeconomica, vol. 2, pp. 33–46, 2013.
13
The Healthy Life Years Lost Methodology (HLYL) was presented in two books in the Springer
Series on Demographic Methods and Population Analysis, vol. 45 and 46, titled: “Exploring the
Health State of a Population by Dynamic Modeling Methods” and “Demography and Health Issues:
Population Aging, Mortality and Data Analysis” and was applied in particular chapters of these
books. Here the main part of the methodology, with more details and illustrations, is applied by
developing and extending a life table, important to estimate the healthy life years lost along with
fitting to the health expenditure, in the case of Greece in 2009. The application results are quite
promising and important to support decision-makers and health agencies with a powerful tool to
improve the health expenditure allocation and future predictions.
13.1. Introduction
The percentage of expenditure per capita and age group starts from the first
year of age where an extra cost is related to the newborn. It follows a reduced
cost for the years from 5–9 to 10–15, and then the costs grow steadily with
a sharp exponential growth after 55–59 and 60–64 years of age, reaching a
maximum at 85–85 and 90–94 years of age. An important point is related to
the minimum level of per capita expenditure, at 1.46%, corresponding to the
age group 10–14 years, where the minimum mortality appears.
The per capita and age group estimation should be multiplied by the
population per age group to provide the health expenditure per age group of
the same year. Figure 13.2 shows that the population in 2009 has a maximum
at 30–34 years of age, with close figures for the age groups 35–39 and
40–44 years, followed by a decline. The population of the age group 65–69
years shows an irregular decline.
Methodology for an Optimum Health Expenditure Allocation 217
Figure 13.1. Per capita and age group health expenditure in Greece in
2009. For a color version of this figure, see
www.iste.co.uk/makrides/data4
Figure 13.2. Population per age group in Greece in 2009. For a color
version of this figure, see www.iste.co.uk/makrides/data4
218 Data Analysis and Applications 4
The percentage of the total expenditure per age group starts from the first
year of age where an extra cost is related to the newborn. It follows a reduced
cost for the next years with a minimum at 10–14 years of age, and then the
costs grow steadily, with a sharp exponential growth after 45–49 years of age
reaching a maximum at 75–79 years. Then a steady decline follows, due to
the declining number of people for the following age groups. An important
point is related to the minimum level of per age group expenditure, at 1.95%,
corresponding to the age group 10–14 years, where the minimum mortality
appears (Figure 13.3). This is the also an estimate of the minimum operational
level of the health system of Greece.
Figure 13.4. Births per 1000 women in Greece. For a color version of
this figure, see www.iste.co.uk/makrides/data4
The calculations for Greek Health expenditure for the year 2009 are based
on a table provided in Figure 13.5, where the death probability is included in
the third column, and in the fourth column, the data for the fertility rate are
added after the appropriate transformation. The main calculations are done in
the sixth column, where a correction parameter k is estimated via a regression
analysis. Column seven is set for the minimum expenditure level, column
eight includes the estimated values, and column nine includes the data sets
provided by OECD. The tenth column is set for the estimation of the sum of
squared errors and the calculation of R2 and the standard error se. The main
task is to produce a standard system for the calculation of the health
expenditure allocation per age group for a specific year in a country. Although
the main health expenditure is devoted to high and very high ages, it is fair to
have a correct idea of the allocation per age group, including the expenditure
for the very sensitive young ages and newborns. Following our estimates,
6.9% of the health expenditure for 2009 in Greece is allocated to the
population from 1 to 14 years of age, 28.6% is allocated to the population
from 15 to 49 years of age, including the costs for the period of women’s
fertility, and 21.1% is allocated to people aged 50–64 years. A considerable
cost of 28.1% is allocated to people from 65 to 79 years of age, and the
remaining 15.3% is devoted to the people aged 80+ years. Note that the health
expenditure for the 35 years from 15 to 49 is almost equal for the expenditure
for the 15 years ranging from 65 to 79 years of age. The remaining
220 Data Analysis and Applications 4
expenditure for the very old, 80+, is slightly higher than double the costs
spent for years 1–14. However, the real health expenditure for the very old
includes personal and family spending for health care that is not included in
the main health statistics. Our methodology is based on the estimation of the
health expenditure allocation in a country via the healthy life years lost, a
parameter that is provided by Eurostat for the European Union and from the
World Health Organization (WHO) for all world countries. Both estimates
differ due to the methodology used. The WHO estimates provide an estimate
for the healthy life years lost due to severe causes of disability, whereas the
Eurostat estimates are higher, including severe and moderate disability
causes. Following the WHO and Eurostat estimates, Greece is ranked in
different places. According to Eurostat, Greece is ranked below OECD25
Countries for the healthy life years lost at age 65, with 7.9 years of healthy
age for men and 7.7 for women in 2015 after Eurostat estimates (see Figure
13.6). However, the WHO estimates give 17.28 healthy years for men and
19.89 healthy years for women at 60 years of age, also in 2015. Our estimates
in 2009 for the total population (men and women) in Greece found 17.84
healthy life years at 60 years of age and 14.35 years at 65 years of age. The
life expectancy at 60 and 65 years of age is 23.53 and 19.38 years of age.
13.5. Conclusion
tends to be crucial for achieving satisfaction of the people from the provided
health services. The simple resource allocation followed in the previous
health systems turns to be unsatisfactory in the new era. As economic
resources are limited, the quest for the optimum resource allocation is
extremely important. Our methodology is coming to support this issue and
providing vital information to develop the best strategy.
13.6. References
[ATS 08] ATSALAKIS G., N EZIS D., M ATALLIOTAKIS G. et al., “Forecasting mortality
rate using a neural network with fuzzy inference system”, 4th International Workshop on
Applied Probability (IWAP 2008), Compiègne, France, July 7–10 2008.
[GHE 14] G HEORGHE M., O BULQASIM P., VAN BAAL P., Estimating and Forecasting Costs
of Illness in the Netherlands, 2014. Available at: Netherlands-Hospital-o21731-kvz-notitie-
2014-2-report-co-interpolation-and-extrapolation.
[OEC 13] OECD, Public Spending on Health and Long-term Care: A New Set of Projections.
OECD, 2013.
[SKI 07] S KIADAS C., M ATALLIOTAKIS G., S KIADAS C.H., “An Extended Quadratic Health
State Function and the Related Density Function for Life Table Data”, in S KIADAS C.H.
(ed.), Recent Advances in Stochastic Modeling and Data Analysis, pp. 360–369, World
Scientific, 2007.
Methodology for an Optimum Health Expenditure Allocation 223
[SKI 18a] S KIADAS C.H., S KIADAS C., Exploring the Health State of a Population by
Dynamic Modeling Methods, Demographic Methods and Population Analysis, vol. 45,
Springer, Chum, 2018.
[SKI 18b] S KIADAS C.H., S KIADAS C., Demography and Health Issues: Population Aging,
Mortality and Data Analysis, Demographic Methods and Population Analysis, vol. 46.
Springer, Chum, 2018.
14
This chapter presents the ongoing research on the use of Markov models on electronic health
records. The research is aimed at the reliability of the data flow and process flow of the different
stages of the patient status following the health care processes in a health care setting.
Specifically, the use of mathematical modeling of the clinical pathways is examined through
literature research. Clinical pathways translate the best available evidence into practice,
indicating the most widely applicable order of treatment interventions for particular treatment
goals. A special focus will be given through a case study regarding the care of patients who are
under the age of 18 and have respiratory health problems. Applying models for the follow-up of
those patients and generally chronic patients is innovative and of great importance. Special
attention is given to chronic patients as the completion of the clinical pathway depends on
parameters that implicitly relate to the provided health services, such as the adulthood of
children and others. The results of the models could become valuable knowledge tools to help
the health care providers.
14.1. Introduction
According to the World Health Organization [WHO 18], the ultimate goal
of primary health care is better health for all. Primary health care does not
simply lead to an approach that places a greater emphasis on certain areas of
the health system in terms of prevention and out-of-hospital care in general.
Indeed, the strategy proposed by primary health care also influences the
Applications related to old ideas about the health care domain were
presented in a previous work [SPY 08]. The feasibility of an application of
such a model is examined in the field of clinical pathways and electronic
health records (EHRs). This model exploited stochastic modeling approaches
with the ultimate scope to calculate the reliability of clinical pathways.
The work for clinical data extraction and clinical pathways relies on the
area of reliability data analysis. The mathematical models of these techniques
mainly include the binary state and the multi-state system models. More
specifically, the mathematical techniques include (according to [SPY 08,
ROU 02]):
1) failure mode and effect analysis;
2) reliability block diagrams;
3) fault tree analysis;
4) Markov analysis;
5) hybrid techniques; etc.
(A = (I − Q)−1 × C)(2),
Markov models are used in the field of clinical effectiveness along with
cost-effectiveness [WIL 17].
The research part of this chapter is based on data obtained from the
outpatient clinic for respiratory problems in the Georgios Gennimatas
hospital of Thessaloniki, Greece, and it concerns the period 2008–2015. One
230 Data Analysis and Applications 4
main reason for this choice is the nature of the disease, which is a chronic
disease that often requires long-term follow-up with many fluctuations in its
development and hence the accumulation of large amounts of information.
We also chose it due to the patients’ ages and the fact they will probably
require follow-ups for many years. There is therefore a group of patients for
whom an organized and systemic medical history is a necessity, contributing
to a practical record for their care and therefore for their quality of life.
The data were extracted from the paper-based records and were uploaded
to the electronic medical record database. The record included 157 patients at
the time of treatment, and 840 visits by patients aged 2–18 were recorded.
Figure 14.1. The process of applying the DTMC in chronic patients. For
a color version of this figure, see www.iste.co.uk/makrides/data4
means that the patient’s status, as recorded by the treating physician, was
recorded in an encoded form. Therefore, the doctor immediately has a quick
overview of the course of the patient’s illness. This could be used, for
example, in association with symptoms which, in some cases, such as
outbreaks of bronchial asthma with viruses during the winter months, may be
severe. It is of course intended to make possible any necessary interventions
at both the individual and social levels after the necessary epidemiological
investigations.
As described in Figure 14.1, after completing the first step of the “EHR
data extraction”, a state transition diagram was created in order to apply the
DTMC model and calculate the overall reliability of the pathway and transition
probabilities among states.
14.3.2. Results
The results of the study included 471 patient visits in state 1, of which 328
reverted to status 1. Also, 264 visits resulted in non-attendance. Among the
limitations is that it is accepted that there is an initial situation for our scenario
as well as a terminal, which is considered to be state 1 (monitoring), which
symbolizes the successful termination (T) of the script. There are also other
termination states that are unsuccessful termination situations, which in our
case involved non-turnout (No Visit, NV).
Among other results, the probability that a patient will switch from state
1 (steady-state) to NV (non-attendance) is 12%. This is particularly important
for the patient’s progress and shows the success of the quality of the medical
services of the clinic. This is because, in this particular disease, parents often
find their children stable in the course of the disease, and either slow down or
terminate their child’s follow-up. This often results in some initial symptoms
either to divert the attention of the environment or to underestimate it, so as
not to get a timely and best-performing therapeutic intervention.
However, the probability of transition from the initial state (“1”) to the state
where the patients do not visit the clinic again (NV) does not seem to have
any significant effect on the probability for successful termination (“T”) of
the whole clinical pathway. The probability for successful termination is about
0.38 and can reach 0.347 even when the probability from state “1” to state
“NV” reaches 0.1.
232 Data Analysis and Applications 4
Several other results have been examined that lead to the fact that a detailed
investigation among the states and transition probabilities should be notified in
order to have valuable results from our model.
14.4. Conclusion
More generally, managing the health services for chronic diseases has come
to a need to develop models to guide public health care in general.
The health policy area needs analytic models with the scope of making
decisions with valuable outcomes for the stakeholders (doctors, patients,
citizens, etc.).
Probabilistic Models for Clinical Pathways: The Case of Chronic Patients 233
14.5. References
[SOB 11] S OBOLEV B.G., S ANCHEZ V., VASILAKIS C., “Systematic review of the use of
computer simulation modeling of patient flow in surgical care”, J. Med. Syst., vol. 35,
no. 1, pp. 1–16, 2011.
[SOT 02] S OTO C.M., K LEINMAN K.P., S IMON S.R., “Quality and correlates of medical
record documentation in the ambulatory care setting”, BMC Health Services Research,
vol. 2, no. 1, 2002.
[SPY 08] S PYROU S. et al., “A methodology for reliability analysis in health networks”, IEEE
Trans. Inf. Technol. Biomed., vol. 12, no. 3, pp. 377–386, 2008.
[WAZ 01] WAZEKA A. et al., “Impact of a pediatric asthma clinical pathway on hospital cost
and length of stay”, Pediatr. Pulmonol., vol. 32, no. 3, pp. 211–216, 2001.
[WHO 18] WHO, “Primary health care”, [cited 1-2-2018]. Available at:
http://www.who.int/topics/primary_ health_care/en/, 2018.
[WIL 17] W ILLIAMS C. et al., “Estimation of survival probabilities for use in
cost-effectiveness analyses: A comparison of a multi-state modeling survival analysis
approach with partitioned survival and Markov decision-analytic modeling”, Med. Decis.
Making, vol. 37, no. 4, pp. 427–439, 2017.
[ZAC 15] Z ACHARIAH J.P., S AMNALIEV M., “Echo-based screening of rheumatic heart
disease in children: A cost-effectiveness Markov model”, J. Med. Econ., vol. 18, no. 6,
pp. 410–419, 2015.
[ZAI 12] Z AITSEVA E., RUSIN M., “Healthcare system representation and estimation based
on viewpoint of reliability analysis”, Journal of Medical Imaging and Health Informatics,
vol. 2, no. 1, pp. 80–86, 2012.
[ZHA 15] Z HANG Y., PADMAN R., PATEL N., “Paving the COWpath: Learning and
visualizing clinical pathways from electronic health record data”, J. Biomed. Inform.,
vol. 58, pp. 186–197, 2015.
[ZHA 16] Z HANG Y., PADMAN R., “Data-driven clinical and cost pathways for chronic care
delivery”, Am. J. Manag. Care, vol. 22, no. 12, pp. 816–820, 2016.
15
Demographic factors and the global economic crisis play a significant role and greatly affect the
performance of national health care systems. The expenditure on health is an important indicator
for understanding and measuring health care performance. For this purpose, in this chapter, we
address and discuss issues related to health care systems and focus on the performance of such
systems using multivariate statistical techniques, and classify OECD countries in clusters based
on the similarities on the expenditure on health.
15.1. Introduction
Health spending typically includes costs for the prevention and treatment
of diseases as well as the rehabilitation of the patient after treatment for the
disease.
Demographic factors and the global economic crisis forced many countries
to look into these costs and especially to seek ways to “force” these resources
to produce better results. This led to a growing interest in the comparisons of
performance in national health care systems. The expenditure on health is an
important indicator for understanding and measuring health care performance.
For this purpose, in this chapter we will classify OECD countries in clusters
based on the similarities on the expenditures on health.
within each group are homogeneous and the observations of different groups
are as different as possible.
Typical steps in cluster analysis include the selection of the variables, the
classification method, the metric distance and the linkage algorithm, as well
as, after the classification, the selection of an ideal number of clusters and
eventually the evaluation and interpretation of the results.
The questions that arise for the implementation of the above approach are
the way of measuring similarity and distance between groups/clusters, the
way of defining the minimum distance of separation of two groups and the
clustering criterion, since different criteria may give rise to different
classification groups with the same (or very similar) characteristics.
The case of continuous data is perhaps the simplest as well as the one
that offers numerous possibilities. Indeed, there are many measures that have
been used to quantify the distance between continuous data. Some of the most
popular are presented below. It should be pointed out that these measures do
not necessarily satisfy all the properties mentioned above. In fact, it should be
emphasized that for our purposes, only the first of the above three properties
is mandatory, as long as the distance between two objects or functions is non-
negative.
– Euclidean distance is defined by
p
d(X, Z) = |X − Z| = (xi − zi )2 = (X − Z)′ (X − Z).
i=1
and is used in the presence of outliers since the absolute value downweights the
effect of extreme observations, as compared to the square used in the Euclidean
distance.
On Clustering Techniques for Multivariate Demographic Health Data 239
dT = max|xi − zi | , i = 1, ..., p
Another important point for the algorithm is how we measure the distance
between groups. The most common methods of calculating this distance are
nearest neighbor, furthest neighbor, average linkage, weighted average,
centroid and the Ward method.
Note that the Ward method greatly differs from all others, as it is an
ANOVA-based approach, viz. uses variance analysis to calculate the distances
between clusters.
For more details on the classification and clustering techniques, the reader
may refer among others to [AND 73, EVE 80, BER 03].
15.4.1. Data
The original data comes from the World Health Organization (WHO) and
the Global Health Observatory (GHO)1. The full dataset consists of time series
data from 2000 to 2015 and covers most of the countries (98 in total) monitored
by the World Health Organization.
The variables of the dataset can be divided into two distinct groups: the
first group concerns variables that measure the resources, financial, material
and human, flowing into the health system (input dataset) and affect the
1 http://www.who.int/gho/en/
On Clustering Techniques for Multivariate Demographic Health Data 241
health of the population, and the second group concerns outcome variables
(output dataset) of the input variables. More specifically, there are 28 input
variables which can be further divided into three indicator subsets, namely
economic inputs, human resources and infrastructures. A brief description is
presented in Table 15.1. Furthermore, there are 20 output variables which can
be mainly characterized as demographic indices. A brief description is
provided in Table 15.2.
Name Variable
1 Current health expenditure as a percentage of gross domestic product
2 Domestic general government health expenditure as a percentage of current health
expenditure
3 Domestic private health expenditure as a percentage of current health expenditure
4 External health expenditure as a percentage of current health expenditure
5 Out-of-pocket expenditure as a percentage of current health expenditure
6 Pharmaceutical personnel density per 1,000 population
7 Physician density per 1,000 population
8 Laboratory health worker density per 1,000 population
9 Dentistry personnel density per 1,000 population
10 Community and traditional health worker density per 1,000 population
11 Other health worker density per 1,000 population
12 Environmental and public health worker density per 1,000 population
13 Health management & support worker density per 1,000 population
14 Nursing and midwifery personnel density per 1,000 population
15 Hospital beds per 1,000 population
16 Legislation
17 Cooperation
18 Surveillance
19 Response
20 Preparedness
21 Risk communication
22 Human resources
23 Laboratory
24 Points of entry
25 Zoonosis
26 Food safety
27 Chemical risk
28 Radionuclear risk
This section presents the key results of cluster analysis. The results are
presented in two stages: initially, the results of each method are presented
separately and then an analysis is presented for comparative purposes.
Table 15.3 gives the breakdown of countries into two clusters. It is shown
that the largest concentration of countries in a cluster is observed using the
hierarchical clustering with the complete linkage (HCA-COM) method. The
highest separation almost 50–50 is observed via the K-means (KM) algorithm.
On Clustering Techniques for Multivariate Demographic Health Data 243
By comparing the classification of the countries with the HCA COM and
the Ward methods, we observe that 44.9% of the countries have the same
classification in the 2 hierarchical methods; while comparing their
classification with K-means, we observe that the highest classification rate is
54% using the complete linkage. The Ward method (WARD) comes second
with 44%.
Methods Similar classification %
HCA COM vs. Ward 44 44.9%
HCA COM vs. KM 54 55.1%
HCA WARD vs. KM 44 44.9%
Figures 15.5–15.7 present the classification results for the output dataset.
For the hierarchical clustering, the complete and Ward linkages have been
used.
Table 15.5 gives the breakdown of countries into the two clusters. We
observe that the largest concentration of countries in a cluster is done using
the HCA COM method (79%). We observe that all three methods separate the
countries in a similar way, with each method selecting approximately 70%
(from 65 to 82%) of the countries to be classified in the first cluster.
246 Data Analysis and Applications 4
While comparing the classification of the countries with the HCA COM
and Ward methods, we observe that 86.6% of the countries have the same
classification in the two hierarchical methods; while comparing their
classification and the method K-means, we observe that 89.07% of the
countries have the common classification in all methods.
Methods Similar classification %
HCA COM vs. Ward 103 86.6%
KM vs. HCA.All 106 89.07%
HCA COM vs. KM 107 89.9%
HCA Ward vs. KM 115 96.6%
15.5. Conclusion
15.6. References
[ALD 84] A LDENDERFER M.S., B LASHFIELD R.K., Cluster Analysis, Sage Publications,
Newbury Park, 1984.
[AND 73] A NDERBERG M.R., Cluster Analysis for Applications, Academic Press, New York,
1973.
[BER 03] B ERKES I., H ORVÁTH L., KOKOSZKA P., “GARCH processes: Structure and
estimation”, Bernoulli, vol. 9, no. 2, pp. 201–227, 2003.
[CLE 54] C LEMENTS F.E., Use of Cluster Analysis with Anthropological Data, 1954.
Available at: https://anthrosource.onlinelibrary.wiley.com/doi/pdf/10.1525/aa.1954.56.2.02
a00040.
[DRI 32] D RIVER H.E., K ROEBER A.L., Quantitative Expression of Cultural Relationships,
University of California Publications in American Archaeology and Ethnology, vol. 31,
pp. 211–56, 1932.
[EVE 80] E VERITT B., Cluster Analysis, Wiley, Chichester, 1980.
250 Data Analysis and Applications 4
[WAR 63] WARD J.H., “Hierarchical groupings to optimize an objective function”, Journal of
the American Statistical Association, vol. 58, pp. 236–244, 1963.
16
Smoking is a popular habit in the Greek population. Because of that – not forgetting the problem
of second-hand smoking – there is an elevated risk for the development of tobacco-related
diseases. Thus, the mortality caused by diseases of the circulatory system, such as neoplasms
was studied. A combined procedure was applied for the estimation of the years lost regarding
longevity depending on age and gender under the action of each tobacco-related cause of death.
According to this procedure, the typical method of elimination of related causes of death is
applied along with the method developed by Arriaga [ARR 84] for the examination of the
differences of life expectancy between two life tables. The results of the analysis indicate that
there is a decreasing effect of the tobacco-related diseases of the circulatory system on
longevity. A more or less stagnated picture emerges for neoplasms. Contrarily, the effects of the
diseases of the respiratory system become more significant over time. Overall, the problem
remains significant for the Greek population, a fact that outlines the need for the intensification of
the existing policies or the application of new ones concerning smoking and second-hand
smoking.
16.1. Introduction
2007; Kyrlesi et al. [KYR 07]). The prevalence of smoking and second-hand
smoking is also high in the newly recruited soldiers (years 2013 and 2014;
Grapatsas et al. [GRA 17]) as in rural Greece and elsewhere (Stafylis et al.
[STA 18]). According to Eurostat [EUR 16], about one in three individuals
aged 15 and over was a smoker and a vast majority of the population (64.2%)
was exposed daily to tobacco smoke indoors in 2014, despite the relevant anti-
smoking legislation (see Telionatis et al. [TEL 17]), which is seldom enforced
effectively (Filippidis and Tzoulaki [FIL 16]).
Consequently, as Stafylis et al. [STA 18] note, the risk for tobacco-related
diseases must be considered as more than significant. According to the
Harvard School of Public Health [HAR 11], we can speak about the “Greek
tobacco epidemic”, as in 2008, 46.4% of deaths were due to cardiovascular
diseases, 25.7% due to cancer and 9.6% due to diseases of the respiratory
system. Obviously, the cost of the effects of tobacco consumption on the
population’s health and longevity is huge, not forgetting at the same time the
money required by individuals to satisfy this habit as well as the expenditure
on social and medical care as well as the hospitalization of patients.
subsequently compared by applying the Arriaga method [ARR 84, ARR 89]
in order to evaluate the effects of smoking-related deaths on life expectancies
at birth and examine their impact on each age group of the human life span.
The details of this procedure will be discussed in section 16.2.
According to the U.S. Department of Health and Human Services [OFF 04],
smoking can lead to a variety of diseases, i.e. cancer, cardiovascular and
respiratory diseases, and has several other effects on health like, for example,
reproductive ones. Auger et al. [AUG 14] (see also Baliunas et al. [BAL 07])
divides the main causes of death related to some extent with smoking into
seven categories:
1) malignant neoplasms: (a) lip, oral cavity, pharynx (b), esophagus, (c)
larynx, (d) trachea, bronchus, lung, (e) stomach, (f) pancreas, (g) cervix uteri,
(h) kidney, (i) bladder and (j) myeloid leukemia;
2) circulatory system: (a) ischemic heart diseases, (b) other heart diseases
(c) cerebrovascular disease, (d) atherosclerosis (e) aortic aneurism, dissection
(f) other arterial diseases;
3) respiratory system: (a) influenza, pneumonia, (b) bronchitis,
emphysema, (c) other chronic obstructive;
4) digestive: ulcer;
5) mental-behavioral disorders;
6) exposure to smoke, fire, flames;
7) perinatal: (a) length of gestation, fetal growth, (b) sudden infant death
syndrome.
However, besides the fact that this database contains detailed information
about the malignant neoplasms described in the previous section, no data are
available for myeloid leukemia. Thus, this disease was omitted from the
analysis, noting at the same time its very low impact on mortality (only 1.8%
of the reported deaths from smoking-related malignant neoplasms in 2008;
see Harvard School of Public Health [GRA 17, p. 26]).
For the diseases of the respiratory system, the analysis was done in the
following categories: influenza (including swine flu), pneumonia, diseases of
the lower respiratory system (asthma and status asthmaticus and other lower
respiratory diseases) as well as other diseases of the respiratory system
(remainder of J00-J99), as smoking can cause a great variety of pulmonary
diseases (see Centers for Disease Control and Prevention (US) [CEN 10]),
which are not cited separately in the database. Thus, the analysis was focused
exclusively on the diseases that constitute the major causes of tobacco-related
deaths in the Greek population.
The method used for the analysis was briefly discussed in section 16.1. It
must be noted that many others besides the Arriaga method could have been
used for this purpose, like those of Andreev [AND 82], Pollard [POL 82,
POL 88] and Andreev and Shkolnikov [AND 12] (see also Andreev et al.
[AND 02]). The Arriaga method was chosen not only for its reliability as seen
in the published literature (see, for example, Yang et al. [YAN 12];
Auger et al. [AUG 14]; Le et al. [LE 15]; Sunberg et al. [SUN 18] and
Tobacco-related Mortality in Greece 255
Chisumpa et al. [CHI 18]) but also because of its simplicity, as will be proven
later.
We must also emphasize that Arriaga [ARR 84] compared the differences
of life expectancy between two time points; however, this procedure can be
easily applied in order to study longevity differences because of the effect of
the smoking-related causes of death. According to this procedure, the effects
of mortality change on life expectancies can be classified as “direct” and
“indirect”: “the direct effect on life expectancy is due to the change in life
years within a particular age group as a consequence of the mortality change
in that age group.” The indirect effect “consists of the number of life years
added to a given life expectancy because the mortality change within (and
only within) a specific age group will produce a change in the number of
survivors at the end of the age interval.” Another effect springs from the
interaction between the exclusive effect of each age group and the overall
effect.
and
t+n
Tx+i lxt lt
i OE x = − tx+i
lat lxt+n lx+1
In these equations, the terms x and x+i refer to age groups, t and t+n
correspond to time points (years), l to the number of survivors at an exact age
256 Data Analysis and Applications 4
where 2 represents the relevant values of the second life table, in which the
tobacco-related causes of death have been eliminated, and 1 refers to the
observed data (i.e. it includes tobacco-related diseases). n Rxi is the proportion
of deaths from cause i, n ∆x is the contribution of all-cause mortality
differences to age group x to x+n differences in life expectancies and n m2x
corresponds to the age specific mortality rates. Obviously, the term n Rxi,2 of
the above equation related to non-tobacco-related mortality, is 0.
16.3. Results
in the longevity increase is observed in the male and female population of the
country.
In the literature, it has been seen that either a pro-cyclical pattern or an anti-
cyclical pattern of mortality in the era of economic recession may prevail. In
the first case, mortality reduces in times of recession and increases in boom
periods (see, for example, Ruhm [RUH 07]). In the second case, the opposite
happens: mortality reduction in conditions of economic growth and mortality
increases in times of recessions (see Catalano [CAT 97]). However, it must be
taken into account that a “time lag” is needed for the effects of economic crisis
258 Data Analysis and Applications 4
to be seen on mortality (for more on this concept, see Laporte [LAP 04]).
Also, the certain cause- and age-specific effects could be more complicated.
For example, reduced economic activity may lead to a reduction in deaths from
road accidents, while an increase in death rates among middle-aged persons,
especially males, from lifestyle-related causes might be observed. In any case,
the detailed discussion of the effects of economic crises on longevity in Greece
is beyond the scope of this chapter. Instead, the effects of smoking-related
mortality will be presented in the following sections.
It is not surprising then that the overall effect of the diseases of the
circulatory system is huge, being the most important cause of death in Greece
at least during the first years of the study (Figure 16.2). Afterwards, the effect
on longevity decreases in both genders, as their proportion in the observed
life expectancy at birth and the two genders tend to converge.
For the sum of the diseases of the circulatory system studied, a significant
female excess mortality is observed; a well-known phenomenon in the
literature. Garcia et al. [GAR 16] in their review about cardiovascular disease
in women discuss several clinical perspectives which are responsible for this
phenomenon, including hypertension, dyslipidemia, diabetes, smoking,
obesity, physical inactivity, preterm delivery, hypertensive pregnancy
disorders, gestational diabetes and menopausal transition. Female excess
mortality is confirmed in both cerebrovascular diseases and in the “other heart
diseases”.
Tobacco-related Mortality in Greece 259
Figure 16.3. Effects of the diseases of the circulatory system on longevity by broad
age group. Males: white markers and continued lines; females: black markers and
dashed lines. Percentages represent the analogy of the effect of each cause of death
on the observed life expectancy at birth
these diseases to 0.12 years in females and 1.7 years in males. However, the
female excess described in the previous paragraphs is not confirmed for that
age group, as mortality is always higher in males.
A similar situation prevails for the group named “other heart diseases”. In
females, the effect of these diseases tends to increase until 1998 (the effect on
life expectancy at birth is 5.5 years), it fluctuates afterwards and from
5.2 years in 2006 decreases to 2.3 years in 2016. In males, from 3.4 years in
1998, it gradually decreases to 1.5 in 2016. In the age group of 45–64 years, a
similar picture as the one described in the previous paragraph is seen: the
female excess is not confirmed, and mortality is higher in males. However, a
significant difference exists. In females, the effect on longevity of this group
of diseases from 0.2 years in 1994 decreases to 0.1 years in 2002, and
afterwards, it remains practically unchangeable. In males, it decreases from
0.3 years in 1994 to 0.2 years in 2003, and afterwards, it remains almost
stable.
According to Zeegers et al. [ZEE 00], the risk of developing bladder cancer
in smokers is two to four times more than non-smokers, while smoking is
responsible for 23% of cases in females and 50% in males (Zeegers et al.
[ZEE 04]). This picture is confirmed concerning the bladder cancer mortality
rate in Greece. The years lost fluctuates between 0.3 and 0.4 years in males,
while in females, the effect is always lower than 0.1 years.
1 https://www.lung.org/lung-health-and-diseases/lung-disease-lookup/lung-cancer/learn-
about-lung-cancer/what-is-lung-cancer/what-causes-lung-cancer.html.
264 Data Analysis and Applications 4
from 0.26 in 1994 to 0.36 in 2016. In females, they increase from 0.2 to 0.3 in
the same years.
As for laryngeal cancers, according to Jones et al. [JON 16], alcohol and
tobacco act in their development. The years lost because of larynx cancers is
limited: below 0.16 in males and 0.02 in females. The same gender
diversification is observed in kidney cancers. While the risk of developing
such cancer is related to tobacco smoking (Zeegers et al. [ZEE 00]), the
mortality caused by it, in males besides its general elevating temporal trend,
remains at 0.15–0.17 years in 2014–2016. In females, the effect is minimal,
less than 0.1 years lost for the period studied.
The other three groups of cancers, those of lip, oral cavity and pharynx,
of esophagus and cervix uteri (for females) always have minimal effects on
longevity. A small increasing trend of mortality is observed under the action
of lip, oral cavity and throat cancers.
According to Behr and Nowak [BEH 02], tobacco smoking affects the
respiratory tract in two ways (mechanisms): 1) by the induction of
inflammation and 2) through its mutagenic/carcinogenic action. The second
mechanism was discussed in the previous section. As for the first mechanism,
smoking is the most important factor for the development of chronic
obstructive pulmonary disease and interstitial lung diseases (idiopathic
pulmonary fibrosis/usual interstitial pneumonia, desquamative interstitial
pneumonia, respiratory bronchiolitis-associated interstitial lung disease, etc.).
It must be stressed that in some diseases, a decreased incidence or severity is
observed in smokers (e.g. hypersensitivity pneumonitis). In any case, several
details of the smoking-related diseases can be found in Centers for Disease
Control and Prevention [CEN 08]. Also, it must not be forgotten that other
agents may have played an important role in the development of the
respiratory diseases, including environmental pollution, socioeconomic status,
the number of family members in relation to residential space, the existence
and nature of respiratory disease of co-habitants and their smoking habits, and
the kind of heating and cooking sources (Sichletidis et al. [SIC 05]).
In that way, the loss of longevity because of the respiratory diseases tends
to increase with fluctuations in males from 1.4 in 1994 to 2.4 years in 2009,
and after a small retreat afterwards, it remains 2.3–2.4 years in the period
2014–2016 (Figure 16.6). In females, the same general trend is followed:
while 1.2 years are lost because of the diseases of the respiratory system in
1994, the relevant value is 2.6 years in 2009 and 2.3–2.4 years in 2014–2016.
Besides the fact that a male excess mortality is evident for the first years
of the study, females quickly converge with males, and in the period 2002–
2012, the effects of respiratory diseases on their mortality are significantly
higher than males in absolute numbers. However, the relative effect of these
diseases on longevity remains lower most of the time, and only in the period
2002–2004 is it marginally higher. Overall, these developments can reasonably
be attributed to the changing habits of women in Greece and the increase of
smoking among them.
266 Data Analysis and Applications 4
The general pattern observed so far for the total of the respiratory diseases
is largely followed in the “other diseases of the respiratory system”. In males,
the effect on longevity from 1 year in 1994 reaches 1.6 years in 2009, retreats
slightly afterwards, and after an increasing course, it reaches 1.6 years in 2015
2 See https://ec.europa.eu/eurostat/ramon/nomenclatures/index.cfm?TargetUrl=LST_NOM_DTL&
StrNom=COD_2012&StrLanguageCode=EN&IntPcKey=31100702&StrLayoutCode=
HIERARCHIC&IntCurrentPage=1.
Tobacco-related Mortality in Greece 267
and 1.5 years in 2016. In females, from 0.9 years in 1994 it reaches 1.8 in
2008, and following a parallel course with males, it will be 1.7 years or more
in the period 2014–2016. An analogous course is described by the relative
frequencies of Figure 16.6. However, even if the differences between the two
genders were very small in the period 1994–2000, afterwards, the effects both
in relative and in absolute numbers tend to be more important in the female
population of the country. As pointed out before, this might be attributed to a
progressively increased use of tobacco products by women.
Heart diseases are the leading cause of amenable and preventable deaths.
In total, over 77% of all amenable deaths in the EU were caused by ischemic
heart diseases, cerebrovascular diseases, colorectal cancer, breast cancer,
hypertensive diseases and pneumonia. Preventable deaths are due to ischemic
heart diseases, lung cancer, accidental injuries, alcohol-related diseases,
colorectal cancer, suicides and self-inflicted injuries and chronic obstructive
Tobacco-related Mortality in Greece 269
mainly by the effects on older people (65+ years) and, to a lesser degree, by the
relevant effects on people aged 45–64 years, while the male excess mortality
is confirmed in all of them.
– Diseases of the respiratory system: following some significant
fluctuations, the effect of these diseases tends to increase over time. This is the
case for the diseases of the lower respiratory system; however, among them,
the deaths due to “asthma and status asthmaticus” are in decreasing order,
and only a marginal elevation of their effects on longevity is observed in the
most recent years. Several fluctuations and a general increasing trend are also
observed for “the other diseases of the respiratory system” in which, in contrast
to the diseases of the lower respiratory system, a female excess of mortality is
observed. As for pneumonia and influenza (including swine flu), an irregular
pattern is followed, with outbreaks and recessions of the effects of these
diseases on the longevity of population. Also, as seen in the previous diseases
studied, all the developments are governed primarily by the developments in
the older age group (65+ years) and secondarily from those of the people aged
45–64 years.
16.5. References
[COO 00] C OOPER G.M., “The development and causes of cancer”, in C OOPER G.M. (ed.),
The Cell: A Molecular Approach, 2nd edition, Sinauer Associates, Sunderland, MA, USA.
Available at: https://www.ncbi.nlm.nih.gov/books/NBK9963/, 2000.
[EUR 16] E UROSTAT, “Tobacco consumption 1 in every 4 persons aged 15 or over in the
European Union is a smoker. . . and 1 in 5 confronted by passive smoking”. Available at:
https://ec.europa.eu/eurostat/documents/2995521/7762296/3-07122016-AP-EN.pdf, 2016.
[EUR 17] E UROPEAN H EART N ETWORK, European Cardiovascular Disease Statistics.
Available at: http://www.ehnheart.org/cvd-statistics/cvd-statistics-2017.html, 2017.
[EUR 18] E UROSTAT, “Amenable and preventable deaths statistics: Statistics Explained”.
Available at: https://ec.europa.eu/eurostat/statistics-explained/ pdfscache/41683.pdf, 2018.
[EZZ 05] E ZZATI M., H ENLEY S.J., T HUN M.J. et al., “Role of smoking in global and
regional cardiovascular mortality”, Circulation, vol. 112, no. 4, pp. 489–97, 2005.
[FIL 16] F ILIPPIDIS F.T., T ZOULAKI I., “Greece giving up on tobacco control”, Addiction,
vol. 111, pp. 1306–1307, 2016.
[FIL 17] F ILIPPIDIS F.T., G EROVASILI V., M AN W.D.-C. et al., “Trends in mortality from
respiratory system diseases in Greece during the financial crisis”, European Respiratory
Journal, vol. 48, pp. 1487–1489, 2017.
[GAR 16] G ARCIA M., M ULVAGH S.L., M ERZ C.N. et al., “Cardiovascular disease in
women: Clinical perspectives”, Circulation Research, vol. 118, no. 8, pp. 1273–1293, 2016.
[GRA 17] G RAPATSAS K., L EIVADITIS V., D IMOPOULOS E. et al., “Characteristics of
smoking among young Greek males at the time of their recruitment in the army”, Annals of
Research Hospitals, vol. 1, no. 10, 2017.
[HAR 11] H ARVARD S CHOOL OF P UBLIC H EALTH, The Greek Tobacco Epidemic: Prepared
by the Faculty of the Harvard School of Public Health, Harvard University Press,
Cambridge, MA, USA. Available at: https://www.who.int/fctc/reporting/party_reports/
greece_annex1_the_greek_tobacco_epidemic_2011.pdf, 2011.
[HOE 13] H OEK G., K RISHNAN R.M., B EELEN R. et al., “Long-term air pollution exposure
and cardio-respiratory mortality: A review”, Environmental Health, vol. 12, no. 1, p. 43,
2013.
[IOD 08] I ODICE S., G ANDINI S., M AISONNEUVE P. et al., “Tobacco and the risk of
pancreatic cancer: A review and meta-analysis”, Langenbeck’s Archives of Surgery,
vol. 393, pp. 535–45, 2008.
[JAC 99] JACOBS J R . D.R., A DACHI H., M ULDER I. et al., “Cigarette smoking and mortality
risk: Twenty-five-year follow-up of the Seven Countries Study”, Archives of Internal
Medicine, vol. 159, no. 7, pp. 733–40, 1999.
[JON 16] J ONES T.M., D E M., F ORAN B. et al., “Laryngeal cancer: United Kingdom
National Multidisciplinary guidelines”, The Journal of Laryngology and Otology, vol. 130,
no. S2, pp. S75–S82, 2016.
[KEN 14] K ENTIKELENIS A., K ARANIKOLOS M., R EEVES A. et al., “Greece’s health crisis:
From austerity to denialism”, Lancet, vol. 383, no. 9918, pp. 748–753, 2014.
274 Data Analysis and Applications 4
[KHA 18] K HANI Y., P OURGHOLAM -A MIJI N., A FSHAR M. et al., “Tobacco smoking and
cancer types: A review”, Biomedical Research and Therapy, vol. 5, no. 4, pp. 2142–2159,
2018.
[KYR 07] K YRLESI A., S OTERIADES E.S., WARREN C.W. et al., “Tobacco use among
students aged 13–15 years in Greece: The GYTS project”, BMC Public Health, vol. 7,
no. 3, 2007.
[LAL 16] L ALIOTIS I., I OANNIDIS J.P.A., S TAVROPOULOU K.C., “Total and cause-specific
mortality before and after the onset of the Greek economic crisis: An interrupted time-
series analysis”, Lancet Public Health. Available at: http://dx.doi.org/10.1016/S2468-
2667(16)30018-4, 2016.
[LAP 04] L APORTE A., “Do economic cycles have a permanent effect on population health?
Revisiting the Brenner hypothesis”, Health Economics, vol. 13, no. 8, pp. 767–779, 2004.
[LE 15] L E Y., R EN J., S HEN J. et al., “The changing gender differences in life expectancy in
chinese cities 2005–2010”, PLoS ONE, vol. 10, no. 4, p. e0123320, 2015.
[MAA 10] M AAS A.H., A PPELMAN Y.E., “Gender differences in coronary heart disease”,
Netherlands Heart Journal: Monthly Journal of the Netherlands Society of Cardiology and
the Netherlands Heart Foundation, vol. 18, no. 12, pp. 598–602, 2010.
[MIC 18] M ICHAS G., K ARVELAS G., T RIKAS A., “Cardiovascular disease in Greece;
the latest evidence on risk factors”, Hellenic Journal of Cardiology. Available at:
https://doi.org/10.1016/j.hjc.2018.09.006, 2018.
[NAM 87] NAMBOODIRI N.K., S UCHINDRAN C.M., Life Table Techniques and Their
Applications (chapters 6 and 9), Academic Press Orlando, Florida, 1987.
[NIK 04] N IKOLAIDIS G., Z AVRAS D., B ONIKOS D. et al., “Trends of mortality rates during
the last thirty years in Greece”, Journal of Medical Systems, vol. 28, no. 6, pp. 607–616,
2004.
[OFF 04] O FFICE OF THE S URGEON G ENERAL , O FFICE ON S MOKING AND H EALTH, The
Health Consequences of Smoking: A Report of the Surgeon General, Centers for Disease
Control and Prevention, Atlanta, GA, USA, 2004.
[POL 82] P OLLARD J.H., “The expectation of life and its relationship to mortality”, Journal
of the Institute of Actuaries, vol. 109, pp. 225–240, 1982.
[POL 88] P OLLARD J.H., “On the decomposition of changes in expectation of life and
differentials in life expectancy”, Demography, vol. 25, pp. 265–276, 1988.
[PRE 01] P RESTON S.H., H EUVELINE P., G UILLOT M., Demography: Measuring and
Modeling Population Processes, Blackwell Publishers, Malden, MA, USA, 2001.
[RON 10] RONTOS K., G IAVRIMIS P., ROUMELIOTOU M. et al., “Spatial and cause-specific
analysis of age at death in Greece”, Portuguese Journal of Quantitative Methods, vol. 1,
no. 1, pp. 73–90, 2010.
[ROS 13] ROSTRON B., “Smoking-attributable mortality by cause in the United States:
Revising the CDC’s data and estimates”, Nicotine & Tobacco Research, vol. 15, no. 1,
pp. 238–246, 2013.
Tobacco-related Mortality in Greece 275
[RUH 07] RUHM C.J., “A healthy economy can break your heart”, Demography, vol. 44, no. 4,
pp. 829–848, 2007.
[SIC 05] S ICHLETIDIS L., T SIOTSIOS I., G AVRIILIDIS A. et al., “The effects of
environmental pollution on the respiratory system of children in Western Macedonia,
Greece”, Journal of Investigational Allergology and Clinical Immunology, vol. 15, no. 2,
pp. 117–123, 2005.
[SIF 17] S IFAKI -P ISTOLLA D., L IONIS C., G EORGOULIAS V. et al., “Lung cancer and
tobacco smoking in Crete, Greece: Reflections from a population-based cancer registry
from 1992 to 2013”, Tobacco Induced Diseases, vol. 15, no. 6, 2017.
[SIM 14] S IMOU E., KOUTSOGEORGOU E., “Effects of the economic crisis on health and
healthcare in Greece in the literature from 2009 to 2013: A systematic review”, Health
Policy, vol. 115, nos 2–3, pp. 111–119, 2014.
[STA 18] S TAFYLIS C., R ACHIOTIS G., K ATSIOULIS A. et al., “Prevalence and determinants
of smoking and secondhand smoke exposure in a rural population of central Greece:
A cross-sectional study”, Rural and Remote Health, vol. 18, no. 4218. Available at:
https://doi.org/10.22605/RRH4218, 2018.
[SUN 18] S UNDBERG L., AGAHI N., F RITZELL J. et al., “Why is the gender gap in life
expectancy decreasing? The impact of age- and cause-specific mortality in Sweden 1997–
2014”, International Journal of Public Health, vol. 63, pp. 673–681, 2018.
[TEL 17] T ELONIATIS S., T ZORTZI A., L IOZIDOU A. et al., “Smoking prevalence,
compliance and attitudes towards smoking bans among School Teachers in Attica, Greece
2014”, Pneumon, vol. 30, no. 4, pp. 227–235, 2017.
[YAN 12] YANG S., K HANG Y.-H., C HUN H. et al., “The changing gender differences in
life expectancy in Korea 1970–2005”, Social Science & Medicine, vol. 75, pp. 1280–1287,
2012.
[ZAF 19] Z AFEIRIS K.N., KOSTAKI A., “Recent mortality trends in Greece”,
Communications in Statistics – Theory and Methods, vol. 48, no. 1, pp. 112–126,
2019.
[ZEE 00] Z EEGERS M.P., TAN F.E., D URANT E. et al., “The impact of characteristics of
cigarette smoking on urinary tract cancer risk: A meta-analysis of epidemiologic studies”,
Cancer, vol. 89, pp. 630–639, 2000.
[ZEE 04] Z EEGERS M.P., K ELLEN E., B UNTINX F. et al., “The association between
smoking, beverage consumption, diet and bladder cancer: A systematic literature review”,
World Journal of Urology, vol. 21, pp. 392–401, 2004.
List of Authors
A, C D, E, G
adjusted evaluation measures, 3, 5, 6, default rate distribution, 33–35, 42,
14, 29, 32 43, 45, 47
alternative data, 51, 52, 54, 57, 59 demographic health data, 235
annual salary enhancements, 199 distance, 185–187, 189, 195
asymptotic availability, 119, 129, measures, 238
130, 132, 143, 144 expectation-maximization
chi-square distribution, 185, 186 algorithm, 61
chronic patients, 225, 229, 230, 232 Expected Percentage Shortfall (EPS),
clinical pathways, 225–228, 232–234 3, 9, 12, 13, 15, 16, 23, 24, 28, 29
coefficient of economic extremal index, 89, 90, 92, 93, 95,
revaluation, 199 99–101
complex system, 77, 87 extreme
computational linguistics, 191 events, 89, 100
Conjoint Analysis, 199, 201, 204, value theory, 89–91, 100, 101
211, 212 Gauss-Laguerre quadrature formula,
correlation matrix, 167–169, 171, 61, 67, 68
174–176, 182 Gaussian distribution, 167
credit Generalized Jackknife methodology,
behavior data, 51, 52 89, 93, 94
risk assessment model, 51 Greek
risk model, 51 health expenditure, 215–222
scoring, 51, 52, 54, 59, 60 population, 251, 254, 261, 263, 271
critical state, 77, 79
in
Innovation, Entrepreneurship and Management
2020
ANDREOSSO-O’CALLAGHAN Bernadette, DZEVER Sam, JAUSSAUD Jacques,
TAYLOR Richard
Sustainable Development and Energy Transition in Europe and Asia
(Innovation and Technology Set – Volume 9)
CERDIN Jean-Luc, PERETTI Jean-Marie
The Success of Apprenticeships: Views of Stakeholders on Training and
Learning
(Human Resources Management Set – Volume 3)
DIDAY Edwin, GUAN Rong, SAPORTA Gilbert, WANG Huiwen,
Advances in Data Science
(Big Data, Artificial Intelligence and Data Analysis Set – Volume 4)
DOS SANTOS PAULINO Victor
Innovation Trends in the Space Industry
(Smart Innovation Set – Volume 25)
GUILHON Bernard
Venture Capital and the Financing of Innovation
(Innovation Between Risk and Reward Set – Volume 6)
MASSOTTE Pierre, CORSI Patrick
Complex Decision-Making in Economy and Finance
2019
AMENDOLA Mario, GAFFARD Jean-Luc
Disorder and Public Concern Around Globalization
BARBAROUX Pierre
Disruptive Technology and Defence Innovation Ecosystems
(Innovation in Engineering and Technology Set – Volume 5)
DOU Henri, JUILLET Alain, CLERC Philippe
Strategic Intelligence for the Future 1: A New Strategic and Operational
Approach
Strategic Intelligence for the Future 2: A New Information Function
Approach
FRIKHA Azza
Measurement in Marketing: Operationalization of Latent Constructs
FRIMOUSSE Soufyane
Innovation and Agility in the Digital Age
(Human Resources Management Set – Volume 2)
GAY Claudine, SZOSTAK Bérangère L.
Innovation and Creativity in SMEs: Challenges, Evolutions and Prospects
(Smart Innovation Set – Volume 21)
GORIA Stéphane, HUMBERT Pierre, ROUSSEL Benoît
Information, Knowledge and Agile Creativity
(Smart Innovation Set – Volume 22)
HELLER David
Investment Decision-making Using Optional Models
(Economic Growth Set – Volume 2)
HELLER David, DE CHADIRAC Sylvain, HALAOUI Lana, JOUVET Camille
The Emergence of Start-ups
(Economic Growth Set – Volume 1)
HÉRAUD Jean-Alain, KERR Fiona, BURGER-HELMCHEN Thierry
Creative Management of Complex Systems
(Smart Innovation Set – Volume 19)
LATOUCHE Pascal
Open Innovation: Corporate Incubator
(Innovation and Technology Set – Volume 7)
LEHMANN Paul-Jacques
The Future of the Euro Currency
LEIGNEL Jean-Louis, MÉNAGER Emmanuel, YABLONSKY Serge
Sustainable Enterprise Performance: A Comprehensive Evaluation Method
LIÈVRE Pascal, AUBRY Monique, GAREL Gilles
Management of Extreme Situations: From Polar Expeditions to Exploration-
Oriented Organizations
MILLOT Michel
Embarrassment of Product Choices 2: Towards a Society of Well-being
N’GOALA Gilles, PEZ-PÉRARD Virginie, PRIM-ALLAZ Isabelle
Augmented Customer Strategy: CRM in the Digital Age
NIKOLOVA Blagovesta
The RRI Challenge: Responsibilization in a State of Tension with Market
Regulation
(Innovation and Responsibility Set – Volume 3)
PELLEGRIN-BOUCHER Estelle, ROY Pierre
Innovation in the Cultural and Creative Industries
(Innovation and Technology Set – Volume 8)
PRIOLON Joël
Financial Markets for Commodities
QUINIOU Matthieu
Blockchain: The Advent of Disintermediation
RAVIX Joël-Thomas, DESCHAMPS Marc
Innovation and Industrial Policies
(Innovation between Risk and Reward Set – Volume 5)
ROGER Alain, VINOT Didier
Skills Management: New Applications, New Questions
(Human Resources Management Set – Volume 1)
SAULAIS Pierre, ERMINE Jean-Louis
Knowledge Management in Innovative Companies 1: Understanding and
Deploying a KM Plan within a Learning Organization
(Smart Innovation Set – Volume 23)
SERVAJEAN-HILST Romaric
Co-innovation Dynamics: The Management of Client-Supplier Interactions
for Open Innovation
(Smart Innovation Set – Volume 20)
SKIADAS Christos H., BOZEMAN James R.
Data Analysis and Applications 1: Clustering and Regression, Modeling-
estimating, Forecasting and Data Mining
(Big Data, Artificial Intelligence and Data Analysis Set – Volume 2)
Data Analysis and Applications 2: Utilization of Results in Europe and
Other Topics
(Big Data, Artificial Intelligence and Data Analysis Set – Volume 3)
VIGEZZI Michel
World Industrialization: Shared Inventions, Competitive Innovations and
Social Dynamics
(Smart Innovation Set – Volume 24)
2018
BURKHARDT Kirsten
Private Equity Firms: Their Role in the Formation of Strategic Alliances
CALLENS Stéphane
Creative Globalization
(Smart Innovation Set – Volume 16)
CASADELLA Vanessa
Innovation Systems in Emerging Economies: MINT – Mexico, Indonesia,
Nigeria, Turkey
(Smart Innovation Set – Volume 18)
CHOUTEAU Marianne, FOREST Joëlle, NGUYEN Céline
Science, Technology and Innovation Culture
(Innovation in Engineering and Technology Set – Volume 3)
CORLOSQUET-HABART Marine, JANSSEN Jacques
Big Data for Insurance Companies
(Big Data, Artificial Intelligence and Data Analysis Set – Volume 1)
CROS Françoise
Innovation and Society
(Smart Innovation Set – Volume 15)
DEBREF Romain
Environmental Innovation and Ecodesign: Certainties and Controversies
(Smart Innovation Set – Volume 17)
DOMINGUEZ Noémie
SME Internationalization Strategies: Innovation to Conquer New Markets
ERMINE Jean-Louis
Knowledge Management: The Creative Loop
(Innovation and Technology Set – Volume 5)
GILBERT Patrick, BOBADILLA Natalia, GASTALDI Lise,
LE BOULAIRE Martine, LELEBINA Olga
Innovation, Research and Development Management
IBRAHIMI Mohammed
Mergers & Acquisitions: Theory, Strategy, Finance
LEMAÎTRE Denis
Training Engineers for Innovation
LÉVY Aldo, BEN BOUHENI Faten, AMMI Chantal
Financial Management: USGAAP and IFRS Standards
(Innovation and Technology Set – Volume 6)
MILLOT Michel
Embarrassment of Product Choices 1: How to Consume Differently
PANSERA Mario, OWEN Richard
Innovation and Development: The Politics at the Bottom of the Pyramid
(Innovation and Responsibility Set – Volume 2)
RICHEZ Yves
Corporate Talent Detection and Development
SACHETTI Philippe, ZUPPINGER Thibaud
New Technologies and Branding
(Innovation and Technology Set – Volume 4)
SAMIER Henri
Intuition, Creativity, Innovation
TEMPLE Ludovic, COMPAORÉ SAWADOGO Eveline M.F.W.
Innovation Processes in Agro-Ecological Transitions in Developing
Countries
(Innovation in Engineering and Technology Set – Volume 2)
UZUNIDIS Dimitri
Collective Innovation Processes: Principles and Practices
(Innovation in Engineering and Technology Set – Volume 4)
VAN HOOREBEKE Delphine
The Management of Living Beings or Emo-management
2017
AÏT-EL-HADJ Smaïl
The Ongoing Technological System
(Smart Innovation Set – Volume 11)
BAUDRY Marc, DUMONT Béatrice
Patents: Prompting or Restricting Innovation?
(Smart Innovation Set – Volume 12)
BÉRARD Céline, TEYSSIER Christine
Risk Management: Lever for SME Development and Stakeholder
Value Creation
CHALENÇON Ludivine
Location Strategies and Value Creation of International
Mergers and Acquisitions
CHAUVEL Danièle, BORZILLO Stefano
The Innovative Company: An Ill-defined Object
(Innovation between Risk and Reward Set – Volume 1)
CORSI Patrick
Going Past Limits To Growth
D’ANDRIA Aude, GABARRET Inés
Building 21st Century Entrepreneurship
(Innovation and Technology Set – Volume 2)
DAIDJ Nabyla
Cooperation, Coopetition and Innovation
(Innovation and Technology Set – Volume 3)
FERNEZ-WALCH Sandrine
The Multiple Facets of Innovation Project Management
(Innovation between Risk and Reward Set – Volume 4)
FOREST Joëlle
Creative Rationality and Innovation
(Smart Innovation Set – Volume 14)
GUILHON Bernard
Innovation and Production Ecosystems
(Innovation between Risk and Reward Set – Volume 2)
HAMMOUDI Abdelhakim, DAIDJ Nabyla
Game Theory Approach to Managerial Strategies and Value Creation
(Diverse and Global Perspectives on Value Creation Set – Volume 3)
LALLEMENT Rémi
Intellectual Property and Innovation Protection: New Practices
and New Policy Issues
(Innovation between Risk and Reward Set – Volume 3)
LAPERCHE Blandine
Enterprise Knowledge Capital
(Smart Innovation Set – Volume 13)
LEBERT Didier, EL YOUNSI Hafida
International Specialization Dynamics
(Smart Innovation Set – Volume 9)
MAESSCHALCK Marc
Reflexive Governance for Research and Innovative Knowledge
(Responsible Research and Innovation Set – Volume 6)
MASSOTTE Pierre
Ethics in Social Networking and Business 1: Theory, Practice
and Current Recommendations
Ethics in Social Networking and Business 2: The Future and
Changing Paradigms
MASSOTTE Pierre, CORSI Patrick
Smart Decisions in Complex Systems
MEDINA Mercedes, HERRERO Mónica, URGELLÉS Alicia
Current and Emerging Issues in the Audiovisual Industry
(Diverse and Global Perspectives on Value Creation Set – Volume 1)
MICHAUD Thomas
Innovation, Between Science and Science Fiction
(Smart Innovation Set – Volume 10)
PELLÉ Sophie
Business, Innovation and Responsibility
(Responsible Research and Innovation Set – Volume 7)
SAVIGNAC Emmanuelle
The Gamification of Work: The Use of Games in the Workplace
SUGAHARA Satoshi, DAIDJ Nabyla, USHIO Sumitaka
Value Creation in Management Accounting and Strategic Management:
An Integrated Approach
(Diverse and Global Perspectives on Value Creation Set –Volume 2)
UZUNIDIS Dimitri, SAULAIS Pierre
Innovation Engines: Entrepreneurs and Enterprises in a Turbulent World
(Innovation in Engineering and Technology Set – Volume 1)
2016
BARBAROUX Pierre, ATTOUR Amel, SCHENK Eric
Knowledge Management and Innovation
(Smart Innovation Set – Volume 6)
BEN BOUHENI Faten, AMMI Chantal, LEVY Aldo
Banking Governance, Performance And Risk-Taking: Conventional Banks
Vs Islamic Banks
BOUTILLIER Sophie, CARRÉ Denis, LEVRATTO Nadine
Entrepreneurial Ecosystems (Smart Innovation Set – Volume 2)
BOUTILLIER Sophie, UZUNIDIS Dimitri
The Entrepreneur (Smart Innovation Set – Volume 8)
BOUVARD Patricia, SUZANNE Hervé
Collective Intelligence Development in Business
GALLAUD Delphine, LAPERCHE Blandine
Circular Economy, Industrial Ecology and Short Supply Chains
(Smart Innovation Set – Volume 4)
GUERRIER Claudine
Security and Privacy in the Digital Era
(Innovation and Technology Set – Volume 1)
MEGHOUAR Hicham
Corporate Takeover Targets
MONINO Jean-Louis, SEDKAOUI Soraya
Big Data, Open Data and Data Development
(Smart Innovation Set – Volume 3)
MOREL Laure, LE ROUX Serge
Fab Labs: Innovative User
(Smart Innovation Set – Volume 5)
PICARD Fabienne, TANGUY Corinne
Innovations and Techno-ecological Transition
(Smart Innovation Set – Volume 7)
2015
CASADELLA Vanessa, LIU Zeting, DIMITRI Uzunidis
Innovation Capabilities and Economic Development in Open Economies
(Smart Innovation Set – Volume 1)
CORSI Patrick, MORIN Dominique
Sequencing Apple’s DNA
CORSI Patrick, NEAU Erwan
Innovation Capability Maturity Model
FAIVRE-TAVIGNOT Bénédicte
Social Business and Base of the Pyramid
GODÉ Cécile
Team Coordination in Extreme Environments
MAILLARD Pierre
Competitive Quality and Innovation
MASSOTTE Pierre, CORSI Patrick
Operationalizing Sustainability
MASSOTTE Pierre, CORSI Patrick
Sustainability Calling
2014
DUBÉ Jean, LEGROS Diègo
Spatial Econometrics Using Microdata
LESCA Humbert, LESCA Nicolas
Strategic Decisions and Weak Signals
2013
HABART-CORLOSQUET Marine, JANSSEN Jacques, MANCA Raimondo
VaR Methodology for Non-Gaussian Finance
2012
DAL PONT Jean-Pierre
Process Engineering and Industrial Management
MAILLARD Pierre
Competitive Quality Strategies
POMEROL Jean-Charles
Decision-Making and Action
SZYLAR Christian
UCITS Handbook
2011
LESCA Nicolas
Environmental Scanning and Sustainable Development
LESCA Nicolas, LESCA Humbert
Weak Signals for Strategic Intelligence: Anticipation Tool for Managers
MERCIER-LAURENT Eunika
Innovation Ecosystems
2010
SZYLAR Christian
Risk Management under UCITS III/IV
2009
COHEN Corine
Business Intelligence
ZANINETTI Jean-Marc
Sustainable Development in the USA
2008
CORSI Patrick, DULIEU Mike
The Marketing of Technology Intensive Products and Services
DZEVER Sam, JAUSSAUD Jacques, ANDREOSSO Bernadette
Evolving Corporate Structures and Cultures in Asia: Impact
of Globalization
2007
AMMI Chantal
Global Consumer Behavior
2006
BOUGHZALA Imed, ERMINE Jean-Louis
Trends in Enterprise Knowledge Management
CORSI Patrick et al.
Innovation Engineering: the Power of Intangible Networks