Ibook - Pub Seasonal Adjustment Methods and Real Time Trend Cycle Estimation
Ibook - Pub Seasonal Adjustment Methods and Real Time Trend Cycle Estimation
Estela Bee Dagum
Silvia Bianconcini
Seasonal
Adjustment
Methods and
Real Time Trend-
Cycle Estimation
Statistics for Social and Behavioral Sciences
Series editor
Stephen E. Fienberg
Carnegie Mellon University
Pittsburgh
Pennsylvania
USA
Statistics for Social and Behavioral Sciences (SSBS) includes monographs and
advanced textbooks relating to education, psychology, sociology, political science,
public policy, and law.
Seasonal Adjustment
Methods and Real Time
Trend-Cycle Estimation
123
Estela Bee Dagum Silvia Bianconcini
Department of Statistical Sciences Department of Statistical Sciences
University of Bologna University of Bologna
Bologna, Italy Bologna, Italy
In order to assess the current stage of the business cycle at which the economy
stands, real time trend-cycle estimates are needed. The basic approach is that of
assessing the real time trend-cycle of major socioeconomic indicators (leading,
coincident, and lagging) using percentage changes, based on seasonally adjusted
data, calculated for months and quarters in chronological sequence. The main goal
is to evaluate the behavior of the economic indicators during incomplete phases
by comparing current contractions or expansions with corresponding phases in the
past. This is done by measuring changes of single time series (mostly seasonally
adjusted) from their standing at cyclical turning points with past changes over a
series of increasing spans. This differs from business cycle studies where cyclical
fluctuations are measured around a long-term trend to estimate complete business
cycles. The real time trend corresponds to an incomplete business cycle and is
strongly related to what is currently happening on the business cycle stage. Major
changes of global character in the financial and economic sector have introduced
high levels of variability in time series making difficult to detect the direction of
the short-term trend by simply looking at seasonally adjusted data, and the use of
trend-cycle data or smoothed seasonally adjusted series has been supported. Failure
in providing reliable real time trend-cycle estimates could give rise to dangerous
drift of the adopted policies. Therefore, a reliable estimation is of fundamental
importance.
This book includes two opening chapters, Chap. 1 which is a general introduction
and Chap. 2 on time series components. The remaining nine chapters are divided
in two parts, one on seasonal adjustment and the other on real time trend-cycle
estimation.
Since the input for trend-cycle estimation is seasonally adjusted data, Part I of this
book thoroughly discusses the definitions and concepts involved with three major
seasonal adjustment methods as follows:
Chapter 3. Seasonal adjustment, meaning, purpose, and methods.
Chapter 4. Linear smoothing or moving average seasonal adjustment methods.
Chapter 5. Seasonal adjustment based on ARIMA model decomposition:
TRAMO-SEATS.
vii
viii Preface
results are empirically corroborated with a set of leading, coincident, and lagging
indicators of the US economy.
Chapter 11 deals with the effects of the seasonal adjustment methods when the
real time trend is predicted with nonparametric kernel filters. The seasonal adjust-
ments compared are the two officially adopted by statistical agencies, X12ARIMA
and TRAMO-SEATS, applied to a sample of US leading, coincident, and lagging
indicators.
The eleven chapters have been written as complete as possible, and each one can
be read rather independently. We have also introduced a uniform notation all along
the chapters to facilitate the reading of the book as a whole.
This book will prove useful for graduate and final-year undergraduate courses
in econometrics and time series analysis and as a reference book for researchers
and practitioners in statistical agencies, other government offices, and business.
The prerequisites are a good knowledge of linear regression, matrix algebra, and
knowledge of ARIMA modeling.
We are indebted to participants and students who during many seminars and
presentations raised valuable questions answered in this book. Our most sincere
gratitude goes to our colleagues who encouraged us to write this book and all those
who provided many valuable and useful suggestions through lively discussions or
written comments. Our thanks also go to Veronika Rosteck, associate statistics
editor of Springer, for her sustained and valuable support while writing the book.
We are solely responsible for any errors and omissions.
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 1
1.1 Book Structure.. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 1
1.2 Part I: Seasonal Adjustment Methods . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 2
1.2.1 Smoothing Linear Seasonal Adjustment Methods . . . . . . . . 3
1.2.2 ARIMA Model-Based Seasonal Adjustment Method . . . . 9
1.2.3 Structural Time Series Models .. . . . . . . .. . . . . . . . . . . . . . . . . . . . 14
1.3 Part II: Real Time Trend-Cycle Estimation . . . . .. . . . . . . . . . . . . . . . . . . . 21
References .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 26
2 Time Series Components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 29
2.1 Time Series Decomposition Models.. . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 30
2.2 The Secular or Long-Term Trend.. . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 33
2.2.1 Deterministic Trend Models . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 34
2.2.2 Stochastic Trends . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 34
2.3 The Business Cycle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 36
2.3.1 Deterministic and Stochastic Models
for the Business Cycle . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 39
2.4 The Seasonal Variations.. . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 39
2.4.1 Seasonal Adjustment Methods .. . . . . . . .. . . . . . . . . . . . . . . . . . . . 41
2.5 Calendar Variations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 47
2.5.1 The Moving Holiday Component .. . . . .. . . . . . . . . . . . . . . . . . . . 47
2.5.2 The Trading Day Component . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 49
2.6 The Irregular Component . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 52
2.6.1 Redistribution of Outliers and Strikes .. . . . . . . . . . . . . . . . . . . . 52
2.6.2 Models for the Irregulars and Outliers .. . . . . . . . . . . . . . . . . . . . 53
References .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 56
xi
xii Contents
Glossary . . . . .. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. . . . . . . . . . . . . . . . . . . . 279
Acronyms
xv
xvi Acronyms
Abstract This chapter introduces the main topics of the book that are divided into
two parts: (I) seasonal adjustment methods and (II) real time trend-cycle estimation.
Most socioeconomic series and mainly indicators (leading, lagging, and coincident)
are affected by the impact of seasonal variations. In order to assess what is the
stage of the cycle at which the economy stands it is necessary to identify, estimate,
and remove the effect of seasonality from the data. The seasonal adjustment
methods discussed in Part I are grouped into three main categories: smoothing linear
methods, ARIMA model-based and structural time series models. There are three
seasonal adjustment softwares, one for each method, that are currently available.
Two are officially adopted by statistical agencies, named X12ARIMA and TRAMO-
SEATS, and the third major software, called STAMP, is mainly used in econometric
studies and for academic research. In Part II, the trend-cycle estimates from these
seasonal adjustment methods are analyzed and compared to a new set of trend-
cycle symmetric and asymmetric filters based on Reproducing Kernel Hilbert Space
(RKHS) methodology.
This book is divided into two parts: one on seasonal adjustment and the other on
real time trend-cycle estimation. It has eleven chapters as follows:
1. Introduction.
2. Time series components.
PART I: SEASONAL ADJUSTMENT.
3. Seasonal adjustment, meaning, purpose, and methods.
4. Linear smoothing or moving averages seasonal adjustment methods.
5. Seasonal adjustment based on ARIMA model decomposition: TRAMO-
SEATS.
6. Seasonal adjustment based on structural time series models.
PART II: REAL TIME TREND-CYCLE ESTIMATION.
7. Trend-cycle estimation.
8. Recent developments on nonparametric trend-cycle estimation.
A great deal of data in business, economics, and natural sciences occurs in the
form of time series where observations are dependent, and where the nature of
this dependence is of interest in itself. The time series is generally compiled for
consecutive and equal periods, such as weeks, months, quarters, and years.
From a statistical point of view, a time series is a sample realization of a
stochastic process, i.e., a process controlled by probability laws. The observations
made as the process continues indicate the way it evolves. In the analysis of
economic time series, Persons [48] was the first to distinguish four types of
evolution, namely: (a) trend, (b) cycle, (c) seasonal variations, and (d) irregular
fluctuations.
The trend corresponds to a variation in some defined sense persisting over a long
period of time, that is, a period which is long in relation to the cycle. In some cases,
it is a steady growth; while in others, the trend may move downward as well as
upward. The cycle, usually referred to as the business cycle, is a quasi-periodic
oscillation characterized by alternating periods of expansion and contraction. The
seasonal variations represent the composite effect of climatic and institutional events
which repeat more or less regularly each year. These three types of fluctuation are
assumed to follow systematic patterns (they are the signal of the process).
On the other hand, the irregulars are unforeseeable movements related to events
of all kinds. In general, they have a stable random appearance but, in some
series, extreme values may be present. These extreme values or outliers have
identifiable causes, e.g., floods, unseasonal weather, and strikes; and, therefore, can
be distinguished from the much smaller irregular variations.
Among all these components, the influence of the seasonal fluctuations in the
human activity has been felt from earlier times. The organization of society, the
means of production and communication, the habits of consumption, and other
social and religious events have been strongly conditioned by both climatic and
conventional seasons. The seasonal variations in agriculture, the high pre-Easter and
pre-Christmas retail sales, and the low level of winter construction are all general
knowledge.
The main causes of seasonality, such as climatic and institutional factors, are
mainly exogenous to the economic system and cannot be controlled or modified
by the decision makers in the short run. The impact of seasonality in the economic
activity, however, is usually not independent on the stages of the business cycle. It
is well-known for example, that the seasonal unemployment among adult males is
1.2 Part I: Seasonal Adjustment Methods 3
much larger during periods of recession than during periods of prosperity. Another
main feature of seasonality is that the phenomenon repeats with certain regularity
every year but it may evolve. The latter is mainly due to systematic intervention
of the government and to technological and institutional changes as well; and it is
more the general case than the exception for economic time series. Therefore, the
assumption of stable seasonality, i.e., of seasonality being represented by a strictly
periodic function, can be used as a good approximation for few series only. In effect,
even in the extreme cases of those activities where seasonality is mainly caused
by climatic conditions, e.g., agriculture, fishing, forestry, the seasonal variations
change, for weather itself measured by the temperature, and quantity of precipitation
changes.
Once the assumption of stable seasonality is abandoned, new assumptions must
be made regarding the nature of its evolution; if seasonality changes, is it slowly or
rapidly? Is it gradually or abruptly? Is it in a deterministic or a stochastic manner?
Today, the most widely accepted hypothesis is that seasonality moves gradually,
slowly, and in a stochastic manner. A third characteristic of seasonality is that the
phenomenon can be separated from other forces (trend, cycle, and irregulars) that
influence the movement of an economic time series and can be measured.
The seasonal variations are distinguished from trend by their oscillated character,
from the cycle by being confined within the limits of an annual period, and from the
irregulars, by the fact of being systematic. Most socioeconomic indicators, such as
leading, coincident, and lagging, are affected by the impact of seasonal variations.
In order to assess what is the stage of the current economic conditions it is necessary
to identify, estimate, and remove the effect of seasonality. Seasonal adjustment
methods can be grouped into three main categories: smoothing linear methods,
ARIMA model-based and structural time series models. There are three seasonal
adjustment softwares, one for each method, that are currently available but only
two are officially adopted by statistical agencies, named X12ARIMA and TRAMO-
SEATS. The third major software, called STAMP, is mainly used in econometric
studies and for academic research. These topics are largely discussed in Chaps. 4, 5,
and 6, respectively, and only brief summaries are given below.
The best known and most often applied seasonal adjustment methods are based on
smoothing linear filters or moving averages applied sequentially by adding (and
subtracting) one observation at a time. These methods assume that the time series
components change through time in a stochastic manner. Given a time series yt ; t D
1; : : : ; n, for any t far removed from both ends, say m < t < n m, the seasonally
adjusted value yat is obtained by applying, in a moving manner, a symmetric average
4 1 Introduction
X
m X
m
yat D W.B/yt D wj Bj yt D wj ytj ; (1.1)
jDm jDm
where the weights wj are symmetric, that is, wj D wj , and the length of the average
is 2m C 1. For current and recent data (n m < t n), a symmetric linear filter
cannot be applied, and therefore truncated asymmetric filters are used. For example,
for the last available observation yn , the seasonally adjusted value is given by
X
m
yan D W0 .B/yn D w0;j ynj : (1.2)
jD0
The asymmetric filters are time-varying in the sense that different filters are applied
to the m first and last observations. The end estimates are revised as new observa-
tions are added because of: (a) the new innovations and (b) the differences between
the symmetric and asymmetric filters. The estimates obtained with symmetric filters
are often called “final.”
The development of electronic computers contributed to major improvements
in seasonal adjustment based on moving averages and facilitated their massive
application. In 1954, Julius Shiskin of the US Bureau of Census developed a
software called Method I, based mainly on the works of Macauley [44] already
being used by the US Federal Reserve Board. Census Method I was followed by
Census Method II and 11 more experimental versions (X1, X2, : : :, X11). The best
known and widely applied was the Census Method II-X11 variant developed by
Shiskin et al. [51], but produced poor seasonally adjusted data at the end of the
series which is of crucial importance to assess the direction of the short-term trend
and the identification of real time turning points in the economy.
Estela Bee Dagum [13] working at Statistics Canada developed the X11ARIMA
mainly to correct for this serious limitation and, later, Findley et al. [26] developed
X12ARIMA that offers a regARIMA option to estimate deterministic components,
such as trading day variations, moving holiday effects and outliers, simultaneously
with the ARIMA model for extrapolation. It also includes new diagnostic tests and
spectral techniques to assess the goodness of the results. The X12ARIMA method
is today the one most often applied by statistical agencies in the world. The US
Bureau of Census continued research on the development of seasonal adjustment
methods, and recently produced a version called X13ARIMA-SEATS which enables
the estimation of the seasonal component either via linear filters as those available
in X12ARIMA or based on an ARIMA decomposition model.
All the seasonal adjustment methods based on moving averages discussed in this
book have nonlinear elements. Hence, the seasonally adjusted total of an aggregated
series is not equal to the algebraic sum of the seasonally adjusted series that enter
1.2 Part I: Seasonal Adjustment Methods 5
into the aggregation. The main causes of nonlinearity are generated by:
1. a multiplicative decomposition model for the unobserved components,
2. the identification and replacement of outliers,
3. the ARIMA extrapolations, and
4. automatic selection of the moving average length for the estimation of the trend-
cycle and seasonality.
The properties of the combined linear filters applied to estimate the various
components were originally calculated by Young [54] for the standard option
of Census II-X11 variant. Later, Dagum et al. [18] calculated and analyzed all
possible filter combinations of Census II-X11 and X11ARIMA. Given the fact that
X11ARIMA and X12ARIMA are methods based on the Census Method II-X11
[51], widely discussed in Chap. 4, a brief summary is given here. The reader is
referred to Ladiray and Quenneville [41] for a very detailed description of the X11
method.
The X11 method assumes that the main components of a time series follow a
multiplicative or an additive model, that is,
(multiplicative model) yt D TCt St It ,
(additive model) yt D TCt C St C It ,
where yt stands for the original series, TCt for the trend-cycle, St for the seasonality,
and It for the irregulars.
There are no mixed models in this program, such as yt D TCt St C It or other
possible combinations. The estimation of the components is made with different
kinds of smoothing linear filters. The filters are applied sequentially and the whole
process is repeated twice.
The X11 variant produced seasonal factor forecasts for each month which were
given by
1
Sj;tC1 D Sj;t C .Sj;t Sj;t1 /; j D 1; : : : ; 12;
2
where j is the month and t the current year. The use of seasonal factor forecasts was
very popular in statistical bureaus till the end of the 1970s. In 1975, Dagum [12]
proposed the use of concurrent seasonal factors obtained from data that included
the most recent values and this was adopted by Statistics Canada. Gradually, other
statistical agencies applied concurrent seasonal factors and this is now the standard
practice.
The final trend-cycle is obtained by a 9-, 13-, or 23-term Henderson moving
average applied to the final monthly seasonally adjusted series (a 5- or 7-term
Henderson filter for quarterly data). The selection of the appropriate filter is made
on the basis of a preliminary estimate of the I=C ratio (the ratio of the average
absolute month-to-month change in the irregular to that in the trend-cycle). A
9-term is applied to less irregular series and a 23-term to highly irregular series.
The linear smoothing filters applied by Census Method II-X11 variant to produce
6 1 Introduction
seasonally adjusted data can be classified according to the distribution of their set
of weights in symmetric (two-sided) and asymmetric (one-sided). The symmetric
moving averages are used to estimate the component values that fall in the middle
of the span of the average, say 2m C 1, and the asymmetric moving averages, to the
m first and last observations.
The sum of the weights of both kinds of filters is one, and thus, the mean of
the original series is unchanged in the filtering process. The sum of the weights of
a filter determines the ratio of the mean of the smoothed series to the mean of the
unadjusted series assuming that these means are computed over periods long enough
to ensure stable results. It is very important in filter design that the filter does not
displace in time the components of the output relative to those of the input. In other
words, the filter must not introduce phase shifts.
Symmetric moving averages introduce no time displacement for some of the
components of the original series and a displacement of 180ı for others. A phase
shift of 180ı is interpreted as a reverse in polarity which means that maxima are
turned into minima and vice versa. In other words, peaks (troughs) in the input
are changed into troughs (peaks) in the output. For practical purposes, however,
symmetric moving averages act as though the time displacement is null. This is so
because the sinusoids that will have a phase shift of 180ı in the filtering process are
cycles of short periodicity (annual or less) and moving averages tend to suppress or
significantly reduce their presence in the output. In spectral analysis, the phase is a
dimensionless parameter that measures the displacement of the sinusoid relative to
the time origin. Because of the periodic repetition of the sinusoid, the phase can be
restricted to ˙180ı . The phase is a function of the frequency of the sinusoid, being
the frequency equal to the reciprocal of the length of time or period required for one
complete oscillation.
It is inherent to any moving average procedure that the m first and last points
of an original series cannot be smoothed with the same set of symmetric 2m C 1
weights applied to the middle values.
The X11 method uses one-sided filters to smooth these end points. Moving
averages with asymmetric weights are bounded to introduce phase shifts for all the
components of the original series. This is a very undesirable characteristic since it
may cause, for example, a business cycle in the smoothed series to turn down before
or after it actually does in the unadjusted data. The asymmetric weights associated
with the Henderson filters no longer estimate a cubic within their span, but only a
linear trend-cycle within the length of the filters.
The X11ARIMA is a modified version of the Census Method II-X11 variant that
was developed by Estela Bee Dagum, at Statistics Canada, mainly to produce a
better current seasonal adjustment of series with seasonality that changes rapidly in
a stochastic manner. The latter characteristic is often found in main socioeconomic
indicators, e.g., retail trade, imports and exports, unemployment, and so on.
A detailed description of the X11ARIMA method of seasonal adjustment is given
1.2 Part I: Seasonal Adjustment Methods 7
where s denotes the periodicity of the seasonal component (equal to 12 for monthly
series and 4 for quarterly data), B is the backshift operator, such that Byt D yt1 ,
p .B/ D .1 1 B p Bp / is the nonseasonal autoregressive (AR) operator of
order p, ˚P .Bs / D .1 ˚1 Bs ˚P BPs / is the seasonal autoregressive (SAR)
operator of order P, q .B/ D .1 1 B q Bq / is the nonseasonal moving
average (MA) operator of order q, Q .Bs / D .11 Bs Q BQs / is the seasonal
moving average (SMA) operator of order Q, and the at ’s are i.i.d. with mean zero
and variance a2 (white noise). The .1 B/d .1 Bs /D term implies nonseasonal
differencing of order d and seasonal differencing of order D. If d D D D 0 (no
differencing), it is common to replace yt in (1.3) by deviations from its mean. The
general multiplicative model (1.3) is said to be of order .p; d; q/.P; D; Q/s.
The X12ARIMA is today the most often applied seasonal adjustment method by
statistical agencies. It was developed by Findley et al. [26], and it is an enhanced
version of the X11ARIMA method.
The major modifications concern: (1) extending the automatic identification and
estimation of ARIMA models for the extrapolation option to many more than the
three models available in X11ARIMA and (2) estimating trading day variations,
moving holidays, and outliers in what is called regARIMA.
RegARIMA consists of regression models with autoregressive integrated moving
average (ARIMA) errors. More precisely, they are models in which the mean
function of the time series (or its logs) is described by a linear combination of
regressors, and the covariance structure of the series is that of an ARIMA process.
8 1 Introduction
If no regressors are used, indicating that the mean is assumed to be zero, the
regARIMA model reduces to an ARIMA model.
Whether or not special problems requiring the use of regressors are present in
the series to be adjusted, a fundamentally important use of regARIMA models is
to extend the series with forecasts (and backcasts) in order to improve the seasonal
adjustments of the most recent (and the earliest) data. Doing this reduces problems
inherent in the trend-cycle estimation and asymmetric seasonal averaging processes
of the type used by the X11 method near the ends of the series. The provision of
this extension was the most important improvement offered by the X11ARIMA
program. Its theoretical and empirical benefits have been documented in many
publications, such as Dagum [14], Bobbit and Otto [5], and the references therein.
The X12ARIMA method has all the seasonal adjustment capabilities of the
X11ARIMA variant. The same seasonal and trend moving averages are available,
and the program still offers the X11 calendar and holiday adjustment routines
incorporated in X11ARIMA. But several important new options have been included.
The modeling module is designed for regARIMA model building with seasonal
socioeconomic time series. To this end, several categories of predetermined regres-
sion variables are available, including trend constants or overall means, fixed
seasonal effects, trading day effects, holiday effects, pulse effects (additive outliers),
level shifts, temporary change outliers, and ramp effects. User-specified regression
variables can also be included in the models.
The specification of a regARIMA model requires specification of both the regres-
sion variables to be included in the model and the type of ARIMA model for the
regression errors (i.e., the order .p; d; q/.P; D; Q/s ). Specification of the regression
variables depends on the user knowledge about the series being modeled. Identi-
fication of the ARIMA model for the regression errors follows well-established
procedures based on examination of various sample autocorrelation and partial
autocorrelation functions produced by the X12ARIMA program. Once a regARIMA
model has been specified, X12ARIMA estimates its parameters by maximum like-
lihood using an Iterative Generalized Least Squares (IGLS) algorithm. Diagnostic
checking involves examination of residuals from the fitted model for signs of model
inadequacy. X12ARIMA produces several standard residual diagnostics for model
checking, as well as provides sophisticated methods for detecting additive outliers
and level shifts.
Trading day effects occur when a series is affected by the different day-of-the-
week compositions of the same calendar month in different years. Trading day
effects can be modeled with seven variables that represent (no. of Mondays), . . . ,
(no. of Sundays) in month t. Bell and Hillmer [2] proposed a better parameterization
of the same effects using six variables defined as (no. of Mondays) (no. of
Sundays), . . . , (no. of Saturdays) (no. of Sundays), along with a seventh variable
for length of month (LOM) or its deseasonalized version, the leap-year regressor
(lpyear). In X12ARIMA the six variables are called the tdnolpyear variables.
Instead of using a seventh regressor, a simpler and often better way to handle
multiplicative leap-year effects is to re-scale the February values of the original
time series before transformation to m N Feb yt =mt , where yt is the original time series
1.2 Part I: Seasonal Adjustment Methods 9
before transformation, mt is the length of month t (28 or 29), and m N Feb D 28:25 is
the average length of February. If the regARIMA model includes seasonal effects,
these can account for the length-of-month effect except in Februaries, so the trading
day model only has to deal with the leap-year effect. When this is done, only the
tdnolpyear variables need be included in the model. X12ARIMA allows explicit
choice of either approach, as well as an option (td) that makes a default choice
of how to handle length-of-month effects. When the time series being modeled
represents the aggregation of some daily series (typically unobserved) over calendar
months they are called monthly flow series. If the series instead represents the value
of some daily series at the end of the month, called a monthly stock series, then
different regression variables are appropriate.
Holiday effects in a monthly flow series arise from holidays whose dates vary
over time if: (1) the activity measured by the series regularly increases or decreases
around the date of the holiday and (2) this affects 2 (or more) months depending on
the date the holiday occurs each year. Effects of holidays with a fixed date, such as
Christmas, are indistinguishable from fixed seasonal effects. Easter effects are the
most frequently found holiday effects in American and European economic time
series, since the date of Easter Sunday varies between March 22 and April 25. Labor
Day and Thanksgiving also are found in American and Canadian time series.
X12ARIMA provides four other types of regression variables to deal with abrupt
changes in the level of a series of temporary or permanent nature: additive outliers
(AO), level shifts (LS), temporary changes (TC), and ramps. Identifying the location
and nature of potential outliers is the object of the outlier detection methodology
implemented. This methodology can be used to detect AOs, TCs, and LSs (not
ramps). Any outlier detected is automatically added to the model as regression
variable. Prespecified AOs, LSs, TCs, and ramps are actually simple forms of
interventions as discussed by Box and Tiao [6].
Figure 1.1 shows the original values of the time series US Unemployment Rate
for Males (16 years and over) observed from January 1992 to December 2013
together with the final seasonally adjusted series obtained with the X12ARIMA
software using the default option. Figure 1.2 exhibits the corresponding seasonal
factors.
Peter Burman [7] was the first to develop a seasonal adjustment method based on
ARIMA model decomposition, named SIGEX (Signal Extraction). Later, working
on the same topic Hillmer and Tiao [35] developed what is known as ARIMA
model-based seasonal adjustment, largely discussed in Bell and Hillmer [2].
An ARIMA model is identified from the observed data and by imposing certain
restrictions, models for each component are derived. Since the components are
unknown, to obtain a unique solution Hillmer and Tiao proposed a canonical
decomposition which has the property of maximizing the variance of the irregulars
10 1 Introduction
Original series
12
10
8
6
4 X12ARIMA seasonally adjusted series
Fig. 1.1 Original and X12ARIMA (default option) seasonally adjusted US Unemployment Rate
for Males (16 years and over) observed from January 1992 to December 2013
115
105
95
90
Fig. 1.2 Seasonal factors of the US Unemployment Rate for Males (16 years and over) estimated
using the X12ARIMA software (default option)
1.2 Part I: Seasonal Adjustment Methods 11
yt D x0t ˇ C t (1.4)
where B is the backshift operator, .B/ and .B/ are finite polynomials in B, and at
is assumed to be normally identically distributed, that is, NID(0, a2 ). The polyno-
mial .B/ contains the unit roots associated with regular and seasonal differencing,
as well as the polynomial with stationary autoregressive roots (and complex unit
roots, if present). .B/ denotes the invertible moving average polynomial. The
regression variables xt can be given by the user or generated by the program. In the
latter case, the variables are for trading day variations, Easter effects, and outliers.
Outliers reflect the effect of some special, non regularly repetitive events, such
as implementation of a new regulation, major political or economic changes,
modifications in the way the variable is measured, occurrence of a strike or natural
disaster, etc. Consequently, discordant observations and various types of abrupt
12 1 Introduction
changes are often present in times series data. The location and type of outliers are
“a priori” unknown. TRAMO uses an improved Chen and Liu [8] type procedure
for outlier detection and correction. The effect of an outlier is modeled by yt D
! .B/It .t0 / C t , where .B/ is a quotient of polynomials in B that models the type
of outlier (its effect over time), and It .t0 / is an indicator variable of the occurrence
of the outlier, that is,
1 if t D t0
It .t0 / D
0 otherwise
! represents the impact of the outlier at time t0 and t is the outlier free series which
follows the model specified in Eq. (1.5). In the automatic detection and correction,
by default, three types of outliers can be considered
(1) Additive Outlier (AO): .B/ D 1;
(2) Transitory Change (TC): .B/ D 1=.1 ıB/, where, by default, ı D 0:7;
(3) Level Shift (LS): .B/ D 1=.1 B/.
One can also include a fourth type of outlier, that is, Innovational Outlier (IO) for
which .B/ D .B/=.B/ı.B/ that resembles a shock in the innovations at .
The procedure followed by TRAMO consists of:
1. Pretest for the log/level specification, that is, a trimmed range mean regression
test is performed to select whether the original series will be transformed into log
or maintain the level.
2. Pretest for trading days and Easter effects made with regressions using the
default model for the noise. If the model is subsequently changed, the test is
redone. Thus, the output file of TRAMO may say at the beginning “Trading
day is not significant,” but the final model estimated may contain trading day
variables (or vice versa).
3. Automatic detection and correction of outliers. The program has a facility for
detecting outliers and for removing their effects. The outliers can be entered by
the user or they can be automatically identified by the program.
4. Automatic model selection. The program further performs an automatic identi-
fication of the ARIMA model. This is done in two steps. The first one yields
the nonstationary polynomial .1 B/d .1 Bs /D of model (1.5). This is done by
iterating on a sequence of AR and ARIMA models (with constant) which have
a multiplicative structure when the data is seasonal. The procedure is based on
results of Tiao and Tsay [52] and Tsay [53]. Regular and seasonal differences
are obtained up to a maximum order of .1 B/2 .1 Bs /. The second step
identifies an ARMA model for the stationary series (modified for outliers and
regression-type effects) following the Hannan–Rissanen procedure [31], with
an improvement which consists of using the Kalman filter instead of zeros to
calculate the first residuals in the computation of the estimator of the variance
of the innovations of model (1.5). Finally, the program combines the facilities
for automatic identification and correction of outliers and automatic ARIMA
1.2 Part I: Seasonal Adjustment Methods 13
model identification just described so that it has an option for automatic model
identification of a nonstationary series in presence of outliers.
5. Diagnostic checks. The main diagnostic tests are on the residuals. The estimated
residuals aO t are analyzed to test the hypothesis that at are independent and
identically normally distributed with zero mean and constant variance a2 .
Besides inspection of the residual graph, the corresponding sample autocorre-
lation function is examined. The lack of residual autocorrelation is tested using
the Ljung–Box test statistics, and skewness and kurtosis tests are applied to
test for normality of the residuals. Specific out-of-sample forecast tests are also
performed to evaluate if forecasts behave in agreement with the model.
6. Optimal forecasts. If the diagnostics are satisfied, the model is used to compute
optimal forecasts for the series, together with their mean square error (MSE).
These are obtained using the Kalman filter applied to the original series.
TRAMO is to be used as a preadjustment process to eliminate all the determinis-
tic components such that the residual is a linearized series that can be modeled with
an ARIMA process. The latter is decomposed by SEATS in stochastic trend-cycle,
seasonality, and irregulars. Both programs can handle routine applications to a large
number of series and provide a complete model-based solution to the problems of
forecasting, interpolation, and signal extraction for nonstationary time series.
SEATS belongs to the class of procedures based on ARIMA models for the
decomposition of time series into unobserved components and consists of the
following steps:
1. ARIMA model estimation. SEATS starts by fitting an ARIMA model to a series
not affected by deterministic components, such as trading day variations, moving
holidays, and outliers. Let yt denote this linearized series, and consider an
additive decomposition model (multiplicative if applied to the log transformation
of yt ). The complete model can be written in detailed form as
yt D Tt C Ct C St C It ; (1.8)
where Tt denotes the trend component, Ct the cycle, St represents the seasonal
component, and It the irregulars.
The decomposition is done in the frequency domain. The spectrum (or
pseudospectrum) is partitioned into additive spectra, associated with the different
components which are determined mainly from the AR roots of the model. The
trend component represents the long-term evolution of the series and displays
a spectral peak at frequency 0, whereas the seasonal component captures the
spectral peaks at seasonal frequencies (e.g., for monthly data these are 0.524,
1.047, 1.571, 2.094, 2.618, and 3.142). The cyclical component captures periodic
fluctuations with period longer than a year, associated with a spectral peak
for a frequency between 0 and (2=s), and short-term variation associated
with low order MA components and AR roots with small moduli. Finally, the
irregular component captures white noise behavior, and hence has a flat spectrum.
The components are determined and fully derived from the structure of the
(aggregate) ARIMA model (1.7) for the linearized series directly identified from
the data. The program is aimed at monthly or quarterly frequency data and the
maximum number of observations that can be processed is 600.
One important assumption is that of orthogonality among the components, and
each one will have in turn an ARIMA model. In order to identify the components,
the canonical decomposition is used which implies that the variance of the
irregulars is maximized, whereas the trend, seasonal, and cycle are as stable as
possible.
The canonical condition on the trend, seasonal, and cyclical components
identifies a unique decomposition, from which the ARIMA models for the
components are obtained (including the component innovation variances).
Figure 1.3 shows the original values of the time series US Unemployment Rate for
Males (16 years and over) observed from January 1992 to December 2013 together
with the final seasonally adjusted series obtained with the TRAMO-SEATS software
using the default option. Figure 1.4 exhibits the corresponding seasonal factors.
The structural time series approach involves decomposing a series into components
which have a direct interpretation. A structural model, also called unobserved
component (UC) model, consists of a number of stochastic linear processes that
stand for the trend, cycle, seasonality, and remaining stationary dynamic features
in an observed time series. The trend component typically represents the longer
1.2 Part I: Seasonal Adjustment Methods 15
12
Original series
TRAMO-SEATS seasonally adjusted series
10
8
6
4
Fig. 1.3 Original and TRAMO-SEATS (default option) seasonally adjusted US Unemployment
Rate for Males (16 years and over) observed from January 1992 to December 2013
115
105
95
90
Fig. 1.4 Seasonal factors of the US Unemployment Rate for Males (16 years and over) estimated
using the TRAMO-SEATS software (default option)
term movement of the series and is often specified as a smooth function of time.
The recurring but persistently changing patterns within the years are captured by the
seasonal component. In economic time series, the cycle component can represent the
16 1 Introduction
dynamic features associated with the business cycle. Each component is represented
as an ARIMA process. Hence, the typical parameters that need to be estimated are
the variances of the innovations driving the components, and a selection of other
coefficients associated with the ARIMA processes.
The key to handling structural time series models is the state space form with
the state of the system representing the various unobserved components such
as trend, cycle, and seasonality. The estimate of the unobservable state can be
updated by means of a filtering procedure as new observations become available.
Predictions are made by extrapolating these estimated components into the future,
while smoothing algorithms give the best estimate of the state at any point within the
sample. The statistical treatment can therefore be based on the Kalman filter and its
related methods. A detailed discussion of the methodological and technical concepts
underlying structural time series models is given in Chap. 6. The reader is also
referred to the monographs by Harvey [32], Kitagawa and Gersh [39], and Kim and
Nelson [37]. Durbin and Koopman [24] also provide an exhaustive overview on state
space methods for time series, whereas an introduction is given by Commandeur and
Koopman [11].
We now introduce the basic structural time series model with explicit specifica-
tions for each component. For quarterly or monthly observations yt ; t D 1; : : : ; n, it
is given by
yt D Tt C St C It ; (1.9)
where Tt stands for the trend, St for seasonality, and It is the irregular component
generally assumed to be NID.0; I2 /. All these components are stochastic and the
disturbances driving them are assumed to be mutually uncorrelated. The definitions
of the components are given below, but a full explanation of the underlying rationale
can be found in Harvey [32]. The effectiveness of structural time series models
compared to ARIMA-type models is discussed in Harvey et al. [33].
The trend component can be specified in many different ways, but the most
common specification is an extension of the random walk trend by including a
stochastic drift component
where the disturbance series aTt is normally independently distributed. The initial
values T1 and ˇ1 are treated as unknown coefficients. When Tt is given by (1.10),
Harvey [32] defines yt D Tt C It as the local linear trend model. In case
2 D 0, the
trend (1.10) reduces to an I.1/ process given by TtC1 D Tt Cˇ1 CaTt , where the drift
ˇ1 is fixed. This specification is referred as a random walk plus drift process. If in
addition T2 D 0, the trend reduces to the deterministic linear trend TtC1 D T1 C ˇ1 t.
When T2 D 0 and
2 > 0, the trend Tt in (1.10) remains an I.2/ process.
Common specifications for the seasonal component St are provided in the
following.
1.2 Part I: Seasonal Adjustment Methods 17
Dummy seasonal. Since the seasonal effects should sum to zero over a year, a
basic model for this component is given by
X
s1
St D Stj C !t t D s; : : : ; n; (1.12)
jD1
where s denotes the number of “seasons” in a year. In words, the seasonal effects are
allowed to change over time by letting their sum over the previous year be equal to a
random disturbance term !t with mean zero and variance !2 . Writing out Eq. (1.12)
in terms of the lag operator B gives
When s is even, the sine term disappears for j D s=2, and so the number of
trigonometric parameters, the Sj and Sj , is always .s1/=2, which is the same as the
number of coefficients in the seasonal dummy formulation. A seasonal pattern (1.14)
is the sum of [s=2] cyclical components that can be reformulated as follows:
! ! !
Sj;t cos j sin j Sj;t1 !j;t
D C (1.15)
Sj;t sin j cos j Sj;t1 !j;t
with !j;t and !j;t ; j D 1; : : : ; Œs=2
, being zero mean white noise processes which are
uncorrelated with each other and have a common variance !2 . Note that, when s is
even, the component at j D s=2 collapses to
yt D zTt ˛t C It ; t D 1; : : : ; n (1.16)
˛tC1 D t ˛t C Rt "t : (1.17)
Formal proofs of the Kalman filter can be found in Anderson and Moore [1], Harvey
[32], and Durbin and Koopman [24].
The model (1.9) may provide a good description of the time series, although it
may sometimes be necessary to include additional components. Seasonal economic
time series are often affected by trading day effects and holiday effects which can
influence the dynamic behavior of the series. Hence, a set of explanatory variables
need to be included in the model to capture specific (dynamic) variations in the time
series, as well as outliers and breaks. Therefore, Koopman and Ooms [40] suggest
to extend model (1.9) as follows:
Figure 1.5 shows the original values of the time series US Unemployment Rate
for Males (16 years and over) observed from January 1992 to December 2013
together with the final seasonally adjusted series obtained with the STAMP software.
Figure 1.6 exhibits the corresponding seasonal component estimated using the
default option (additive decomposition).
20 1 Introduction
12
Original series
STAMP seasonally adjusted series
10
8
6
4
Fig. 1.5 Original and STAMP (default option) seasonally adjusted US Unemployment Rate for
Males (16 years and over) observed from January 1992 to December 2013
0.8
0.4
0.0
-0.4
Fig. 1.6 Seasonal effects of the US Unemployment Rate for Males (16 years and over) estimated
using the STAMP software (default option)
1.3 Part II: Real Time Trend-Cycle Estimation 21
The second part of this books deals with recent developments on trend-cycle
estimation on real time. It comprises Chaps. 7–11. The basic approach to the
analysis of current economic conditions, known as recession and recovery analysis,
is that of assessing the real time trend-cycle of major socioeconomic indicators
(leading, coincident, and lagging) using percentage changes, based on seasonally
adjusted units, calculated for months and quarters in chronological sequence. The
main goal is to evaluate the behavior of the economic indicators during incomplete
phases by comparing current contractions or expansions with corresponding phases
in the past. This is done by measuring changes of single time series (mostly
seasonally adjusted) from their standing at cyclical turning points with past changes
over a series of increasing spans. This differs from business cycle studies where
cyclical fluctuations are measured around a long-term trend to estimate complete
business cycles. The real time trend corresponds to an incomplete business cycle
and is strongly related to what is currently happening on the business cycle
stage.
In recent years, statistical agencies have shown an interest in providing trend-
cycle or smoothed seasonally adjusted graphs to facilitate recession and recovery
analysis. Among other reasons, this interest originated from the recent crisis and
major economic and financial changes of global nature which have introduced more
variability in the data. The USA entered in recession in December 2007 until June
2009, and this has produced a chain reaction all over the world, with great impact in
Europe and China. There are no evidence of a fast recovery as in previous recession:
the economic growth is sluggish and with high levels of unemployment. It has
become difficult to determine the direction of the short-term trend (or trend-cycle)
as traditionally done by looking at month to month (quarter to quarter) changes
of seasonally adjusted values, particularly to assess the upcoming of a true turning
point. Failure in providing reliable real time trend-cycle estimates could give rise
to dangerous drift of the adopted policies. Therefore, a consistent prediction is of
fundamental importance. It can be done by means of either univariate parametric
models or nonparametric techniques. Since the majority of the statistical agencies
use seasonally adjusted software, such as the Census II-X11 method and its variants
X11/12ARIMA and X13, this book deals with the nonparametric techniques.
The linear filter developed by Henderson [34] is the most frequently applied
and has the property that fitted to exact cubic functions will reproduce their
values, and fitted to stochastic cubic polynomials it will give smoother results than
those estimated by ordinary least squares. The properties and limitations of the
Henderson filters have been extensively discussed by many authors, among them,
[10, 14, 15, 20, 21, 25, 29, 36, 41, 43], and Dagum and Bianconcini [16] who were
the first to represent the Henderson filter using Reproducing Kernel Hilbert Space
(RKHS) methodology.
22 1 Introduction
At the beginning and end of the sample period, the Henderson filter of length, say,
2m C 1 cannot be applied to the m data points, hence only asymmetric filters can
be used. The estimates of the real time trend are then subject to revisions produced
by the innovations brought by the new data entering in the estimation and the time-
varying nature of the asymmetric filters, in the sense of being different for each of
the m data points.
The asymmetric filters applied to the first and last m observations associated with
the Henderson filter were developed by Musgrave [45] on the basis of minimizing
the mean squared revision between the final estimates, obtained with the symmetric
Henderson weights, and preliminary estimates from the asymmetric weights, subject
to the constraint that the sum of these weights is equal to one. The assumption made
is that at the end of the series, the seasonally adjusted values do not follow a cubic
polynomial, but a linear trend-cycle plus a purely random irregular. Several authors
have studied the statistical properties and limitations of the Musgrave filters, among
others, [4, 20, 21, 23, 30, 42, 49], and Dagum and Bianconcini [16, 17] were the
first to introduce an RKHS representation of the asymmetric filters developed by
Musgrave [45].
Dagum [15] developed a nonlinear smoother to improve on the classical Hen-
derson filter. The NonLinear Dagum Filter (NLDF) results from applying the
13-term symmetric Henderson filter (H13) to seasonally adjusted series where
outliers and extreme observations have been replaced and which have been extended
with extrapolations from an ARIMA model. The main purpose of the ARIMA
extrapolations is to reduce the size of the revisions of the most recent estimates
whereas that of extreme values replacement is to reduce the number of unwanted
ripples produced by H13. An unwanted ripple is a 10-month cycle (identified by the
presence of high power at ! D 0:10 in the frequency domain) which, due to its
periodicity, often leads to the wrong identification of a true turning point. In fact, it
falls in the neighborhood between the fundamental seasonal frequency and its first
harmonic. On the other hand, a high frequency cycle is generally assumed to be
part of the noise pertaining to the frequency band 0:10 ! < 0:50. The problem
of the unwanted ripples is specific of H13 when applied to seasonally adjusted
series. Studies by Dagum, Chhab, and Morry [15] and Chhab et al. [9] showed the
superior performance of the NLDF respect to both structural and ARIMA standard
parametric trend-cycle models applied to series with different degrees of signal-
to-noise ratios. In another study, the good performance of the NLDF is shown
relative to nonparametric smoothers, namely: locally weighted regression (LOESS),
Gaussian kernel, cubic smoothing spline, and supersmoother [19].
Dagum and Luati developed in 2009 what they called a Cascade Linear Filter
(CLF), and distinguished between the Symmetric (SLF) and the Asymmetric Linear
Filter (ALF).
A linear filter offers many advantages over a nonlinear one. For one, its
application is direct and hence does not require knowledge of ARIMA model
identification. Furthermore, linear filtering preserves the crucial additive constraint
by which the trend of an aggregated variable should be equal to the algebraic
addition of its component trends, thus avoiding the problem of direct versus indirect
1.3 Part II: Real Time Trend-Cycle Estimation 23
adjustment. Finally, the properties of a linear filter concerning signal passing and
noise suppression can always be compared to those of other linear filters by means
of spectral analysis.
These authors studied the properties of the new CLF relative to 13-term
Henderson filter (H13) and have shown that the CLF is an optimal trend-cycle
estimator among a variety of second and higher order kernels restricted to be of
the same length. The theoretical properties were presented by means of spectral
analysis, whereas the empirical properties were evaluated on a large sample of real
time series pertaining to various socioeconomic areas and with different degrees
of variability. It should be noticed that the theoretical properties of CLF cannot be
compared with those of NLDF since the latter is data dependent.
From another perspective Dagum and Bianconcini in 2008 and 2013 were the
first to introduce an RKHS representation of the symmetric Henderson and the
asymmetric filters developed by Musgrave.
An RKHS is a Hilbert space characterized by a kernel that reproduces, via
an inner product, every function of the space or, equivalently, a Hilbert space of
real valued functions with the property that every point evaluation functional is
bounded and linear. Parzen [47] was the first to introduce an RKHS approach in time
series analysis applying the famous Loéve theorem by which there is an isometric
isomorphism between the closed linear span of a second order stationary stochastic
process and the RKHS determined by its covariance function. Parzen demonstrated
that the RKHS approach provides a unified framework to three fundamental
problems related with: (1) least squares estimation; (2) minimum variance unbiased
estimation of regression parameters; and (3) identification of unknown signals
perturbed by noise. Parzen’s approach is parametric, and basically consists of
estimating the unknown signal by generalized least squares in terms of the inner
product between the observations and the covariance function. A nonparametric
approach of the RKHS methodology was developed by De Boor and Lynch [22]
in the context of cubic spline approximation. Later, Kimeldorf and Wahba[38]
exploited both developments and treated the general spline smoothing problem from
a RKHS stochastic equivalence perspective. These authors proved that minimum
norm interpolation and smoothing problems with quadratic constraints imply an
equivalent Gaussian stochastic process.
In this book we show how nonparametric estimators can be transformed into
kernel functions of order two, that are probability densities, and from which
corresponding hierarchies of estimators are derived. The density function provides
the “initial weighting shape" from which the higher order kernels inherit their
properties. This kernel representation enables the comparison of estimators based
on different smoothing criteria, and has important consequences in the derivation
of the asymmetric filters which can be applied to the most recent observations. In
particular, those obtained by means of RKHS are shown to have superior properties
from the view point of signal passing, noise suppression, and revisions relative to
the classical ones.
24 1 Introduction
X
n
yO t D wt;i yi ; t D 1; 2; : : : ; n (1.18)
iD1
where wt;i denotes the weights to be applied to the observations yi to get the estimate
yO t for each point in time t; t D 1; : : : ; n.
Once a symmetric span 2mC1 of the neighborhood has been selected, the weights
for the observations corresponding to points falling out of the neighborhood of any
target point are null or approximately null. Hence, the estimates of the n2m central
observations are obtained by applying 2mC1 symmetric weights to the observations
neighboring the target point. That is,
X
m
yO t D wj ytCj t D m C 1; : : : ; n m:
jDm
.m C 1/
f0H .t/ D W..m C 1/t/; t 2 Œ1; 1
k
R .mC1
where k D .mC1/ W. j/dj and j D .m C 1/t.
These authors also found that the biweight density function gives almost
equivalent results to those obtained with the exact density function without the need
to be calculated any time that the Henderson smoother length changes. Another
important advantage is that the biweight density function belongs to the well-known
Beta distribution family, that is,
r
f .t/ D .1 jtjr /s IŒ1;1
.t/;
2B.s C 1; 1r/
R1
where B.a; b/ D 0 ta1 .1 t/b1 dt with a; b > 0 is the Beta function.
The orthonormal polynomials associated to the biweight function are the Jacobi
polynomials, for which explicit expressions for computation are available and their
properties have been widely studied in literature.
For the m first and last filters we discuss time-varying bandwidth parameters
since the asymmetric filters are time-varying. Three specific criteria of bandwidth
selection are chosen based on the minimization of
1. the distance between the transfer functions of asymmetric and symmetric filters,
2. the distance between the gain functions of asymmetric and symmetric filters, and
3. the phase shift function over the domain of the signal.
We deal only with the reduction of revisions due to filter changes that depend on
how close the asymmetric filters are respect to the symmetric one [14, 15] and do not
consider revisions introduced by the innovations in the new data. Another important
aspect discussed deals with the capability of the asymmetric filters to signal the
upcoming of a true turning point that depends on the time delay for its identification.
This is obtained by calculating the number of months (quarters) it takes for the
last trend-cycle estimate to signal a true turning point in the same position of the
final trend-cycle data. An optimal asymmetric filter should have a time path that
converges fast and monotonically to the final estimate as new observations are added
to the series.
It is shown in Chap. 10 that, applied to a set of US leading, coincident, and
lagging indicators, the new set of asymmetric kernel filters reduced by one half the
size of the revisions and by one third the time delay to detect the June 2009 turning
point relative to the Musgrave filters.
26 1 Introduction
References
1. Anderson, B. D. O., & Moore, J. B. (1979). Optimal filtering. Englewood Cliffs: Prentice-Hall.
2. Bell, W. H., & Hillmer, S. C. (1984). Issues involved with the seasonal adjustment of economic
time series. Journal of Business and Economic Statistics, 2, 291–320.
3. Berlinet, A. (1993). Hierarchies of higher order kernels. Probability Theory and Related Fields,
94, 489–504.
4. Bianconcini, S., & Quenneville, B. (2010). Real time analysis based on reproducing kernel
Henderson filters. Estudios de Economia Aplicada, 28(3), 553–574.
5. Bobbit, L., & Otto, M. C. (1990). Effects of forecasts on the revisions of seasonally
adjusted values using the X11 seasonal adjustment procedure. In Proceedings of the American
Statistical Association Business and Economic Statistics Session (pp. 449–453).
6. Box, G. E. P., & Tiao, G. C. (1975). Intervention analysis with applications to economics and
environmental problems. Journal of the American Statistical Association, 70, 70–79.
7. Burman, J. P. (1980). Seasonal adjustment by signal extraction. Journal of the Royal Statistical
Society Series A, 143, 321–337.
8. Chen, C., & Liu, L. M. (1993). Joint estimation of model parameters and outlier effects in time
series. Journal of the American Statistical Association, 88, 284–297.
9. Chhab, N., Morry, M., & Dagum, E. B. (1999). Results on alternative trend-cycle estimators
for current economic analysis. Estadistica, 49–51, 231–257.
10. Cholette, P. A. (1981). A comparison of various trend-cycle estimators. In O. D. Anderson &
M. R. Perryman (Eds.), Time series analysis (pp. 77–87). Amsterdam: North Holland.
11. Commandeur, J. J. F., & Koopman, S. J. (2007). An introduction to state space time series
analysis. Oxford: Oxford University Press.
12. Dagum, E. B. (1975). Seasonal factor forecasts from ARIMA models. In 40th Session
International Statistical Institute, Warsaw (Vol. 3, pp. 203–216).
13. Dagum, E. B. (1978). Modelling, forecasting and seasonally adjusting economic time series
with the X-11 ARIMA method. The Statistician, 27(3), 203–215.
14. Dagum, E. B. (1988). The X-11 ARIMA/88 seasonal adjustment method-foundations and user’s
manual. Ottawa, ON, Canada: Time Series Research and Analysis Centre, Statistics Canada.
15. Dagum, E. B. (1996). A new method to reduce unwanted ripples and revisions in trend-cycle
estimates from X11ARIMA. Survey Methodology, 22, 77–83.
16. Dagum, E. B., & Bianconcini, S. (2008). The Henderson smoother in reproducing kernel
Hilbert space. Journal of Business and Economic Statistics, 26(4), 536–545.
17. Dagum, E. B., & Bianconcini, S. (2013). A unified probabilistic view of nonparametric
predictors via reproducing kernel Hilbert spaces. Econometric Reviews, 32(7), 848–867.
18. Dagum, E. B., Chhab, N., & Chiu, K. (1996). Derivation and analysis of the X11ARIMA and
census X11 linear filters. Journal of Official Statistics, 12(4), 329–347.
19. Dagum, E. B., & Luati, A. (2000). Predictive performance of some nonparametric linear and
nonlinear smoothers for noisy data. Statistica, LX(4), 635–654.
20. Dagum, E. B., & Luati, A. (2009). A cascade linear filter to reduce revisions and turning points
for real time trend-cycle estimation. Econometric Reviews, 28(1–3), 40–59.
21. Dagum, E. B., & Luati, A. (2012). Asymmetric filters for trend-cycle estimation. In W. R.
Bell, S. H. Holan, & T. S. McElroy (Eds.), Economic time series: Modeling and seasonality
(pp. 213–230). Boca Raton: Chapman & Hall.
22. De Boor, C., & Lynch, R. (1966). On splines and their minimum properties. Journal of
Mathematics and Mechanics, 15, 953–969.
23. Doherty, M. (2001). The surrogate Henderson filters in X-11. Australian and New Zealand
Journal of Statistics, 43(4), 385–392.
24. Durbin, J., & Koopman, S. J. (2001). Time series analysis by state space methods. Oxford:
Oxford University Press.
References 27
25. Findley, D. F., & Martin, D. E. K. (2006). Frequency domain analysis of SEATS and X-11/X-
12-ARIMA seasonal adjustment filters for short and moderate length time series. Journal of
Official Statistics, 22, 1–34.
26. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., & Chen, B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
27. Gomez, V., & Maravall, A. (1994). Estimation, prediction and interpolation for nonstationary
series with the Kalman filter. Journal of the American Statistical Association, 89, 611–624.
28. Gomez, V., & Maravall, A. (1996). Program TRAMO and SEATS: Instructions for Users.
Working paper 9628. Service de Estudios, Banco de Espana.
29. Gray, A., & Thomson, P. (1996). Design of moving-average trend filters using fidelity and
smoothness criteria. In P. M. Robinson & M. Rosenblatt (Eds.), Time series analysis (in
memory of E.J. Hannan) Vol. II (Vol. 115, pp. 205–219). Springer Lecture notes in statistics.
New York: Springer.
30. Gray, A. G., & Thomson, P. J. (2002). On a family of finite moving-average trend filters for the
ends of series. Journal of Forecasting, 21, 125–149.
31. Hannan, E. J., & Rissanen, J. (1982). Recursive estimation of mixed autoregressive moving
average order. Biometrika, 69, 81–94.
32. Harvey, A. C. (1989). Forecasting, structural time series models and the Kalman filter.
Cambridge: Cambridge University Press.
33. Harvey, A. C., Koopman, S. J., & Penzer, J. (1998). Messy time series: A unified approach. In
T. B. Fomby & R. Carter Hill (Eds.), Advances in econometrics (Vol. 13, pp. 103–143). New
York, NY, USA: JAI Press.
34. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
35. Hillmer, S. C., & Tiao, G. C. (1982). An ARIMA model-based approach to seasonal
adjustment. Journal of the American Statistical Association, 77, 63–70.
36. Kenny, P., & Durbin, J. (1982). Local trend estimation and seasonal adjustment of economic
and social time series. Journal of the Royal Statistical Society A, 145, 1–41.
37. Kim, C. J., & Nelson, C. R. (1999). State space models with regime switching. Cambridge,
MA: MIT Press.
38. Kimeldorf, G. S., & Wahba, G. (1970). A correspondence between Bayesian estimation on
stochastic processes and smoothing by splines. Annals of Mathematical Statistics, 41, 495–
502.
39. Kitagawa, G., & Gersch, W. (1996). Smoothness priors analysis of time series. New York:
Springer.
40. Koopman, S. J., & Ooms, M. (2011). Forecasting economic time series using unobserved
components time series models. In M. Clements & D. F. Hendry (Eds.), Oxford handbook
of economic forecasting. Oxford: Oxford University Press.
41. Ladiray, D., & Quenneville, B. (2001). Seasonal adjustment with the X11 method. Lecture
notes in statistics (Vol. 158). New York: Springer.
42. Laniel, N. (1985). Design Criteria for the 13-term Henderson End Weights. Working paper.
Methodology Branch, Statistics Canada, Ottawa.
43. Loader, C. (1999). Local regression and likelihood. New York: Springer.
44. Macaulay, F. R. (1931). The smoothing of time series. New York: National Bureau of Economic
Research.
45. Musgrave, J. (1964). A Set of End Weights to End All End Weights. Working paper. U.S.
Bureau of Census, Washington, DC.
46. Newbold, P., & Granger, C. W. J. (1974). Experience with forecasting univariate time series
and the combination of forecasts (with discussion). Journal of the Royal Statistical Society.
Series A, 137, 131–165.
47. Parzen, E. (1959). Statistical Inference on Time Series by Hilbert Space Methods. Technical
Report No. 53, Statistics Department, Stanford University, Stanford, CA.
48. Persons, W. M. (1919). Indices of business conditions. Review of Economic Statistics, 1, 5–107.
28 1 Introduction
49. Quenneville, B., Ladiray, D., & Lefrancois, B. (2003). A note on Musgrave asymmetrical trend-
cycle filters. International Journal of Forecasting, 19(4), 727–734.
50. Reid, D. J. (1975). A review of short-term projection techniques. In H. A. Gorden (Ed.),
Practical aspects of forecasting (pp. 8–25). London: Operational Research Society.
51. Shiskin, J., Young, A. H., & Musgrave, J. C. (1967). The X-11 Variant of the Census Method
II Seasonal Adjustment Program. Technical Paper 15 (revised). US Department of Commerce,
Bureau of the Census, Washington, DC.
52. Tiao, G. C., & Tsay, R. S. (1983). Consistency properties of least squares estimates of
autoregressive parameters in Arma models. The Annals of Statistics, 11, 856–871.
53. Tsay, R. S. (1984). Regression models with time series errors. Journal of the American
Statistical Association, 79, 118–124.
54. Young, A. H. (1968). A linear approximation to the Census and BLS seasonal adjustment
methods. Journal of the American Statistical Association, 63, 445–457.
Chapter 2
Time Series Components
yt D Tt C Ct C St C It ; t D 1; : : : n (2.1)
where yt denotes the observed series at time t, Tt the long-term trend, Ct the business
cycle, St seasonality, and It the irregulars.
If there is dependence among the latent components, this relationship is specified
through a multiplicative model
yt D Tt Ct St It ; t D 1; : : : n (2.2)
usually assigned to the seasonal component. The trading day component is usually
considered as negligible in quarterly series and even more so in yearly data.
Another important calendar variation is the moving holiday or moving-festival
component. That component is associated with holidays which change date from
year to year, e.g., Easter, causing a displacement of activity from 1 month to the
previous or the following month. For example, an early date of Easter in March or
early April can cause an important excess of activity in March and a corresponding
short-fall in April, in variables associated with imports, exports, and tourism.
Under models (2.1) and (2.2), the trading day and moving festival components (if
present) are implicitly part of the irregular. In 1965, Young developed a procedure
to estimate trading day variations which was incorporated in the X11 seasonal
adjustment method [39] and its subsequent versions, the X11ARIMA [17, 18] and
X12ARIMA [21] methods. The latter two versions also include models to estimate
moving holidays, such as Easter.
If the new components are present, the additive decomposition model becomes
yt D Tt C Ct C St C Dt C Ht C It (2.3)
where Dt and Ht denote the trading day and moving holiday components, respec-
tively. Similarly, the multiplicative decomposition model becomes
yt D Tt Ct St Dt Ht It (2.4)
yt D t C et (2.5)
where t and et are referred to as the signal and the noise, according to the electrical
engineering terminology. The signal t comprises all the systematic components of
models (2.1) and (2.4), i.e., Tt ; Ct ; St ; Dt , and Ht .
Model (2.5) is classical in signal extraction where the problem is to find the best
estimates of the signal t given the observations yt corrupted by noise et . The best
estimates are usually defined as minimizing the mean square error.
32 2 Time Series Components
Finally, given its fundamental role in time series modeling, we summarize the
well-known decomposition theorem due to Herman Wold [41].
A stochastic process fYt g is second order stationary or weakly stationary if the
first two moments are not time dependent, that is, the mean and the variance are
constant, and the autocovariance function depends only on the time lag and not on
the time origin:
E.Yt / D (2.6)
Yt D Zt C Vt (2.9)
where
1
X 1
X
2
Zt D j atj ; 0 D 1; j <1 (2.10)
jD0 jD1
where is the constant mean of process fYt g and f˛j g; fˇj g are mutually uncorre-
lated white noise processes. The process fVt g is called deterministic because it can
be predicted in the future without error from a single realization by means of an
infinite linear combination of past values.
Wold theorem demonstrates that the property of stationarity is strongly related
to that of linearity. It provides a justification for autoregressive moving average
(ARMA) models [5] and some extensions, such as the autoregressive integrated
moving average (ARIMA) and regression-ARIMA models (regARIMA).
2.2 The Secular or Long-Term Trend 33
The concept of trend is used in economics and other sciences to represent long-
term smooth variations. The causes of these variations are often associated with
structural phenomena such as population growth, technological progress, capital
accumulation, and new practices of business and economic organization. For most
economic time series, the trends evolve smoothly and gradually, whether in a
deterministic or stochastic manner. When there is a sudden change of level and/or
slope this is referred to as a structural change. However, it should be noticed that
series at a higher levels of aggregation are less susceptible to structural changes. For
example, a technological change is more likely to produce a structural change for
some firms than for the whole industry.
The identification and estimation of the secular or long-term trend have posed
serious challenges to statisticians. The problem is not of statistical or mathematical
character but originates from the fact that the trend is a latent (non-observable)
component and its definition as a long-term smooth movement is statistically vague.
The concept of a long-period is relative, since a trend estimated for a given series
may turn out to be just a long business cycle as more years of data become available.
To avoid this problem statisticians have used two simple solutions. One is to estimate
the trend and the business cycles together, calling it the trend-cycle. The other
solution is to estimate the trend over the whole series, and to refer to it as the longest
non-periodic variation.
It should be kept in mind that many systems of time series are redefined every
15 years or so in order to maintain relevance. Hence, the concept of long-term trend
loses importance. For example, the Canadian system of Retail and Wholesale Trade
series was redefined in 1989 to adopt the 1980 Standard Industrial Classification
(SIC), and in 2003 to conform to the North American Industrial Classification
System (NAICS), following the North American Free Trade Agreement. The
following examples illustrate the need for such reclassifications. The 1970 Standard
Industrial Classification (SIC) considered computers as business machines, e.g.,
cash registers and desk calculators. The 1980 SIC rectified the situation by creating
a class for computers and other goods and services. The last few decades witnessed
the birth of new industries involved in photonics (lasers), bio-engineering, nano-
technology, and electronic commerce. In the process, new professions emerged, and
classification systems had to keep up with these new realities.
There is a large number of deterministic and stochastic models which have
been proposed for trend estimation (see Dagum and Dagum [15]). Deterministic
models are based on the assumption that the trend can be well approximated by
mathematical functions of time, whereas stochastic trends models assume that the
trend can be better modeled by applying finite differences of low order together with
autoregressive components and moving average errors.
34 2 Time Series Components
yt D Tt C et ; (2.12)
where Tt denotes the trend and fet g is assumed to follow a stationary process, often
white noise. The polynomial trend can be written as
Tt D ˛0 C ˛1 t C C ˛p tp (2.13)
Stochastic models are appropriate when the trend is assumed to follow a nonstation-
ary stochastic process. The nonstationarity is modeled with finite differences of low
order (cf. [24] and [35]).
A typical stochastic trend model often used in structural time series modeling is
the so-called random walk with constant drift. In the classical notation, the model is
Cubic trend
15000
10000
5000 Annual Real Gross Domestic product
Fig. 2.1 Cubic trend fitted to the annual values of the Real Gross Domestic Product (billions of
dollars)
where Tt denotes the trend, ˇ a constant drift, and faTt g is a normal white noise
process. Solving the difference equation (2.14) and assuming aT0 D 0, we obtain
1
X
Tt D ˇt C 1 aTt D ˇt C aTtj ; t D 1; : : : ; n; (2.15)
jD0
which show that a random walk with constant drift consists of a linear deterministic
trend plus a nonstationary infinite moving average.
Another type of stochastic trend belongs to the ARIMA (p; d; q) class, where p
is the order of the autoregressive polynomial, q is the order of the moving average
polynomial, and d the order of the finite difference operator .1 B/. The backshift
operator B is such that Bn Tt D Ttn . The ARIMA(p; d; q) model is written as
which means that after applying first order differences twice, the transformed series
can be modeled by an autoregressive process of order one.
to hold less capital and to be less concerned about liquidity positions. This, in turn,
is thought to have been a factor in encouraging increased debt levels and a reduction
in risk premium required by investors.
The period of the Great Moderation ranges between 1987 and 2007, and it
is characterized by predictable policy, low inflation, and modest business cycles.
However, at the same time various regions in the world have experienced prolonged
depressions, most dramatically the economic crisis in former Eastern Bloc countries
following the end of the Soviet Union in 1991. For several of these countries the
period 1989–2010 has been an ongoing depression, with real income still lower
than in 1989. In economics a depression is a more severe downturn than a recession,
which is seen by economists as part of a normal business cycle.
Considered a rare and extreme form of recession, a depression is characterized
by its length, and by abnormally large increases in unemployment, falls in the
availability of credit, shrinking output and investment, numerous bankruptcies,
significantly reduced amounts of trade and commerce, especially international,
as well as highly volatile relative currency value fluctuations most often due to
devaluations.
In 1946, economists Arthur F. Burns and Wesley C. Mitchell [9] provided the new
standard definition of business cycles in their book “Measuring Business Cycles”:
“Business cycles are a type of fluctuation found in the aggregate economic activity
of nations that organize their work mainly in business enterprises: a cycle consists of
expansions occurring at about the same time in many economic activities, followed
by similarly general recessions, contractions, and revival which merge into the
expansion phase of the next cycle; in duration business cycles vary from more than
one year to ten to twelve years; they are not divisible into shorter cycles of similar
characteristics with amplitudes approximating their own.”
In 1954 Schumpeter [38] stated that an economic cycle has four stages : (1)
expansion (increases in production and prices, low interests rates); (2) crisis (stock
exchanges crash and multiple bankruptcies of firms occur); (3) recession (drops
in prices and in output, high interests rates); and (4) recovery (stocks recover
because of the fall in prices and incomes). In this model, recovery and prosperity are
associated with increases in productivity, consumer confidence, aggregate demand,
and prices.
In a broad sense, there have been two ways by which economic and business
cycles have been studied, one analyzing complete cycles and the other, studying the
behavior of the economic indicators during incomplete phases by comparing current
contractions or expansions with corresponding phases in the past in order to assess
current economic conditions.
Because completed economic and business cycles are not directly observable, for
their identification and estimation, it is needed to remove from seasonally adjusted
series a long-term trend. This was originally done by means of a long moving
average using the Bry and Boschan method [7] adopted by the National Bureau of
Economic Research (NBER). Later, other methods were developed, such as those
of Hodrick and Prescott [28], Baxter and King [2], and Buttherworth [11] filters.
38 2 Time Series Components
Similarly to the trend, the models for cyclical variations can be deterministic or
stochastic. Deterministic models often consist of sine and cosine functions of
different amplitude and periodicities. For example, denoting the cycle by ct , a
deterministic model is
2
X
Ct D ˛j cos.j t/ C ˇj sin.j t/ ; (2.18)
jD1
where 1 D 2=60 and 2 D 2=40. Model (2.18) takes into consideration two
dominant cycles found in the European and American economies, those of 60 and
40 months, respectively.
Stochastic models of the ARIMA type, involving autoregressive models of order
2 with complex roots, have also been used to model the trend-cycle. For example,
where Ct denotes the cycle, faCt g is assumed Gaussian white noise, and the
following conditions apply to the parameters: 1 C 2 < 1, 2 1 < 1, and
1 < 1 < 0 (see [5]).
largely exogenous to the economic system but can be partially offset by human
intervention. For example, seasonality in money supply can be controlled by central
bank decisions on interest rates. In other cases, the seasonal effects can be offset
by international and inter-regional trade. To some extent seasonality can evolve
through technological and institutional changes. For example, the developments
of appropriate construction materials and techniques made it possible to continue
building in winter. The development of new crops, which better resist cold and
dry weather, has influenced the seasonal pattern. The partial or total replacement
of some crops by chemical substitutes, e.g., substitute of sugar, vanilla, and other
flavors, reduces seasonality in the economy.
As for institutional change, the extension of the academic year to include some
summer months affected the seasonal pattern of unemployment for the population
of 15–25 years of age. Similarly, the practice of spreading holidays over the whole
year impacted on seasonality.
The changing industrial mix of an economy also transforms the seasonal pattern,
because some industries are more seasonal than others. In particular, economies
which diversify and depend less on “primary” industries (e.g., fishing and agricul-
ture) typically become less seasonal.
Another important characteristic of seasonality is the feasibility of its identifi-
cation even if it is a latent (not directly observable) variable. The identification
and estimation of seasonality, however, is not done independently on the other
components affecting the time series under study, that is, trend, business cycle,
trading day variations, moving holidays, and irregulars. The seasonal variations can
be distinguished from the trend by their oscillatory character, from the business
cycle by having annual periodicity, and from the irregulars by being systematic.
Seasonality entails large costs to society and businesses. One cost is the necessity
to build warehouses to store inventories of goods to be sold as consumers require
them, for example, grain elevators. Another cost is the under-use and over-use
of the factors of production: capital and labor. Capital in the form of unused
equipment, buildings, and land during part of the year has to be financed regardless.
For example, this is the case in farming, food processing, tourism, and electrical
generation. The cold climate increases the cost of buildings and infrastructure, e.g.,
roads, transportation systems, water and sewage systems, schools, and hospitals; not
to mention the damage to the same caused by the action of ice. The labor force is
over-used during the peak seasons of agriculture and construction for example; and,
under-used in trough seasons sometimes leading to social problems.
In order to determine whether a series contains seasonality, it is sufficient to
identify at least 1 month (or quarter) which tends to be systematically higher or
lower than other month.
For illustrative purposes, Fig. 2.2 exhibits the seasonal variations affecting the
US New Orders for Durable Goods (NODG) from February 1992 till December
2013 that are characterized by three main peaks and two main troughs. The peaks
are observed in March, June and, in recent years, also in December when orders are
highest, whereas the troughs correspond to the months of January and July.
2.4 The Seasonal Variations 41
110
100
95
90
85
Fig. 2.2 Seasonal pattern of US New Orders for Durable Goods (January 1992–December 2013)
The seasonal pattern measures the relative importance of the months of the year.
The constant 100 % represents an average month or a nonseasonal month. The peak
month in March 1993 is 108.6 or 8.6 % larger than on an average month; the trough
month is July with 83.2 in 1992, almost 17 % lower than on an average month.
The seasonal amplitude, the difference between the peak and trough months of the
seasonal pattern, has changed from 25.4 % in 1992 to 19.3 % in 2013. It is apparent
that the seasonal amplitude is decreasing since 2005.
The seasonal adjustment methods developed so far are based on univariate time
series decomposition models with no causal explanation. It is difficult to classify
existing methods into mutually exclusive categories. However, it is possible to group
the majority of seasonal adjustment methods into two main classes: one based on
moving averages or linear filters and the other on explicit models with a small
number of parameters for each component. These methods will be extensively
discussed in the following chapters and only a brief summary is provided in this
section.
42 2 Time Series Components
The best known and most often applied seasonal adjustment methods are based on
moving averages or smoothing linear filters applied sequentially by adding (and
subtracting) one observation at a time. These methods assume that the time series
components change through time in a stochastic manner. Given a time series, yt ; t D
1; : : : ; n, for any t far removed from both ends, say mC1 t nm, the seasonally
adjusted value yat is obtained by application of a symmetric moving average W.B/,
as follows:
X
m
yat D W.B/yt D wj ytj (2.20)
jDm
where the weights wj are symmetric, that is, wj D wj , and the length of the average
is 2m C 1.
For current and recent data (n m < t n) a symmetric linear filter cannot be
applied, and therefore truncated asymmetric filters are used. For example, for the
last available observation yn , the seasonally adjusted value is given by
X
m
yan D W0 .B/yn D w0;j ynj : (2.21)
jD0
The asymmetric filters are time-varying in the sense that different filters are
applied for the m C 1 first and last observations. The end estimates are revised
as new observations are added because of: (1) the new innovations and (2) the
differences between the symmetric and asymmetric filters. The estimates obtained
with symmetric filters are often called “final.”
The development of electronic computers contributed to major improvements
in seasonal adjustment based on moving averages and facilitated their massive
application. In 1954, Julius Shiskin of the US Bureau of Census developed a
software called Method I, based mainly on the works of Macauley [33] already
being used by the US Federal Reserve Board. Census Method I was followed by
Census Method II and eleven more experimental versions (X1 , X2,: : :, X11). The
best known and widely applied was the Census Method II-X11 variant developed by
Shiskin et al. [39], but produced poor seasonally adjusted data at the end of the series
which is of crucial importance to assess the direction of the short-term trend and the
identification of turning points in the economy. Estela Bee Dagum developed in
1978 [16] a variant she called the X11ARIMA to correct for this serious limitation.
2.4 The Seasonal Variations 43
The best known model-based methods are: (1) the regression methods with global
or locally deterministic models for each component, and (2) stochastic model-based
methods that use ARIMA models.
yt D S t C et ; t D 1; : : : ; n
X
s X
s
St D j djt subject to j D 0 (2.22)
jD1 jD1
here the dj ’s are dummy variables taking a value of unity in season j and a value of
zero otherwise; fet g WN.0; e2 / is a white noise process with zero mean, constant
variance, and non-autocorrelated. The seasonal parameters j ’s can be interpreted
as the expectation of Yt in each season and represent the seasonal effects of the
series under question. Since the parameters are constant, the seasonal pattern repeats
exactly over the years, that is,
St D Sts : (2.23)
2.4 The Seasonal Variations 45
Equation (2.22), often included in econometric models, assumes that the effects
of the seasonal variations are deterministic because they can be predicted with
no error. However, for most economic and social time series, seasonality is not
deterministic but changes gradually in a stochastic manner. One representation of
stochastic seasonality is to assume in regression (2.22) that the j ’s are random
variables instead of constant. Hence, the relationship given in Eq. (2.23) becomes
where f!t g WN.0; !2 / and E.!t et / D 0; 8t. The stochastic seasonal balance
P
constraint is here given by s1jD0 Stj D !t with expected value equal to zero.
Equation (2.24) assumes seasonality to be generated by a nonstationary stochas-
tic process. In fact, it is a random walk process and can be made stationary by
seasonal differencing .1 Bs /, where B denotes the backward shift operator. In
reduced form, the corresponding regression model is a linear combination of white
noise processes of maximum lag s. Since .1 Bs / D .1 B/.1 C B P C C Bs1 /,
s1 j
model-based seasonal adjustment methods attribute only .B/ D jD0 B to the
seasonal component leaving .1 B/ as part of a nonstationary trend. Thus, the
corresponding seasonal stochastic model can be written as
.B/St D !t : (2.25)
Equation (2.25) represents a seasonality that evolves in an erratic manner and it was
included in seasonal adjustment methods based on structural model decomposition
(see, e.g., [23] and [31]). This seasonal behavior is seldom found in observed time
series where seasonality evolves smoothly through time. To represent this latter, [27]
introduced a major modification
where the left-hand side of Eq. (2.26) follows an invertible moving average process
of maximum order s 1, denoted by MA.s 1/. Equation (2.26) is discussed
extensively in [3] and can be generalized easily by replacing the MA part with
stationary invertible autoregressive moving average (ARMA) processes of the [5]
type. Depending on the order of the ARMA process a large variety of evolving
stochastic seasonal variations can be modeled.
Other seasonal models used in regression methods, where the components are
assumed to be local polynomials of time, are based on a smoothness criterion. For
example, given yt D SP
P t C et ; t D 1; : : : ; n, restrictions are imposed on St , such that
n 2 2
tD1 .S t S ts / and t Œ.B/St
be small. The solution is given by minimizing a
weighted linear combination of both criteria as follows:
X
n X
n X
n
.yt St /2 C ˛ .St Sts /2 C ˇ ŒS.B/St
2 (2.27)
tD1 tD1 tD1
46 2 Time Series Components
where the ill-posed problem of choosing the ˛ and ˇ parameters was solved by
Akaike [1] using a Bayesian model.
Stochastic model-based methods were mainly developed during the 1980s following
two different approaches. One, originally known as seasonal adjustment by SIGEX
(Signal Extraction), was developed by Burman [8] and as ARIMA model-based
seasonal adjustment by Hillmer and Tiao [27], largely discussed by Bell and Hillmer
[3]. The other is referred to as structural model decomposition method (see, e.g.,
[23] and [31]). The main difference between these two approaches is that in the
latter simple ARIMA models are directly specified for each unobserved component,
whereas in the former an overall ARIMA model is obtained from observable data
and, by imposing certain conditions, models for each component are derived. Since
the components are unknown, to obtain a unique decomposition Hillmer and Tiao
proposed a canonical decomposition which has the properties of maximizing the
variance of the irregulars and minimizing the variance of the other components.
ARIMA models identification and estimation are sensitive to outliers or extreme
values and cannot deal with deterministic components, such as trading days and
moving holidays. Therefore, further developments were made by combining dummy
variables regression models with ARIMA models to deal with these cases. In
this regard, Gomez and Maravall [22] developed at the Bank of Spain a seasonal
adjustment software called TRAMO-SEATS which is currently applied mainly
by European statistical agencies. TRAMO stands for “Time Series Regression
with ARIMA noise, Missing observations and Outliers” and SEATS for “Signal
Extraction in ARIMA Time Series.” First, TRAMO estimates via regression dummy
variables and direct ARIMA modeling, the deterministic components which are
after removed from the input data. In a second round, SEATS estimates the season-
ality and trend-cycle components from an ARIMA model fitted to the modified data
where the deterministic components are removed. SEATS uses the filters derived
from the linearized ARIMA model that describes the behavior of the time series. It
should be mentioned that Eurostat, with the collaboration of the Bank of Belgium,
the US Bureau of the Census, the Bank of Spain, and the European Central Bank,
is developing an interface of TRAMO-SEATS and X12ARIMA called Demetra+,
and more recently JDemetra+. At the Bank of Spain and Eurostat web site it is also
possible to find a considerable number of papers relevant to TRAMO-SEATS as
well as in the European Statistical System (ESS) Guidelines.
On the other hand, the structural model decomposition method starts directly
with an observation equation (sometimes called measurement equation) that relates
the observed time series to the unobserved components. Simple ARIMA or stochas-
tic trigonometric models are assumed for each component. The structural model
is cast in a state space form and estimated with Kalman filtering and smoothing.
Koopman et al. [32] developed STAMP, that stands for Structural Time series
Analyzer, Modeler and Predictor, and includes several types of models for each
2.5 Calendar Variations 47
The main calendar variations are moving holidays and trading days.
102
100
99
98
97
the effect is absent because Easter fell too late in April. The dates of Easter appear
in Table 2.1.
In the case illustrated, the effect is felt 1 day before Easter and on Easter Sunday
but not after Easter. This is evidenced by years 1985, 1986, 1989, 1991, 1997,
2002, 2005, and 2013, where Easters falls early in April or in March. Imports are
substantially affected by Easter, because customs do not operate from Good Friday
2.5 Calendar Variations 49
to Easter Monday. Easter can also significantly affect quarterly series, by displacing
activity from the second to the first quarter.
Generally, festival effects are difficult to estimate, because the nature and
the shape of the effect are often not well-known. Furthermore, there are few
observations, i.e., one occurrence per year.
Time series may be affected by other variations associated with the composition of
the calendar. The most important calendar variations are the trading day variations,
which are due to the fact that the activities of some days of the week are more
important than others. Trading day variations imply the existence of a daily pattern
analogous to the seasonal pattern. However, these daily factors are usually referred
to as daily coefficients.
Depending on the socioeconomic variable considered, the activity of some days
may be 60 % more important than an average day and other days, 80 % less
important.
If the most important days of the week appear five times in a month (instead
of four), the month registers an excess of activity ceteris paribus. If the least
important days appear five times, the month records a short-fall. As a result, the
monthly trading day component can cause variations of +8 % or 8 % (say) between
neighboring months and also between same months of neighboring years. The
trading day component is usually considered as negligible and very difficult to
estimate in quarterly series.
Figure 2.4 displays the monthly trading day component obtained from the
following daily pattern: 101.55, 102.26, 100.75, 99.99, 97.85, 97.73, and 99.95 for
Monday to Sunday (in percentage), respectively. The daily pattern indicates that
Tuesday, Friday, and Saturday are approximately 2.2 % more important than an
average day (100 %); and that Wednesday, Thursday, and Sunday are less important.
For the multiplicative, the log-additive and the additive time series decomposition
models, the monthly trading day component is respectively obtained in the following
manner:
X X
Dt D d =nt D .2800 C d /=nt (2.28)
2t 2t5
! " ! #
X X
Dt D exp d =nt D exp d =nt (2.29)
2t 2t5
X X
Dt D d D d (2.30)
2t 2t5
50 2 Time Series Components
102
101
100
99
98
where nt is the number of days in month t, t5 is the set of days of the week that
appears five times in month t, and d are the daily coefficients in the month. Setting
nt equal to the number of days in month t implies that the length-of-month effect
is captured by the multiplicative seasonal factors, except for Februaries. To adjust
Februaries for the length-of-month, the seasonal factors of that month are multiplied
by 29/28.25 and 28/28.25 for the leap and non-leap years, respectively.
The other option is to set nt equal to 30.4375, so that the multiplicative trading
day component also accounts for the length-of-month effect. The number 2800 in
Eq. (2.28) is the sum of the first 28 days of the months expressed in percentage.
The monthly trading day estimates of the US Imports of Goods from Canada
shown in Fig. 2.4 were obtained with the log-additive model (2.29). They display
a drop of 5.45 % between February and March 1996 and an increase of 6.52 %
between January and February 2000.
One can identify several instances where the change between same-months
is significant. Indeed, same-month year-ago comparisons are never valid in the
presence of trading day variations, not even as a rule of thumb. Furthermore, it
is apparent that the monthly trading day factors in the figure are identical for quite a
few months. Indeed for a given set of daily coefficients, there are only 22 different
monthly values for the trading day component, for a given set of daily coefficients:
seven values for 31-day months (depending on which day the month starts), seven
for 30-day months, seven for 29-day months, and one for 28-day months. In other
words, there are at most 22 possible arrangements of days in monthly data.
2.5 Calendar Variations 51
Many goods and services are affected by daily patterns of activity, which
entail higher costs for producers, namely through the need of higher inventories,
equipment, and staff on certain days of the week. For example, there is evidence that
consumers buy more gasoline on certain days of the week, namely on Thursdays,
Fridays, Saturdays, and holidays, which results in line-ups and shortages at the
pumps. In order to cope with the problem, gasoline retailers raise their price on
those days to promote sales on other days. Furthermore, the elasticity of demand for
gasoline is low. In other words, to reduce consumption by a small percentage, prices
must be raised by a disproportionate percentage, which upsets some consumers.
On the other hand, consumers can buy their gasoline on other days. The alternative
for retailers is to acquire larger inventories, larger tanks, more pumps, and larger
fleets of tanker trucks, all of which imply higher costs and translate into much
higher prices. In other words, there are savings associated with more uniform daily
patterns; and costs, with scattered daily patterns. A similar consumer behavior
prevails for the purchases of food, which probably results in more expensive
prices, namely through higher inventories, larger refrigerators, more numerous cash
registers, and more staff, than otherwise necessary. Deaths also occur more often
on certain days of the week. Car accidents, drowning, skiing, and other sporting
accidents tend to occur on weekend days and on holidays. In principle, stock series
pertaining to 1 day display a particular kind of trading day variations. Among other
things, inventories must anticipate the activity (flow) of the following day(s). For
such stock series, the monthly trading day factor coincides with the daily weight of
the day.
A classical deterministic model for trading day variations was developed by Young
[42],
y t D Dt C e t ; t D 1; : : : ; n (2.31)
7
X
Dt D ˛j Njt (2.32)
jD1
P
where et WN.0; e2 /; 7jD1 ˛j D 0; ˛j ; j D 1; : : : ; 7, denote the effects of the 7
days of the week, Monday to Sunday, and Njt is the number of times day j is present
P
in month t. Hence, the length of the month is Nt D 7jD1 Njt , and the cumulative
P
monthly effect is given by (2.32). Adding and subtracting ˛N D . 7jD1 ˛j /=7 to
Eq. (2.32) yields
7
X
Dt D ˛N
N tC .˛j ˛/N
N jt : (2.33)
jD1
52 2 Time Series Components
P
with the effect of Sunday being ˛7 D 6jD1 ˛j .
Deterministic models for trading day variations assume that the daily activity
coefficients are constant over the whole range of the series. Stochastic model for
trading day variations has been rarely proposed. Dagum et al. [20] developed a
model where the daily coefficients change over time according to a stochastic
difference equation.
Some events can cause displacements of activity from one month to the next months,
or vice versa. This phenomenon is referred to as redistribution of outliers. We also
deal with the strike effect under this headline. The outliers must be modeled and
temporally removed from the series in order to reliably estimate the systematic
components, namely the seasonal and trading day components. Events such as
2.6 The Irregular Component 53
1.15
1.10
1.05
1.00
0.95
Fig. 2.5 Irregular component of US New Orders for Durable Goods (January 1992–December
2013)
major snow storms and power blackouts usually postpone activity to the next month,
without much longer term effect.
The irregulars are most commonly assumed to follow a white noise process fIt g
defined by
Box and Tiao [6] introduced the following intervention model to deal with
different types of outliers:
1
X 1
X
yt D hj xtj C t D hj Bj xt C t D H.B/xt C t ;
jD0 jD0
where the observed series fyt g consists of an input series fxt g considered a
deterministic function of time and a stationary process ft g of zero mean and
P with fxt g. In such a case the mean of fyt g is given by the deterministic
uncorrelated
function 1 jD0 hj xtj . The type of function assumed for fxt g and weights fhj g depend
on the characteristic of the outlier or unusual event and its impact on the series.
If the outlier at time t0 is additive, in the sense that it will not permanently modify
the level of fyt g, that is, E.yt / D 0; 8t t0 , and also E.yt / ! 0 for t ! t0 , then an
appropriate xt is the impulse function defined by
1; t D t0 ;
xt D Pt .t0 / D
0; t ¤ t0 :
If, instead, the outlier at time t0 is multiplicative in the sense that it will modify the
level of fyt g, that is, E.yt / D 0; 8t t0 , and also E.yt / ! c ¤ 0 for t ! t0 , then an
appropriate xt is the step function defined by
1; t ¤ t0 ;
xt D St .t0 / D
0; t < t0 :
where the weights !0 ; : : : ; !k measure the outlier effects during the k C 1 periods
t0 ; t0 C 1; : : : ; t0 C k. The global effect is such that the sum of all the outlier effects
cancel out.
For a sudden permanent change of level at time t D t0 , a step function can be
used
!; t t0
H.B/xt D !0 St .t0 / D
0; otherwise
where ! represent the level difference of the series before and after the outlier. Since
.1 B/St .n/ D Pt .n/, we can write
! !; t D t0
H.B/xt D Pt .t0 / D
.1 B/ 0; otherwise
P P1
where !.1B/1 D !.1CBCB2 C /Pt .t0 / D 1 jD0 B Pt .t0 / D
j
jD0 !Pt .t0 Cj/:
For a sudden transitory change at time t D t0 , an appropriate model is
8
ˆ
ˆ 0; t < t0
! <
!; t D t0
H.B/xt D Pt .t0 / D
.1 ıB/ ˆ ı 2 !; t D t0 C 1
:̂ 2
ı !; t D t0 C 2
where ! denotes the initial effect, and 0 < ı < 1 is the rate of decrease of the initial
effect.
For a gradual permanent level change at time t D t0 , an appropriate model is
8
ˆ
ˆ 0; t < t0
ˆ
ˆ
ˆ
ˆ !; t D t0
ˆ
<
! .1 C ı/!; t D t0 C 1
H.B/xt D St .t0 / D 2
.1 ıB/ ˆ C ı C ı /!; t
ˆ
ˆ
.1 D t0 C 2
ˆ
ˆ
ˆ
:̂ !
.1ı/ ; t!1
56 2 Time Series Components
The intervention model proposed by Box and Tiao [6] requires the location
and the dynamic pattern of an event to be known. Chen and Liu [12] consider
the estimation problem when both the location and the dynamic pattern are not
known a priori. The proposed iterative procedure is less vulnerable to the spurious
and masking effects during outlier detection and allows to jointly estimate the
model parameters and multiple outlier effects. This procedure is the one actually
implemented in X12ARIMA and TRAMO-SEATS software.
References
20. Dagum, E.B., Quenneville, B., & Sutradhar, B. (1992). Trading-day variations multiple
regression models with random parameters. International Statistical Review, 60(1), 57–73.
21. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., & Chen, B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
22. Gomez, V., & Maravall, A. (1996). Program TRAMO and SEATS. Madrid: Bank of Spain.
23. Harvey, A. C. (1981). Time series models. Oxford, UK: Phillip Allan Publishers Limited.
24. Harvey, A. C. (1985). Trends and cycles in macroeconomic time series. Journal of the Business
and Economic Statistics, 3, 216–227.
25. Harvey, A. C., & Trimbur, T. M. (2003). General model-based filters for extracting cycles and
trends in economic time series. The Review of Economics and Statistics, 85(2), 244–255.
26. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
27. Hillmer, S. C., & Tiao, G. C. (1982). An ARIMA-model-based approach to seasonal adjust-
ment. Journal of the American Statistical Association, 77, 63–70.
28. Hodrick, R. J., & Prescott, E. C. (1997). Postwar U.S. business cycles: An empirical
investigation. Journal of Money, Credit and Banking, 29(1), 1–16.
29. Kaiser, R., & Maravall, A. (2002). A complete model-based interpretation of the Hodrick-
Prescott filter: Spuriousness reconsidered. Banco de Espana Working Papers 0208, Banco de
Espana.
30. Kaiser, R., & Maravall, A. (2005). Combining filter design with model-based filtering (with
an application to business-cycle estimation). International Journal of Forecasting, 21(4), 691–
710.
31. Kitagawa, G., & Gersch, W. (1984). A smoothness priors state-space modelling of time series
with trend and seasonality. Journal of the American Statistical Association, 79, 378–389.
32. Koopman, S. J., Harvey, A. C., Doornik, J. A., & Shephard, N. (1998). Structural time series
analysis STAMP (5.0). London: Thomson Business Press.
33. Macaulay, F. R. (1931). The smoothing of time series. New York: National Bureau of Economic
Research.
34. Malthus, T. R. (1798). An essay on the principle of population. Oxford World’s Classics reprint:
xxix Chronology.
35. Maravall, A. (1993). Stochastic and linear trend models and estimators. Journal of Economet-
rics, 56, 5–37.
36. Moore, G. H. (1961). Business cycle indicators. Princeton, NJ: Princeton University Press.
37. Persons, W. M. (1919). Indices of business conditions. Review of Economic Statistics, 1, 5–107.
38. Schumpeter, J. A. (1954). History of economic analysis. New York: Oxford University Press
Inc.
39. Shiskin, J., Young, A. H., & Musgrave, J. C. (1967). The X-11 variant of the Census Method
II seasonal adjustment program. Technical Paper 15 (revised). US Department of Commerce,
Bureau of the Census, Washington, DC.
40. Stock, J. H., & Watson, M. W. (2003). Has the business cycle changed and why?. NBER
macroeconomics annual 2002 (Vol. 17, pp. 159–230). Cambridge: MIT Press.
41. Wold, H. O. (1938). A study in the analysis of stationary time series. Uppsala, Sweden:
Almquist and Wiksell (Second Edition 1954).
42. Young, A. H. (1965). Estimating trading-day variations in monthly economic time series. US
Bureau of the Census, Technical Report No. 12.
43. Young, A. H. (1968). A linear approximation to the Census and BLS seasonal adjustment
methods. Journal of the American Statistical Association, 63, 445–457.
Part I
Seasonal Adjustment Methods
Chapter 3
Seasonal Adjustment: Meaning, Purpose, and
Methods
Abstract This chapter deals with the causes and characteristics of seasonality.
For gradual changes in seasonality whether of a stochastic or a deterministic type,
various models have been proposed that can be grouped into two broad categories:
(1) models that assume that the generating process of seasonality varies only in
amplitude and (2) models that assume that the generating process varies in both
amplitude and phase. The basic assumptions of both groups of models are studied
with particular reference to the two seasonal adjustment methods officially adopted
by statistical agencies, the X12ARIMA and TRAMO-SEATS. The economic
significance of seasonality and the need for seasonal adjustment are also discussed.
Since seasonality ultimately results mainly from noneconomic forces (climatic and
institutional factors), external to the economic system, its impact on the economy as
a whole cannot be modified in a short period of time. Therefore, it is of interest for
policy making and decision taking to have the seasonal variations removed from the
original series. The main reason for seasonal adjustment is the need of standardizing
socioeconomic series because seasonality affects them with different timing and
intensity.
A great deal of data in business, economics, and natural sciences occurs in the
form of time series where observations are dependent and where the nature of
this dependence is of interest in itself. The time series is generally compiled for
consecutive and equal periods, such as weeks, months, quarters, and years. From a
statistical point of view, a time series is a sample realization of a stochastic process,
i.e., a process controlled by probability laws. The observations made as the process
continues indicate the way it evolves.
In the analysis of economic time series, Persons [44] was the first to distinguish
four types of evolution, namely: (a) the trend, (b) the cycle, (c) the seasonal
variations, and (d) the irregular fluctuations.
Among all these components, the influence of the seasonal fluctuations in the
human activity has been felt from earlier times. The organization of society, the
means of production and communication, the habits of consumption, and other
social and religious events have been strongly conditioned by both climatic and
conventional seasons. The seasonal variations in agriculture, the high pre-Easter and
pre-Christmas retail sales, and the low level of winter construction are all general
knowledge. The main causes of seasonality, the climatic and institutional factors
are mainly exogenous to the economic system and cannot be controlled or modified
by the decision makers in the short run. The impact of seasonality in the economic
activity, however, is usually not independent on the stages of the business cycle.
It is well-known, for example, that the seasonal unemployment among adult males
is much larger during periods of recession than during periods of prosperity.
Another main feature of seasonality is that the phenomenon repeats with certain
regularity every year but it may evolve. The latter is mainly due to systematic
intervention of the government and to technological and institutional changes as
well; and it is more the general case than the exception for economic time series.
Therefore, the assumption of stable seasonality, i.e., of seasonality being repre-
sentable by a strictly periodic function, can be used as a good approximation for few
series only. In effect, even in the extreme cases of those activities where seasonality
is mainly caused by climatic conditions, e.g., agriculture, fishing, forestry, the
seasonal variations change, for weather itself measured by the temperature, and
quantity of precipitation changes. There are many other reasons that can produce
temporal shifts in the seasonal effects. A decline in the participation of the primary
sector in the Gross National Product will modify seasonal patterns in the economy
as a whole, as will a change in the geographical distribution of industry in a country
extending over several climatic zones. Changes in technology alter the importance
of climatic factors. Customs and habits change with a different distribution of
income and, thus, modify the demand for certain goods and its corresponding
seasonal pattern.
Once the assumption of stable seasonality is abandoned, new assumptions must
be made regarding the nature of its evolution; if seasonality changes, is it slowly or
rapidly? Is it gradually or abruptly? Is it in a deterministic or a stochastic manner?
Today, the most widely accepted hypothesis is that seasonality moves gradually,
slowly, and in a stochastic manner. For gradual changes in seasonality whether of a
stochastic or a deterministic type, various models have been studied. These models
can be grouped into two broad categories:
1. models that assume the generating process of seasonality varies in amplitude
only and
2. models that assume the generating process varies in both amplitude and phase.
Kuznets [38] and Wald [49] were the first to propose an estimation procedure that
assumed seasonality changes in amplitude while the pattern (i.e., the proportionality
relationship between the seasonal effects of each month) remained constant over
time.
On the other hand, the majority of the methods officially adopted by statistical
agencies belongs to the second group. Seasonality is assumed to change either
in a stochastic or in a deterministic manner. The stochastic approach has been
analyzed among many others by Hannan [23], Hannan et al. [24], Brewer et al. [5],
3.2 The Economic Significance of Seasonality and the Need for Seasonally. . . 63
Cleveland [9], and Cleveland and Tiao [10]. Pierce [45] proposed a mixed model
where seasonality can be in part deterministic and in part stochastic.
A third characteristic of seasonality is that the phenomenon can be distinguished
from other movements (trend, cycle, and irregulars) that influence an economic time
series and can be measured. The seasonal variations are distinguished from trend
by their oscillating character, from the cycle by being confined within the limits
of an annual period, and from the irregulars, by the fact of being systematic. For
example, the causes of seasonality in labor force series are found in the main factors
that influence their demand and supply. The seasonal variations in employment and
unemployment of adult males are strongly conditioned by the stages of the business
cycle and the weather. There is practically no seasonality in the supply of adult
males labor but there is a large amount of seasonal variation in the demand side.
This group is mainly absorbed by the primary sector (construction, agriculture,
mining, fishing, and forestry) where seasonality is mostly climatic with very large
oscillation, and by the industrial sector, where seasonality is mostly induced by the
seasonality in the primary sector but it is also strongly affected by the stages of the
cycle. On the other hand, seasonality for females and young males originates from
the side of the demand as well as of the supply. In fact, females and young males
are mainly employed by the tertiary sector where seasonality tends to be originated
by institutional and religious events (Christmas, Easter, Federal Taxes deadlines).
Similarly, from the viewpoint of the supply, seasonal variations occur because this
group tends to move in and out of the labor force in accordance with the school year.
The seasonal variations affect the economic sectors in different degrees. Nor is it
only the intensity of the primary effects of the seasons that differs for the various
economic activities, producing seasonal swings of varying size or amplitude, the
timing of these effects is also different, creating different patterns as the season
begins in 1 month or another. Thus, for example, in winter occur both the seasonal
low in construction and the seasonal peak in retail trade. The seasonal amplitudes
for both sectors also differ significantly, being much larger for construction than for
retail trade.
Similar to the business cycle, the presence of seasonality in the economy
introduces a disequilibrium in the allocation of the resources, implying an extra cost.
Because of the seasonal variations, the economy has a supply of labor, equipment,
and raw materials in excess to what it would be required if the activity proceeds
at an even rate throughout the year. Since these surpluses may arise also from
the different impacts of the other components (trend, cycle, and irregulars) on the
economic processes, it becomes very difficult to estimate the actual cost imposed by
seasonality. Some attempts of measurement have been made by Kuznets [38], Baron
64 3 Seasonal Adjustment: Meaning, Purpose, and Methods
[3], Dawson et al. [16], and Judek [33] among many others. For instance, Kuznets
[38] suggested two measures of the degree of overcapacity in industry attributable
to seasonal variations: one gives the total difference in activity between each of the
months and the months of peak output; the other, the amount by which the months
below 100 depart from 100. The basic idea is that the seasonal peak months measure
the normal plus the latent capacity of the industry, therefore, an output below the
seasonal peak indicates an inadequate utilization of the maximum possible capacity.
On the other hand, an output below a seasonal index of 100 indicates an inadequate
utilization of the normal capacity of the industry. The same idea is applicable to the
surplus of labor. If the seasonal peak in employment indicates the number of workers
attached to a given industry, then the average monthly departure from the peak
measures the average monthly seasonal unemployment and hence the surplus of
labor that is in the industry because of seasonal swings. If, however, only the annual
average is considered the correct estimate of labor attached to the industry, then only
departures below 100 should be considered as measuring seasonal unemployment
and excess of labor supply.
Because seasonality ultimately results mainly from noneconomic forces (climatic
and institutional factors), external to the economic system, its impact on the
economy as a whole cannot be modified in a short period of time. Therefore, it is of
interest to decision makers to have the seasonal variations removed from the original
series to obtain a seasonally adjusted series. Seasonal adjustment means the removal
of seasonal variations in the original series jointly with trading day variations and
moving holiday effects. The main reason for seasonal adjustment is the need of
standardizing socioeconomic series because seasonality affects them with different
timing and intensity. In this manner, the variations of a seasonally adjusted series are
due to variations only in the trend, the cycle, and the irregulars. It should be noticed,
however, that the removal of seasonal variations from a time series does not indicate
how the series would have evolved as there had been no seasonal variations, rather
it shows more clearly the trend-cycle abstracting from seasonality.
The information given by seasonally adjusted series plays a very important role
in the analysis of current economic conditions, particularly, in determining the stage
of the cycle at which the economy stands. Such knowledge is useful in forecasting
subsequent cyclical movements and provides the basis for decision making to
control the level of the economic activities. It is particularly important around
turning points since, e.g., failure to recognize the downturn in the business cycle
may lead to the adoption of policies to curb expansion when, in fact, a recession is
under way. Analysts who wish to get a picture of the economic situation not distorted
by exogenous variables related to the seasonal variations, may make comparisons
with the same month of the year before. However, apart from other considerations,
the basic problem with same-month year-ago comparisons is that they show only
what has happened a year after, not what was happening during the year not what is
happening currently. To evaluate the current state of the economy, the analyst should
be able to measure cyclical changes for each month over less than 1 year span, e.g.,
to compare May with April (1-month span) or May with February (3-month span).
3.3 Basic Assumptions of Main Seasonal Adjustment Methods 65
Decision making based on the raw data can lead to wrong policies, particularly
if the series is strongly affected by seasonal variations. The average absolute
month percentage change in the seasonal variation can be much greater than the
corresponding changes in the irregular or trend-cycle. Results of several studies of
selected economic US indicators show that the average absolute month-to-month
percentage changes in the seasonal component run between 3 and 7 times the
average absolute percentage changes in the trend-cycle or in the irregulars over the
same time spans (see Shiskin [46]). The measurement of seasonality is also very
useful for short-term decision making. The knowledge of the seasonal pattern of the
economic activities facilitates a better planning of the economic resources during
the periods of peak loads and of inactivity. This knowledge can also be applied for a
better integration of economic activities characterized by opposite seasonal patterns.
Resources which formerly served only one purpose can serve additional purposes
and thus reduce the burden imposed by seasonality to the whole economy.
The majority of the seasonal adjustment methods developed until now is based
on univariate time series models, where the estimation of seasonality is made in
a simple and mechanical manner and not based on a causal explanation of the
phenomenon under study. Very few attempts have been made to follow the latter
approach and none of them reached further than the experimental stage. Among
others, Mendershausen [43] tried to regress the seasonal effects for each month on a
set of exogenous variables (meteorological and social variables) in order to build an
explanatory model for seasonality but his empirical results were not very fruitful.
On the other hand, univariate time series methods of seasonal adjustment try to
estimate the generating mechanism of the observations under the simple assumption
that the series is composed of a systematic part which is a well-determined function
of time and, a random part which obeys a probability law. The feasibility of
this decomposition was proved in a well-known theorem due to Herman Wold
[52]. Wold showed that any second order stationary process fYt g can be uniquely
represented as the sum of two mutually uncorrelated processes
Yt D Zt C Vt ;
where fZt g is a convergent infinite moving average process, and fVt g a deterministic
process. The decomposition is linear and is determined entirely by the second
moments of the process.
In a broad sense, seasonal methods can be classified as deterministic or stochastic
depending on the assumptions made concerning how seasonality evolves through
time. Deterministic methods assume that the seasonal pattern can be predicted with
66 3 Seasonal Adjustment: Meaning, Purpose, and Methods
no error or with variance of the prediction error null. On the contrary, stochastic
methods assume that seasonality can be represented by a stochastic process, a
process governed by a probability law and, consequently, the variance of the
prediction error is not null. The best known seasonal adjustment methods belong
to the following types:
(a) regression methods which assume global or local simple functions of time,
(b) stochastic-model-based methods which assume simple autoregressive inte-
grated moving average (ARIMA) models, and
(c) moving average methods which are based on linear filtering and hence do not
have explicit parametric models.
Only methods (a), which assume global simple functions for each component,
are deterministic; the others are considered stochastic. Moving average or linear
filtering methods are those adopted mostly by statistical agencies to produce
officially seasonally adjusted series.
The use of regression methods for the decomposition of a time series is old. In
the late 1930s, Fisher [20] and Mendershausen [43] proposed to fit polynomials
by the least squares method to estimate the seasonal effects. To deal with moving
seasonality, Cowden [11] suggested to fit polynomials to each month seasonal
obtained from the ratios (differences) of the original series and a centered twelve
term moving average. Jones [30] and Wald [49] considered the problem of fitting
simultaneously a polynomial of a relatively low degree to estimate the trend plus
twelve constants (one for each month) representing stable seasonality.
In the 1960s, the use of multiple regression techniques for the seasonal adjust-
ment of economic time series was strongly advocated. The main reasons for this
were the widespread interest in the use of monthly and quarterly series for econo-
metric model building and the fast development of electronic computers. In effect,
the former posed the problem of whether to use as input seasonally adjusted data
directly, or to use first, raw data, and then seasonally adjusted. These two alternatives
do not give the same results unless the seasonal adjustment method is strictly
linear (conditioned fulfilled by the regression methods). The electronic computers
reduced significantly the heavy burden imposed by the computations in the least
squares method despite the existence of several simplifying procedures such as
the orthogonal polynomials and the Buys-Ballot table. Important contributions
were made by Hannan [22], Lovell [40], Ladd [39], Jorgenson [31], Henshaw
[28], Stephenson and Farr [48], and Wallis [50]. Their contributions facilitated the
application of least squares methods for the seasonal adjustment of economic time
series.
3.3 Basic Assumptions of Main Seasonal Adjustment Methods 67
The regression methods assume that the systematic part of a time series can be
approximated closely by simple functions of time over the entire span of the series.
In general, two types of functions of time are considered. One is a polynomial
of fairly low degree that fulfills the assumption that the economic phenomenon
moves slowly, smoothly, and progressively through time (the trend). The other type
of function is a linear combination of sines and cosines of different frequencies
representing oscillations, strictly periodic or not, that affect also the total variation
of the series (the cycle and seasonality).
For certain regression functions such as fixed periodic functions, polynomial
time trends, and interaction of the two types, ordinary least squares estimates
are asymptotically efficient if the random part is second order stationary, i.e.,
the mean and variance are constant and the covariance is function only of the
time lag. Therefore, by taking differences to ensure a stationary process, ordinary
least squares are generally adequate. If the relationship among the components is
multiplicative, the standard procedure is to take logarithms and then, differences, in
order to transform the generating mechanism of the series into an additive form with
a stationary random term.
A large number of model specifications have been considered for monthly series.
Young [53] analyzed and compared the seasonal adjustment obtained by regression
methods with the results given by the Census II-X11 variant which belongs to
the category of linear smoothing techniques. The models analyzed have different
degrees of flexibility concerning the pattern of the seasonals and trend-cycle and the
relationship between the components is assumed to be multiplicative.
Model I
2
X
C ˇ1;k tk sin k t C ˇ2;k tk cos k t ; (3.2)
kD1
2k
where k D 12 ,
5
X
log Ct D 0 C k tk C ı1 sin !t C ı2 cos !t; (3.3)
kD1
where ! D 2 72 .
This model contains a deterministic stable seasonality represented by the sines
and cosines of periodicity 12, 6, 4, 3, 2.4, and 2 plus a deterministic moving
seasonality that is assumed to follow a second degree polynomial through time.
The trend-cycle expression contains a fifth degree polynomial and sines and cosines
with a periodicity of 72 months that take into account a 5 year cycle.
68 3 Seasonal Adjustment: Meaning, Purpose, and Methods
Model II
Instead of using sines and cosines, this model contains a deterministic stable
seasonality represented by 12 dummy variables dk where dk D 1 if t is the k-th
month of the year and equals 0 otherwise. The trend-cycle curve is represented by a
10-th degree polynomial.
Model III
12
X 12
X
yt D ˛1;k dk C ˛2;k Œdk C
C It (3.10)
kD1 kD1
3.3 Basic Assumptions of Main Seasonal Adjustment Methods 69
subject to
12
X
˛2;k D 0
kD1
where C is a 12 month lagging average of the series (a trend level estimate) and the
dk ’s are monthly dummies. This model assumes a seasonal pattern that is stable plus
a slowly moving trend conditioned to seasonality for the whole span of the series.
The estimate of the trend C is the same for all the series. The seasonally adjusted
data are calculated by adding the mean of the original series to the residuals It .
This model which is extremely simple as compared to other models specifications
for regression methods did not give better results than those obtained by moving
averages procedures. As suggested by Early [17], the revisions in the seasonally
adjusted values were high and not necessarily monotonic for earlier years.
To overcome the limitation of using global deterministic representations for
the trend, cycle, and seasonality, regression methods were extended to incorporate
stochastic representations by means of local polynomials (spline functions) for suc-
cessive short segments of the series and introducing changing seasonal amplitudes.
A major breakthrough in this direction was made by Akaike [1] who introduced
prior constraints to the degree of smoothness of the various components and
solved the problem with the introduction of a Bayesian model. Another important
contribution is the regression method with locally deterministic models estimated
by the LOESS (locally weighted regression) smoother developed by Cleveland
et al. [8].
The simplest and often studied seasonal adjustment regression method assumes
that the generating process of seasonality can be represented by strictly periodic
functions of time of annual periodicity. The problem is to estimate s seasonal
coefficients (s being the seasonal periodicity, e.g., 12 for monthly series, 4 for
quarterly series) subject to the seasonal constraint that they sum to zero. This
regression seasonal model can be written as
yt D S t C I t ; t D 1; : : : ; n
X
s X
s
St D j djt subject to j D 0; (3.11)
jD1 jD1
where the dj ’s are dummy variables taking a value of unity in season j and a value of
zero otherwise; fIt g WN.0; e2 / is a white noise process with zero mean, constant
variance, and non-autocorrelated. The seasonal parameters j ’s can be interpreted
as the expectation of Yt in each season, and represent the seasonal effects of the
series under question. Since the parameters are constant, the seasonal pattern repeats
exactly over the years, that is,
St D Sts : (3.12)
70 3 Seasonal Adjustment: Meaning, Purpose, and Methods
Equation (3.11), often included in econometric models, assumes that the effects
of the seasonal variations are deterministic because they can be predicted with
no error. However, for most economic and social time series, seasonality is not
deterministic but changes gradually in a stochastic manner. One representation of
stochastic seasonality is to assume in Eq. (3.11) when the j ’s are random variables
instead of constant. Hence, the relationship given in Eq. (3.12) becomes
where f!t g WN.0; !2 / and E.!t It / D 0; 8t. The stochastic seasonal balance
P
constraint is here given by s1jD0 Stj D !t , with expected value equal to zero.
Equation (3.13) assumes seasonality to be generated by a nonstationary stochas-
tic process. In fact, it is a random walk process and can be made stationary by
applying seasonal differences .1Bs /, where B denotes the backward shift operator.
In reduced form, the corresponding regression model is a linear combination of
white noise processes of maximum lag s. Since .1Bs/ D .1B/.1CBC CBs1 /,
model-based seasonal adjustment methods attribute only .1 C B C C Bs1 / D
P s1 j
jD0 B to the seasonal component, leaving .1 B/ as part of a nonstationary trend.
Thus, the corresponding seasonal stochastic model can be written as
Equation (3.14) represents a seasonality that evolves in an erratic manner and it was
included in seasonal adjustment methods based on structural model decomposition
(see, e.g., Harvey [25] and Kitagawa and Gersch [36]). This seasonal behavior is
seldom found in observed time series where seasonality evolves smoothly through
time. To represent this latter, Hillmer and Tiao [29] introduced a major modification
where the left-hand side of Eq. (3.15) follows an invertible moving average process
of maximum order s 1, denoted by MA.s 1/. Equation (3.15) is discussed
extensively in Bell and Hillmer [4], and can be generalized easily by replacing
the MA part with stationary invertible autoregressive moving average (ARMA)
processes of the Box and Jenkins type. Depending on the order of the ARMA
process, a large variety of evolving stochastic seasonal variations can be modeled.
Other seasonal models used in regression methods, where the components are
assumed to be local polynomials of time, are based on a smoothness criterion. For
example, given Yt D SP
P t C It ; t D 1; : : : ; n, restrictions are imposed on St , such that
2
n
tD1 .S t S ts / and t Œ.1 C B C C B
s1
/St
2 have to be small. The solution is
given by minimizing a weighted linear combination of both criteria as follows:
X
n X
n X
n
.yt St /2 C ˛ .St Sts /2 C ˇ Œ.1 C B C C Bs1 /St
2 (3.16)
tD1 tD1 tD1
3.3 Basic Assumptions of Main Seasonal Adjustment Methods 71
Stochastic model-based methods were mainly developed during the 1980s following
two different approaches. One, originally known as seasonal adjustment by SIGEX
(Signal Extraction), was developed by Burman [6] and as ARIMA model-based
seasonal adjustment by Hillmer and Tiao [29], the other is referred to as structural
model decomposition method (see, e.g., Harvey [25] and Kitagawa and Gersch
[36]). The main difference between these two approaches is that in the latter simple
ARIMA models are directly specified for each unobserved component, whereas
in the former an overall ARIMA model is obtained from observable data and, by
imposing certain restrictions, models for each component are derived. Since the
components are unknown, to obtain a unique decomposition Hillmer and Tiao [29]
proposed a canonical decomposition which has the properties of maximizing the
variance of the irregulars and minimizing the variance of the stochastic seasonal
models.
ARIMA models identification and estimation are sensitive to outliers or extreme
values and cannot deal with deterministic components, such as trading days and
moving holidays. Therefore, further developments were made by combining dummy
variables regression models with ARIMA models to deal with these cases. In
this regard, Gomez and Maravall [21] developed at the Bank of Spain a seasonal
adjustment software called TRAMO-SEATS which is currently applied mainly
by European statistical agencies. TRAMO stands for “Time Series Regression
with ARIMA noise, Missing observations and Outliers” and SEATS for “Signal
Extraction in ARIMA Time Series.” Firstly, TRAMO estimates via regression
dummy variables and direct ARIMA modeling, the deterministic components which
are after removed from the input data. In a second round, SEATS estimates the
seasonal and trend-cycle components from an ARIMA model fitted to the data
where the deterministic components are removed. SEATS uses the filters derived
from the ARIMA model that describes the behavior of the linearized time series.
72 3 Seasonal Adjustment: Meaning, Purpose, and Methods
It should be mentioned that Eurostat, with the collaboration of the Bank of Belgium,
the US Bureau of the Census, the Bank of Spain, and the European Central Bank,
has developed an interface of TRAMO-SEATS and X12ARIMA called Demetra+,
and more recently an update version called JDemetra+ . At the Bank of Spain
and Eurostat web site it is also possible to find a considerable number of papers
relevant to TRAMO-SEATS as well as in the European Statistical System (ESS)
Guidelines.
The structural model decomposition method starts directly with an observation
equation, also called measurement equation, that relates the observed time series
to the unobserved components. Simple ARIMA or stochastic trigonometric models
are assumed for each component. The structural model is cast in a state space form
and estimated with Kalman filtering and smoothing. Koopman et al. [37] developed
STAMP, that stands for Structural Time series Analyzer, Modeler and Predictor,
and includes several types of models for each component. STAMP is set up in
an easy-to-use form which enables the user to concentrate on model selection and
interpretation. STAMP 8 is an integrated part of the OxMetrics modular software
system for time series data analysis, forecasting, financial econometric modeling,
and the statistical analysis of cross-section and panel data. STAMP is not used
officially by statistical agencies, but mainly by econometricians and academics.
Linear smoothing filters or moving averages for seasonal adjustment were already
known in the early 1920s but seldom used [18, 35]. One of the main reasons for
this was the fact that the best known seasonal annihilator, the centered 12 months
moving average, was found to be a poor trend-cycle estimator. It cannot follow
closely peaks and troughs of short-term business cycles (of 5 years of periodicity or
less). Moreover, unless the erratic fluctuations are small, it has not enough terms to
smooth the data successfully. Another reason was that moving averages are sensitive
to outliers and, therefore, an “a priori” treatment of extreme values is required.
This sort of limitation lead King [35] and other researchers to use moving medians
instead of moving arithmetic means for the estimation of seasonality.
However, by the end of the 1920s, the development of new smoothing filters and
different techniques of application gave a big push to the use of this procedure for
the seasonal adjustment of economic time series. Joy and Thomas [32] described a
method based on moving averages and applied to one hundred series of the Federal
Reserve Board. According to the authors, the method was able to handle “long-time
changes in seasonal fluctuations” which appear to have occurred in a number of
important industries .
By the same time, Frederick Macaulay [42] developed another method also based
on linear smoothing filters. His book, The Smoothing of Time Series, published
in 1931, became a classic on the subject. Gradually, government agencies and
statistical bureaus started to apply smoothing procedures for the seasonal adjustment
3.3 Basic Assumptions of Main Seasonal Adjustment Methods 73
of their series. The procedure, however, was costly, time consuming, and mainly
subjective, because the adjustments were mostly made by hand.
The basic properties of linear smoothing filters or moving averages were
extensively discussed in the following three classical books: (1) Whittaker and
Robinson [51], (2) Macauley [42], and (3) Kendall and Stuart [34].
Only a summary of the basic principles of smoothing techniques will be
presented here, and given the complexity of the subject, we shall make it as simple
as possible.
The majority of the seasonal adjustment methods officially adopted by statistical
bureaus makes the assumption that although the signal of a time series is a smooth
function of time, it cannot be approximated well by simple mathematical functions
over the entire range. Therefore, they use the statistical technique of smoothing.
A smooth curve is one which does not change its slope in a sudden or erratic manner.
The most commonly used measure of smoothness is based on the smallness of the
sum of squares of the third differences of successive points on the curve [42].
Given a time series, yt ; t D 1; : : : ; n, for any t far removed from both ends, say
m < t < n m, the seasonally adjusted value yat is obtained by application of a
symmetric moving average W.B/
X
m
yat D W.B/yt D wj ytj ; (3.17)
jDm
where the weights wj are symmetric, that is, wm;j D wm;j , and the length of the
average is 2m C 1.
For current and recent data (n m < t n), a symmetric linear filter cannot be
applied, and therefore truncated asymmetric filters are used. For example, for the
last available observation yn , the seasonally adjusted value is given by
X
m
yan D W0 .B/yn D w0;j ynj : (3.18)
jD0
The asymmetric filters are time-varying in the sense that different filters are
applied for the m first and last observations. The end estimates are revised as new
observations are added because of the new innovations and the differences between
the symmetric and asymmetric filters. The estimates obtained with symmetric filters
are often called final.
The development of electronic computers contributed to major improvements
in seasonal adjustment based on moving averages and facilitated their massive
application. In 1954, Julius Shiskin [47] of the US Bureau of Census developed
a software called Method I, based mainly on the works of Macauley [42], already
being used by the US Federal Reserve Board. Census Method I was followed by
Census Method II and eleven more experimental versions (X1 , X2,: : :, X11). The
best known and widely applied was the Census II - X11 variant developed by Shiskin
74 3 Seasonal Adjustment: Meaning, Purpose, and Methods
et al. [47], but produced poor seasonally adjusted data at the end of the series
which is of crucial importance to assess the direction of the short-term trend and the
identification of turning points in the economy. Estela Bee Dagum [12] developed
the X11ARIMA to correct for this serious limitation.
The X11ARIMA method consists of:
(i) modeling the original series with an ARIMA model of the Box and Jenkins
type,
(ii) extending the original series 1–3 years with forecasts from the ARIMA model
that fits and extrapolate well according to well-defined acceptance criteria, and
(iii) estimating each component with moving averages that are symmetric for
middle observations and asymmetric for both end years. The latter are obtained
via de convolution of Census II - X11 variant and the ARIMA model
extrapolations.
For flow series, deterministic components such as trading day variations and
moving holiday effects are estimated with dummy variable regression models, and
removed from the series, so that only the remainder is subject to steps (i)–(iii)
above. The X11ARIMA software was developed by Dagum in 1980 [13] and was
extended again in 1988 [14]. In 1998, Findley et al. [19] developed X12ARIMA
that offers a regARIMA option to estimate deterministic components such as trading
day variations and moving holiday effects, simultaneously with the ARIMA model
for extrapolation. It also includes new diagnostic tests and spectral techniques to
assess the goodness of the results. The X12ARIMA method is today one of the
most often applied by statistical agencies in the world. The US Bureau of Census
continued research on the development of seasonal adjustment methods and recently
produced a beta version called X13ARIMA-SEATS which enables the estimation of
the seasonal component either via linear filters as those available in X12ARIMA or
based on an ARIMA model decomposition.
All the moving average methods here discussed are nonlinear. Hence, the
seasonally adjusted total of aggregated series is not equal to the algebraic sum of
the seasonally adjusted series that enters into the aggregation. The main causes of
nonlinearity are:
(i) a multiplicative decomposition model for the unobserved components,
(ii) the identification and replacement of extreme values,
(iii) the ARIMA extrapolations, and
(iv) automatic selection of moving average length for the estimation of the trend-
cycle and seasonality.
The properties of the combined linear filters applied to estimate the various
components were originally calculated by Young [53] for a standard option. Later,
Dagum et al. [15] calculated and analyzed all possible filter combinations of Census
II - X11 and X11ARIMA. Cleveland and Tiao [10] and Burridge and Wallis [7]
found ARIMA models that approximated well some of the linear filters used for the
trend-cycle and seasonal component of Census II - X11.
3.3 Basic Assumptions of Main Seasonal Adjustment Methods 75
The set of weights of a smoothing linear filter can be obtained following two
different techniques: (a) by fitting a polynomial and (b) by summation formulae.
The process of fitting a polynomial to generate the set of weights of a moving
average requires that the span or length of the average as well as the degree of the
polynomial be chosen in advance. For a given span, say 2m C 1 and a polynomial
of degree p not greater than 2m, the coefficients are estimated by the method of
least squares, and then the polynomial is used to determine the smoothed value in
the middle of the span. The weights are function only of the span of the average
2m C 1 and the degree of the polynomial, p, and do not depend on the values of
the observations. The smoothed value is a weighted average of the observations, the
weights being independent on which part of the series is taken. For a given p, the
variance of the smoothed series decreases with increasing 2m C 1 and, for a given
2m C 1, the variance goes up with increasing p (see, e.g., [2]).
The set of weights for linear smoothing filters based on summation formulae
was mainly developed by actuaries. The basic principle for the summation formulae
is the combination of operations of differencing and summation in such a manner
that when differences above a certain order is neglected, they will reproduce the
functions operated on. The merit of this procedure is that the smoothed values
thus obtained are functions of a large number of observed values whose errors to
a considerable extent cancel out.
The smoothing linear filters developed by actuaries using summation formulae,
have the good properties that when fitted to second or third degree parabolas, will
fall exactly on those parabolas. If fitted to stocastic, non mathematical data, it will
give smoother results than can be obtained from the weights which give the middle
point of a second degree parabola fitted by least squares method. Recognition of the
fact that smoothness of the resulting filtering depends directly on the smoothness of
the weight diagram, led Robert Henderson [26] to develop a formula which makes
the sum of squares of the third differences of the smoothed series a minimum for
any number of terms. To fulfill this requirement is equivalent to minimize the sum
of the squares of the third differences of the set of weights of the filter.
There is a relationship between smoothing by the mid ordinate of third degree
parabolas fitted by least squares and smoothing by the mid ordinate of third degree
parabolas from summation formulae. In the least squares, the assumption made
is that all deviations between observed and fitted values are equally weighted,
and thus the sum of squares of the deviations is made a minimum. In the
summation formulae, the deviations between observed and fitted values are not
equally weighted and if different weights are applied, the sum of squares of the
deviations is made a minimum.
The Henderson’s trend-cycle moving averages give the same results as if
weighted least squares had been applied where the weights assigned to the devi-
ations are those that give the smoothest possible diagram. The latter in the sense
that the sum of squares of the third differences is made a minimum. In smoothing
real data there is always a compromise between: (1) how good the fit should be
(in the sense of minimizing the sum of squares of the deviations between observed
and fitted values); and (2) how smooth the fitted curve should be. The lack of fit is
76 3 Seasonal Adjustment: Meaning, Purpose, and Methods
measured by the sum of squares of the deviations between observed and smoothed
values and the lack of smoothness by the sum of squares of the third differences of
the smoothed curve.
This problem was suggested by Henderson and Sheppard [27], and first analyzed
by Whittaker and Robinson [51], who thus put the theory of smoothing within the
framework of the theory of probability. The degree of fitness was weighted by a
constant k to be determined and the problem was to minimize k times the sum of
squares of the deviations between the observed and fitted curve plus the sum of
squares of the third differences of the fitted (smoothed) curve. That is to minimize:
X
n X
n
k .yt yO t /2 C ..1 B/3 yO t /2 (3.19)
tD1 tD3
The smoothed curve obtained with the Whittaker procedure is such that each of
its values is equal to the corresponding unadjusted values plus 1=k times a sixth
difference of the smoothed curve. The weights were calculated for the whole range
of the data and not in a moving manner. The set of weights eliminates stable seasonal
fluctuations only if k is very small. For k D 0:0009, more than 95 % of the seasonal
effects corresponding to a 12 month sine curve are eliminated, but cycles of less
than 36 month periodicity are very poorly fitted as shown by Macauley [42]. This
problem can be expressed in a more adequate way by minimizing a loss function
given by a convex combination between the lack of smoothness and the lack of
fit. The principle should be applied in a moving manner in order to make it more
flexible. Hence, we want to minimize
X
n X
n
k .yt yO t /2 C .1 k/ ..1 B/3 yt /2 ; 0 k 1: (3.20)
tD1 tD3
References
5. Brewer, K. R.W., Hagan, P. J., & Perazzelli, P. (1975). Seasonal adjustment using Box-Jenkins
models. In Proceedings of International Institute of Statistics, Warsaw. 40th Session (pp. 133–
139). Contributed Papers.
6. Burman, J. P. (1980). Seasonal adjustment by signal extraction. Journal of the Royal Statistical
Society Series A, 143, 321–337.
7. Burridge, P., & Wallis, K. F. (1984). Unobserved components models for seasonal adjustment
filters. Journal of Business and Economic Statistics, 2(4), 350–359.
8. Cleveland, R. B., Cleveland, W. S., McRae, J. E., & Terpenning, I. (1990). STL: A seasonal
trend decomposition procedure based on LOESS. Journal of Official Statistics, 6(1), 3–33.
9. Cleveland, W. P. (1972). Analysis and Forecasting of Seasonal Time Series (215 pp.).
Unpublished doctoral thesis, Department of Statistics, University of Wisconsin.
10. Cleveland, W. P., & Tiao, G. C. (1976). Decomposition of seasonal time series: A model for
the census X-11 program. Journal of the American Statistical Association, 71, 581–587.
11. Cowden, D. J. (1942). Moving seasonal indexes. Journal of the American Statistical Associa-
tion, 37, 523–524.
12. Dagum, E. B. (1978). Modelling, forecasting and seasonally adjusting economic time series
with the X11ARIMA method. The Statistician, 27(3), 203–216.
13. Dagum, E. B. (1980). The X11ARIMA seasonal adjustment method. Ottawa, ON, Canada: Time
Series Research and Analysis Center, Statistics Canada.
14. Dagum, E. B. (1988). The X11ARIMA/88 seasonal adjustment method-foundations and user’s
manual. Ottawa, ON, Canada: Time Series Research and Analysis Centre, Statistics Canada.
15. Dagum, E. B., Chhab, N., & Chiu, K. (1990). Derivation and properties of the X11ARIMA
and census X11 linear filters. Journal of Official Statistics, 12(4), 329–347.
16. Dawson, D. A., Denton, F. T., Feaver, C., & Robb, A. (1975). Seasonal patterns in the
Canadian Labour force (108 pp.). Economic Council of Canada, Discussion Paper. No. 38.
17. Early, J. (1976). Comment. Brooking Papers on Economic Activity, 1, pp. 239–243.
18. Falkner, H. D. (1924). The measurement of seasonal variation. Journal of the American
Statistical Association, 19, 167–179.
19. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., & Chen, B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
20. Fisher, A. (1937). A brief note on seasonal variation. The Journal of Accountancy, 64, 174–199.
21. Gomez, V., & Maravall, A. (1996). Program TRAMO and SEATS. Madrid: Bank of Spain.
22. Hannan, E. J. (1960). The estimation of seasonal variations. Australian Journal of Statistics, 2,
1–15.
23. Hannan, E. J. (1964). The estimation of changing seasonal pattern. Journal of the American
Statistical Association, 59, 1063–1077.
24. Hannan, E. J., Terrell, R. D., & Tuckwell, N. E. (1970). The seasonal adjustment of economic
time series. International Economic Review, 11, 24–52.
25. Harvey, A. C. (1981). Time series models. Oxford, UK: Phillip Allan Publishers Limited.
26. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
27. Henderson, R., & Sheppard, H. N. (1919). Graduation of mortality and other tables. Actuarial
Studies, No. 4. New York: Actuarial Society of America.
28. Henshaw, R. C. (1966). Application of the general linear model to seasonal adjustment of
economic time series. Econometrica, 34, 381–395.
29. Hillmer, S. C., & Tiao, G. C. (1982). An ARIMA model-based approach to seasonal
adjustment. Journal of the American Statistical Association, 77, 63–70.
30. Jones, H. L. (1943). Fitting polynomial trends to seasonal data by the method of least squares.
Journal of the American Statistical Association, 38, 453–465.
31. Jorgenson, D. N. (1964). Minimum variance, linear unbiased, seasonal adjustment of economic
time series. Journal of the American Statistical Association, 59, 402–421.
32. Joy, A., & Thomas, N. (1928). The use of moving averages in the measurement of seasonal
variations. Journal of the American Statistical Association, 23, 242–252.
78 3 Seasonal Adjustment: Meaning, Purpose, and Methods
33. Judek, S. (1975). Canada’s seasonal variations in employment and unemployment, 1953–73
(118 pp.). Research Paper No. 75. Canada: University of Ottawa.
34. Kendall, M. G., & Stuart, A. (1966). The advanced theory of statistics (Vol. 3, 552 pp.). New
York: Hafner Publishing Co.
35. King, W. F. (1924). An improved method for measuring the seasonal factor. Journal of the
American Statistical Association, 19, 301–313.
36. Kitagawa, G., & Gersch, W. (1984). A smoothness priors state-space modelling of time series
with trend and seasonality. Journal of the American Statistical Association, 79, 378–389.
37. Koopman, S. J., Harvey, A. C., Doornik, J. A., & Shephard, N. (1998). Structural time series
analysis STAMP (5.0). London: Thomson Business Press.
38. Kuznets, S. (1933). Seasonal variations in industry and trade (455 pp.). New York: National
Bureau of Economic Research.
39. Ladd, G. W. (1964). Regression analysis of seasonal data. Journal of the American Statistical
Association, 59, 402–421.
40. Lovell, M. C. (1963). Seasonal adjustment of economic time series and multiple regression
analysis. Journal of the American Statistical Association, 58, 993–1010.
41. Lovell, M. C. (1976). Least squares seasonally adjusted unemployment data. Brooking Papers
on Economic Activity, l, pp. 225–237.
42. Macaulay, F. R. (1931). The smoothing of time series. New York: National Bureau of Economic
Research.
43. Mendershausen, H. (1939). Eliminating changing seasonals by multiple regression analysis.
The Review of Economic Studies, 21, 71–77.
44. Persons, W. M. (1919). Indices of business conditions. Review of Economic Statistics, 1, 5–107.
45. Pierce, D. (1978). Seasonal adjustment when both deterministic and stochastic seasonality are
present. In A. Zellner (Ed.), Proceedings of the NBER/Bureau of the Census Conference on
Seasonal Analysis of Economic Time Series, Washington, DC (pp. 242–269).
46. Shiskin, J. (1973). Measuring current economic fluctuations. In Annals of economic and social
measurements (Vol. 2, pp. 1–16.). Cambridge: National Bureau of Economic Research.
47. Shiskin, J., Young, A. H., & Musgrave, J. C. (1967). The X-11 variant of the Census Method II
seasonal adjustment program. Technical Paper 15 (revised). Washington, DC: US Department
of Commerce, Bureau of the Census.
48. Stephenson, J. A., & Farr, H. T. (1972). Seasonal adjustment of economic data by application of
the general linear statistical model. Journal of the American Statistical Association, 67, 37–45.
49. Wald, A. (1936). Berechnung und Ausschaltung von Saisonschwankungen. Beitrage Zur
Konjunkturforschung. 9, Vienna: Verlag Julius Springer.
50. Wallis, K. E. (1978). Seasonal adjustment and multiplicative time series analysis. In A. Zellner
(Ed.), Proceedings of the NBER/Bureau of the Census Conference on Seasonal Analysis of
Economic Time Series,Washington, DC (pp. 347–357).
51. Whittaker, E., & Robinson, G. (1924). The calculus of observation: A treatise on numerical
mathematics (397 pp.). London: Blackie and Sons Ltd. (also published by Dover Publications,
New York, 1967).
52. Wold, H. O. (1938). A study in the analysis of stationary time series. Uppsala, Sweden:
Almquist and Wiksell (Second Edition 1954).
53. Young, A. H. (1968). A linear approximation to the Census and BLS seasonal adjustment
methods. Journal of the American Statistical Association, 63, 445–457.
Chapter 4
Linear Filters Seasonal Adjustment Methods:
Census Method II and Its Variants
Abstract The best known and most often applied seasonal adjustment methods are
based on smoothing linear filters or moving averages applied sequentially by adding
(and subtracting) one observation at a time. This chapter discusses with details the
basic properties of the symmetric and asymmetric filters of the Census Method II-
X11 method which belong to this class. It also discusses the basic assumptions of its
two more recent variants, X11ARIMA and X12ARIMA. The latter consists of two
linked parts: the regARIMA model for estimation of the deterministic components
(mainly calendar effects), and the decomposition part of the linearized series for the
stochastic components (trend-cycle, seasonality, and irregulars) performed using the
X11 filters combined with those of the ARIMA model extrapolations. An illustrative
example of the seasonal adjustment with the X12ARIMA software default option is
shown with the US New Orders for Durable Goods series. The illustrative example
concentrates on the most important tables of this software that enable to assess the
quality of the seasonal adjustment.
4.1 Introduction
The best known and most often applied seasonal adjustment methods are based on
smoothing linear filters or moving averages applied sequentially by adding (and
subtracting) one observation at a time. These methods assume that the time series
components change through time in a stochastic manner. Given a time series, yt ; t D
1; : : : ; n, for any t far removed from both ends, say, m < t < n m, the seasonally
adjusted value yat is obtained by application of a symmetric moving average W.B/
as follows:
X
m X
m
yat D W.B/yt D wj Bj yt D wj ytj ; (4.1)
jDm jDm
where the weights wj are symmetric, that is, wj D wj , and the length of the average
is 2m C 1.
For current and recent data (n m < t n) a symmetric linear filter cannot be
applied, and therefore truncated asymmetric filters are used. For example, for the
X
m
yan D W0 .B/yn D w0;j ynj : (4.2)
jD0
The asymmetric filters are time-varying in the sense that different filters are
applied to the m first and last observations. The end estimates are revised as
new observations are added because of: (1) the new innovations, and (2) the
differences between the symmetric and asymmetric filters. The estimates obtained
with symmetric filters are often called “final.”
The development of electronic computers contributed to major improvements
in seasonal adjustment based on moving averages and facilitated their massive
application.
In 1954, Julius Shiskin of the US Bureau of Census developed software called
Method I, based mainly on the works of Macauley [22] already being used by the
US Federal Reserve Board. Census Method I was followed by Census Method II and
eleven more experimental versions (X1 , X2, : : :, X11). The best known and widely
applied was the Census Method II-X11 variant developed by Shiskin et al. [26], but
produced poor seasonally adjusted data at the end of the series which is of crucial
importance to assess the direction of the short-term trend and the identification of
turning points in the economy. Dagum [6] developed the X11ARIMA mainly to
correct for this serious limitation and, later Findley et al. [9] developed X12ARIMA
that offers a regARIMA option to estimate deterministic components, such as
trading day variations, moving holiday effects and outliers, simultaneously with the
ARIMA model for extrapolation. It also includes new diagnostic tests and spectral
techniques to assess the goodness of the results. The X12ARIMA method is today
the one most often applied by statistical agencies in the world. The US Bureau of
Census continued research on the development of seasonal adjustment methods,
and recently produced a beta version called X13ARIMA-SEATS which enables the
estimation of the seasonal component either via linear filters as those available in
X12ARIMA or based on an ARIMA model decomposition.
All the seasonal adjustment methods based on moving averages discussed in this
book have nonlinear elements. Hence, the seasonally adjusted total of an aggregated
series is not equal to the algebraic sum of the seasonally adjusted series that enter
into the aggregation. The main causes of nonlinearity are generated by:
1. a multiplicative decomposition model for the unobserved components,
2. the identification and replacement of extreme values,
3. the ARIMA extrapolations, and
4. automatic selection of the moving average length for the estimation of the trend-
cycle and seasonality.
The properties of the combined linear filters applied to estimate the various
components were originally calculated by Young [28] for the standard option of
Census II-X11 variant. Later, Dagum et al. [8] calculated and analyzed all possible
filter combinations of Census II-X11 and X11ARIMA.
4.1 Introduction 81
Given the fact that X11ARIMA and X12ARIMA are methods that are based on
the Bureau of the Census Method II-X11 [26], we shall first briefly discuss the
Census II-X11 variant and refer the reader to Ladiray and Quenneville [18] for a
very detailed description.
The Census II-X11 method assumes that the main components of a time series
follow a multiplicative or an additive model, that is,
(multiplicative model) yt D TCt St It
(additive model) yt D TCt C St C It
where yt stands for the original series, TCt for the trend-cycle, St for the seasonal,
and It for the irregular.
There are no mixed models in this program, such as yt D TCt St C It or other
possible combinations. The estimation of the components is made with different
kinds of smoothing linear filters.
The main steps followed to produce monthly seasonally adjusted data are:
1. Compute the ratios (differences) between the original series and a centered 12
months moving average (2 12 m.a.), that is, a 2 months average of 12 months
(average) as a first estimate of the seasonal and irregular components.
2. Apply a weighted 5-term moving average (3 3 m.a.) to the seasonal–irregular
ratios (differences) for each month separately, to obtain an estimate of the
seasonal factors (effects). Compute a 2 12 m.a. of the preliminary factors for
the entire series. To obtain the six missing values at either end of this average,
repeat the first (last) available value six times.
3. Adjust these seasonal factors (differences) to sum to 12 (0.000 approximately)
over any 12-month period by dividing (subtracting) the centered 12 months
average into (from) the factors (differences).
4. Divide (subtract) the seasonal factors (differences) into (from) the seasonal–
irregular ratios (differences) to obtain an estimate of the irregular component.
5. Compute a moving 5-year standard deviation () of the estimates of the
irregular component and test the irregulars in the central year of the 5-year
period against 2:5. Remove values beyond 2:5 as extreme and recompute
the moving 5-year .
6. Assign a zero weight to irregulars beyond 2:5 and a weight of one (full weight)
to irregulars within 1:5. Assign a linearly graduated weight—between 0 and
1 to irregulars between 2:5 and 1:5.
7. For each separate month apply a weighted 5-term moving average (3 3 m.a.)
to the seasonal–irregular ratios (differences) with extreme values replaced by
the corresponding values in step 6, to estimate preliminary seasonal factors
(effects).
8. Repeat step 3.
9. To obtain a preliminary seasonally adjusted series divide (subtract) the series
obtained at step 8 into (from) the original series.
82 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
10. Apply a 13-term Henderson moving average to the seasonally adjusted series
and divide (subtract) the resulting trend-cycle into (from) the original series to
obtain a second estimate of the seasonal–irregular ratios (differences).
11. Apply a weighted 7-term moving average (3 5 m.a.) to the seasonal–irregular
ratios (differences) for each month separately to obtain a second estimate of the
seasonal component.
12. Repeat step 3.
13. Divide (subtract) the series obtained at step 11 into (from) the original series to
obtain a seasonally adjusted series.
The filters are applied sequentially and the thirteen steps are repeated twice.
The Census II-X11 variant produces seasonal factor forecasts for each month
which are given by
1
Sj;tC1 D Sj;t C .Sj;t Sj;t1 /; j D 1; : : : ; 12;
2
where j is the month and t the current year. The use of seasonal factor forecasts was
very popular in statistical Bureaus till the end of 1970s. In 1975, Dagum proposed
the use of concurrent seasonal factors obtained from data that included the most
recent value and this was adopted by Statistics Canada. Gradually, other statistical
agencies applied concurrent seasonal factors and this is now the standard practice.
The final trend-cycle is obtained by 9-, 13-, or 23-term Henderson moving
averages applied to the final monthly seasonally adjusted series (5-term and 7-term
Henderson filters are used for quarterly data). The selection of the appropriate filter
is made on the basis of a preliminary estimate of the I=C ratio (the ratio of the
average absolute month-to-month change in the irregular to that in the trend-cycle).
A 9-term is applied to less irregular series and a 23-term to highly irregular series.
Using a linear approximation of the Census Method II, Young [28] arrived at the
conclusion that a 145-term moving average was needed to estimate one seasonal
factor with central weights if the trend-cycle component is adjusted with a 13-
term Henderson moving average. The first and last 72 seasonal factors (6 years)
are estimated using sets of asymmetrical end weights. It is important to point out,
however, that the weights given to the most distant observations are very small and,
therefore, the moving average can be very well approximated by taking one half
of the total number of terms plus one. So, if a 145-term moving average is used
to estimate the seasonal factor of the central observation, a good approximation
is obtained with only 73 terms, i.e., 6 years of observations. This means that the
seasonal factor estimates from unadjusted series whose observations end at least
3 years later can be considered “final”, in the sense that they will not change
significantly when new observations are added to the raw data.
4.2 Basic Properties of the Two-Sided Linear Smoothing Filters of Census. . . 83
The linear smoothing filters applied by Census Method II-X11 variant to produce
seasonally adjusted data can be classified according to the distribution of their set
of weights in symmetric (two-sided) and asymmetric (one-sided). The symmetric
moving averages are used to estimate the component values that fall in the middle
of the span of the average, say, 2m C 1, and the asymmetric moving averages, to the
m first and last observations.
The sum of the weights of both kinds of filters is one and thus, the mean of
the original series is unchanged in the filtering process. The sum of the weights of
a filter determines the ratio of the mean of the smoothed series to the mean of the
unadjusted series assuming that these means are computed over periods long enough
to ensure stable results. It is very important in filter design that the filter does not
displace in time the components of the output relative to those of the input. In other
words, the filter must not introduce phase shifts.
Symmetric moving averages introduce no time displacement for some of the
components of the original series and a displacement of 180ı for others. A phase
shift of 180ı is interpreted as a reverse in polarity which means that maxima are
turned into minima and vice versa. In other words, peaks (troughs) in the input
are changed into troughs (peaks) in the output. For practical purposes, however,
symmetric moving averages act as though the time displacement is null. This is so
because the sinusoids that will have a phase shift of 180ı in the filtering process are
cycles of short periodicity (annual or less) and moving averages tend to suppress or
significantly reduce their presence in the output. In spectral analysis, the phase is a
dimensionless parameter that measures the displacement of the sinusoid relative to
the time origin. Because of the periodic repetition of the sinusoid, the phase can be
restricted to ˙180ı . The phase is a function of the frequency of the sinusoid, being
the frequency equal to the reciprocal of the length of time or period required for one
complete oscillation.
The next two sections discuss the basic properties of the two-sided and one-
sided filters applied by the Census II-X11 method. The study is based on a spectral
analysis of the corresponding filters.
The centered 12 months moving average used for a preliminary estimate of the
trend-cycle (step 1) reproduces exactly a linear trend and annihilates a stable
seasonal in an additive model. However, if the relationship among the components
is multiplicative, this filter will reproduce only a constant trend and eliminate a
stable seasonality. Unless the cycles are long (5 years or more) and the irregulars
quite small, this filter will not follow closely the peaks and troughs and will not
84 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
smooth the data successfully. If a sine curve of 3-year periodicity and amplitude
100 is input to this filter, the output is a sine curve of equal periodicity but with
amplitude reduced to 82.50. For a 5-year cycle the amplitude is reduced to 93.48
and for a 10-year cycle is 98.33. In
general,
any simple moving average of m equal
2
weights fitted to a parabola falls m121 c inside the parabola, see Macauley [22].
Hence, this filter fitted to a second degree parabola yt D a C bt C ct2 will give a
smoothed curve that falls 12.17 units inside the parabola, i.e., yO t D yt C 12:17c. The
bias introduced by this filter will not affect significantly the preliminary estimate
of the seasonal component if the trend is nearly linear or if the direction of the
curvature changes frequently. In the latter case, the bias will be sometimes positive
and sometimes negative, so it may average out in applying the seasonal filters to the
Seasonal–Irregular (SI) ratios for each month separately over several years.
The Henderson [14] moving averages are applied to obtain an improved estimate
of the trend-cycle (step 11). These linear smoothing filters were developed by
summation formulae. Their weights minimize the sum of squares of the third
differences of the smoothed curve. The latter is equivalent to say that these filters
reproduce the middle value of a function that is a third degree parabola within the
span of the filter. The Henderson’s moving averages give the same results as would
be obtained by smoothing the middle value of a third degree parabola fitted by
weighted least squares, where the weights given to the deviations are as smooth
as possible.
The fact that the trend-cycle is assumed to follow a cubic over an interval of short
duration (between 1 and 2 years approximately) makes these filters very adequate
for economic time series.
None of the Henderson’s filters used by the Census II-X11 program eliminates
the seasonal component, but since they are applied to already seasonally adjusted
data, this limitation becomes irrelevant. On the other hand, they are extremely good
to pass sines of any periodicity longer than a year. Thus, the 13 months Henderson,
which is the most frequently applied, will not reduce the amplitude of sines of
20 months periodicity or more that stand for trend-cycle variations. Moreover, it
eliminates almost all the irregular variations that can be represented by sines of very
short periodicity, 6 months or less.
It is inherent to any moving average procedure that the first and last m points of an
original series cannot be smoothed with the same set of symmetric 2m C 1 weights
applied to the middle values. The Census II-X11 method uses one-sided filters to
smooth these end points. Moving averages with asymmetric weights are bounded to
introduce phase shifts for all the components of the original series. This is a very
undesirable characteristic since it may cause, for example, a business cycle in the
smoothed series to turn down before or after it actually does in the unadjusted data.
The asymmetric weights applied in conjunction with the Henderson symmetric
filters, developed by Musgrave [23], no longer estimate a cubic within their
span. In fact, the weight associated with the 13-term Henderson filter, applied to
estimate the last data point, amplifies sinusoids of 3 to 1 year periodicities up to
10 %, passes between 37 and 17 % of the amplitude of sine waves representing
irregular variations, and it has a linearly decreasing phase shift for sinusoids of
long periodicity. The latter means that the smoothed series lags the original series
a number of units of time equal to the slope of the line. For an 18-month cycle, the
phase shift is about 4 %, i.e., a lag of more than half of a month.
The set of end weights used in conjunction with the Henderson filters can
estimate well only a linear trend-cycle within the length of the filters [23].
86 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
Concerning the other two seasonal one-sided filters, they also introduce phase
shifts and can only estimate stable or nearly stable seasonality. There are other
reasons for the poor quality of both the current and seasonal factor forecasts
generated by the Census II-X11 method. As mentioned above, the program repeats
the first and last six preliminary seasonal factors obtained with the 33 m.a. In other
words, for a series ending, say, in December 1977, the seasonal factors from July to
December of 1977 are equal to the corresponding seasonal factors of the previous
year. This operation is repeated any time that the centered 12-term moving average is
applied to estimate the trend-cycle. In the past, when seasonal factor forecasts were
used to produce concurrent seasonal adjustment, the last 6 months were subject to
larger errors than the corresponding factors for the first 6 months because they were
based on observations lagging 2 years with respect to the forecasting year. It was
then preferable to forecast only 6 months instead of 12. By concurrent seasonal
adjustment is meant the adjustment of the original series made with seasonal factors
obtained from a series that includes the last observed point.
one revision will give seasonal factors that are “final” in a statistical sense. There
are several reasons for the significant reduction of the error in the seasonal factor
forecasts and concurrent seasonal factors. The X11ARIMA produces seasonal
factor forecasts from the combination of two filters:
a. the filters of the autoregressive integrated moving averages (ARIMA) models
used to forecast the raw data; and
b. the filters that Census Method II-X11 variant applies for concurrent seasonal
adjustment.
In this manner, the seasonal factor forecasts are obtained from the forecasted
raw values with filters that even though are still one-sided, they are closer to the
two-sided filters as compared to the forecasting function of the Census II-X11
variant.
5. The trend-cycle estimate for the last observation is made with the symmetric
weights of the Henderson moving average which can reproduce a cubic within
their time span. This is relevant for years with turning points because the
Census II-X11 applies the asymmetric weights of the Henderson filters which
can adequately estimate only a linear trend.
6. Finally, by adding one more year of preliminary data (with no extremes since
they are forecasts) a better estimate of the variance of the irregulars is obtained.
The latter allows a significant improvement in the identification and replacement
of outliers which, as we already know, can severely distort the estimates obtained
with linear moving filters.
For concurrent seasonal factors, the same observations are valid except that the
seasonal filters are closer to the two-sided filters than those corresponding to the
seasonal factor forecasts. For this reason, the number of revisions in the seasonal
factor estimates is also significantly reduced.
We will here analyze the forecasting filters of the ARIMA models. The ARIMA
models applied to seasonal series belong to the general multiplicative Box and
Jenkins type, that is,
where s denotes the periodicity of the seasonal component (equal to 12 for monthly
series and 4 for quarterly data), B is the backshift operator, such that Byt D yt1 ,
p .B/ D .1 1 B p Bp / is the nonseasonal Autoregressive (AR) operator of
order p, ˚P .Bs / D .1 ˚1 Bs ˚P BPs / is the seasonal AR operator of order P,
q .B/ D .11 B q Bq / is the nonseasonal Moving Average (MA) operator of
4.4 The X11ARIMA Method 89
where Et .ytCl / is the conditional expectation of ytCl taken at origin t. Thus, the
forecasts yO t .l/ for each lead time are computed from previous observed y’s, previous
forecast of y’s, and current and previous random shocks a’s. The unknown a’s
are replaced by zeros. In other words, they are weighted averages of previous
observations and forecasts made at previous lead times from the same time origin.
In general, if the moving average operator .B/ D q .B/Q .Bs / is of degree
qCsQ, the forecast equations for yO t .1/; yO t .2/; : : : ; yO t .qCsQ/ will depend directly on
the a’s, but forecasts at longer lead times will not. The latter will receive indirectly
the impact of the a’s by means of the previous forecasts. In fact, yO t .q C sQ C 1/ will
depend on the q C sQ previous yO t which in turn will depend on the a’s.
In the case of X11ARIMA, the lead time of the forecasts for monthly series is
often equal to 12 and the identification of the ARIMA models has a Q generally
equal to 1. Therefore, the forecasts depend directly on the a’s.
Finally, it is important to keep in mind that it is the general autoregressive
operator .B/ defined above that determines whether the forecasting function is to
be a polynomial, a mixture of sines and cosines, a mixture of exponentials, or some
combinations of these functions. The one-step ARIMA forecasts are minimum mean
square error forecasts and are easily updated as new raw values become available.
The ARIMA models for X11ARIMA must fulfill the double condition of (1)
fitting the data well, and (2) generating acceptable forecasts, where by acceptable
it is understood forecasts that have a mean absolute error smaller than 5 % for well
behaved series (e.g., employment males 20 years and over) and smaller than 12 %
for more irregular series (e.g., unemployment females 16–19 years).
90 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
In the automated version of this method, the user has the choice of: (a) submitting
his/her own ARIMA model, or (b) letting the program to select automatically an
ARIMA model that passes the above guidelines.
The automatic selection is made from three ARIMA models that have been
successfully chosen from a set of 12 ARIMA models testing out of sample
extrapolated values for the 4 last years. The sample consisted of 174 monthly and
quarterly economic time series of 15 years of length. The three models passed the
guidelines in about 80 % of all the cases. The 12 models originally tested were:
The automated option chooses for a multiplicative decomposition from the follow-
ing ARIMA models:
(1) log(0,1,1)(0,1,1)s
(2) log(0,2,2)(0,1,1)s
(3) (2,1,2)(0,1,1)s
A set of new statistical tests, tables, and graphs has been incorporated into the
automated version of the X11ARIMA. These tests are used to assess the quality of
the original series and the reliability of the seasonal adjustment. A brief description
of these main improvements follows.
differences between the annual average of the SI, for each year, and the average
of the SI for the whole table, corrected by the corresponding degrees of freedom.
3. The “residual” variance, denoted by r2 , equal to the total variance minus the sum
of the “between months or quarters” variance and the “between years” variance.
The F-ratio for the presence of moving seasonality is the quotient between the
“between years” variance and the “residual” variance.
To calculate the variances in an additive model, the (S C I) data are taken
in absolute value, otherwise the annual average is always equal to zero. For a
multiplicative model, the SI ratios are replaced by absolute deviations from 100,
i.e., by j SI 100 j. Contrary to the previous test, for which a high value of the F-
statistic is a good indication of the presence of identifiable seasonality, a high value
of the F-statistic corresponding to moving seasonality is a bad sign. The F-test,
printed in Table D8, indicates whether moving seasonality is present or not.
The presence of moving seasonality could be taken as an indication of residual
seasonality when Census II-X11 is used, because this program forces stable or
nearly stable seasonality on the first and last 3 years of data. This is partially
corrected by the ARIMA option.
This test combines the previous test for the presence of moving seasonality with the
Census II-X11 test for the presence of stable seasonality and another nonparametric
test for the presence of stable seasonality, the Kruskal–Wallis chi-squared test,
whose value is also printed in Table D8.
The main purpose of this test is to determine whether the seasonality of the
series is “identifiable” or not. For example, if there is little stable seasonality and
most of the process is dominated by highly moving seasonality, the chances are that
the seasonal factors will not be well estimated because they will not be properly
identified by the Census II-X11 program.
This test basically consists of combining the F-values obtained from the two
previous tests as follows:
1. If the test for the presence of stable seasonality, denoted by FS , fails at the 1 %
level of significance, the null hypothesis (that seasonality is not identifiable) is
accepted.
2. If step 1 passes but the test for the presence of moving seasonality, indicated by
FM , fails at the 5 % level of significance, then this FM value is combined with the
FS value from step 1 to give
7 3FM
T1 D and T2 D ;
FM FS FS
4.4 The X11ARIMA Method 93
This is an F-test applied to the values of Table D11 and calculated for the whole
length of the series, and for only the last 3 years. The effect of the trend is removed
by a first order difference of lag three, that is, (Oyt -Oyt3 ) where yO t are the values of
Table D11. Two F-ratios are printed at the end of the table, as well as a message
saying whether residual seasonality is present or not for the last 3 years.
The Census II-X11 program uses the Average Duration of Run (ADR) statistic to
test for autocorrelation in the final estimated residuals obtained from Table D13.
This nonparametric test was developed by Wallis and Moore [27] and is based
on the number of turning points. It is only efficient for testing the randomness of
the residuals against the alternative hypothesis that the errors follow a first order
autoregressive process of the form It D It1 C et , where is the autocorrelation
coefficient and fet g is a white noise process. This test, however, is not efficient
for detecting the existence of periodicity in the residuals, which can occur when
relatively long series are seasonally adjusted, or when the relative variation of the
seasonal component is small compared to that of the irregulars.
To test the independence of the residuals against the alternative hypothesis of
being a periodic process, the normalized cumulative periodogram as developed by
Box and Jenkins [3] has been incorporated.
The normalized cumulative periodogram values are given in a table and also as a
graph. By visual inspection, it is possible to determine whether or not components
with certain periodicity are present in the residuals. If the residuals are a sample
realization of a purely random process, and if the size of the sample tends to infinity,
then the normalized cumulative periodogram tends to coincide with the diagonal of
the square in which it is drawn.
94 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
This new table, D11A, produces a modified seasonally adjusted series, where the
annual totals of the seasonally adjusted values are made equal to those of the raw
data. The discrepancy between the annual totals is distributed over the seasonally
adjusted values of Table D11 in such a way as to minimize the distortion on
the month-to-month or quarter-to-quarter movements of the originally seasonally
adjusted series. The procedure is based on a quadratic minimization of the first
differences of the annual discrepancies, expressed as differences or ratios. This
procedure was originally developed by Hout [16] and improved by Cholette [5].
The Statistics Canada Census II-X11 version, as developed in 1975, had two
statistics called Q1 and Q2 that provided an indication of the size and nature of
the irregular and seasonal components, respectively. A description of these statistics
and their basic assumptions are discussed by Hout and De Fontenay [17].
Considerable research has been carried out since the first set of guidelines was
developed and there is now only one Q statistic, which results from a combination
of several other measures developed by Lothian and Morry [21]. Most of these
are obtained from the summary measures printed in Table F2 of the X11ARIMA
computer program. Their values vary between 0 and 3, and only values less than
1 are considered as acceptable. The set of statistics to produce the final Q statistic
is:
1. the relative contribution of the irregular component over spans of 3 months, as
obtained from Table F2, denoted by M1 ;
2. the relative contribution of the irregular component to the stationary portion of
the variance, denoted by M2 ;
3. the value of the I=C ratio between the average absolute month-to-month (or
quarter-to-quarter) percentage change in the irregular and that in the trend-
cycle, from Table D7 and denoted by M3 ;
4. the value of the average duration of run for the irregulars from Table F2 and
denoted by M4 ;
5. the MCD or QCD (the number of months or quarters it takes the average
absolute change in the trend-cycle to dominate that in the irregular), from Table
F2 and denoted by M5 ;
4.5 The X12ARIMA Method 95
6. the global I=S moving seasonality ratio, obtained as an average of the monthly
moving seasonality ratios from Table D9A and denoted by M6 (it is the ratio of
the average absolute year-to-year percentage change in the irregular factors to
that in the seasonal factors.);
7. the amount of stable seasonality in relation to the amount of moving seasonality
from the tests of Table D8 and denoted by M7 ;
8. a measure of the variation of the seasonal component for the whole series from
Table F2 and denoted by M8 ;
9. the average linear movement of the seasonal component for the whole series,
denoted by M9 ;
10. same as 8, but calculated for recent years only and denoted by M10 ; and
11. same as 9, but calculated for recent years only and denoted by M11 .
Furthermore, there are two new tables that show the outliers, and a D16 table
that shows the total effect of the trading day factors and the seasonal effects. Several
other modifications were introduced to correct for some limitations of the Census
II-X11 variant [7].
The X12ARIMA is today the most often applied seasonal adjustment method by
statistical agencies. It was developed by Findley et al. [9] and is an enhanced version
of the X11ARIMA method.
The major modifications concern: (1) extending the automatic identification and
estimation of ARIMA models for the extrapolation option to many more than the
three models available in X11ARIMA, and (2) estimating trading day variations,
moving holidays, and outliers in what is called regARIMA. The latter consists
of regression models with ARIMA (AutoRegressive Integrated Moving Average)
errors. More precisely, they are models in which the mean function of the time series
(or its logs) is described by a linear combination of regressors, and the covariance
structure of the series is that of an ARIMA process. If no regressors are used,
indicating that the mean is assumed to be zero, the regARIMA model reduces to
an ARIMA model.
Whether or not special problems requiring the use of regressors are present in
the series to be adjusted, a fundamentally important use of regARIMA models is
to extend the series with forecasts (and backcasts) in order to improve the seasonal
adjustments of the most recent (and the earliest) data. Doing this reduces problems
inherent in the trend estimation and asymmetric seasonal averaging processes of the
type used by the Census II-X11 method near the ends of the series. The provision
of this extension was the most important improvement offered by the X11ARIMA
program. Its theoretical and empirical benefits have been documented in many
publications, such as Dagum [7], Bobbit and Otto [2], and references therein.
96 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
The X12ARIMA method has all the seasonal adjustment capabilities of the
X11ARIMA variant. The same seasonal and trend moving averages are available,
and the program still offers the Census II-X11 calendar and holiday adjustment
routines incorporated in X11ARIMA. Several new options have been included,
namely:
1. sliding spans diagnostic procedures, developed by Findley et al. [10];
2. capability to produce the revision history of a given seasonal adjustment;
3. a new Henderson trend filter routine which allows the user to choose any odd
number for the length of the Henderson symmetric filter;
4. new options for seasonal filters;
5. several new outlier detection options for the irregular component of the seasonal
adjustment;
6. a new table of trading day factors by type of day; and
7. a pseudo-additive seasonal adjustment mode.
The modeling module is designed for regARIMA model building with sea-
sonal socioeconomic time series. To this end, several categories of predetermined
regression variables are available, including trend constants or overall means, fixed
seasonal effects, trading day effects, holiday effects, additive outliers, level shifts,
temporary change outliers, and ramp effects. User-specified regression variables can
also be included in the models.
The specification of a regARIMA model requires specification of both the regres-
sion variables to be included in the model and the type of ARIMA model for the
regression errors (i.e., the order .p; d; q/.P; D; Q/s ). Specification of the regression
variables depends on user knowledge about the series being modeled. Identification
of the ARIMA model for the regression errors follows well-established procedures
based on examination of various sample autocorrelation and partial autocorrelation
functions produced by the X12ARIMA program.
Once a regARIMA model has been specified, X12ARIMA estimates its param-
eters by maximum likelihood using an Iterative Generalized Least Squares (IGLS)
algorithm. Diagnostic checking involves examination of residuals from the fitted
model for signs of model inadequacy. X12ARIMA produces several standard
residual diagnostics for model checking, as well as provides sophisticated methods
for detecting additive outliers and level shifts. Finally, X12ARIMA can produce
point forecasts, forecast standard errors, and prediction intervals from the fitted
regARIMA model. In addition to these modeling features, X12ARIMA has an
automatic model selection procedure based mainly on the automatic model selection
procedure of TRAMO [12]. There are also options that use AICC, a version of
Akaike’s AIC that adjusts for the length of the series being modeled, to determine if
user-specified regression variables (such as trading day or Easter regressors) should
be included into a particular series. Also, histories can be generated for likelihood
statistics (such as AICC) and forecasts to facilitate comparisons between competing
models.
4.5 The X12ARIMA Method 97
ARIMA models are frequently used for seasonal time series. A general multi-
plicative seasonal ARIMA model for a time series yt ; t D 1; : : : ; n, is specified
in Eq. (4.3). A useful extension of ARIMA models results from the use of a
time-varying mean function modeled via linear regression effects. More explicitly,
suppose we write a linear regression equation for the time series yt as
X
r
yt D ˇi xit C zt (4.7)
iD1
The regARIMA model (4.8) can be thought of either as generalising P the pure
ARIMA model (4.3) to allow for a regression mean function ( riD1 ˇi xit ), or as
generalising the regression model (4.7) to allow the errors zt to follow the ARIMA
model (4.3). In any case, notice that the regARIMA model implies that first the
regression effects are subtracted from yt to get the zero mean series zt , then the
error series zt is differenced to get a stationary series, say wt , assumed to follow a
stationary ARMA model, .B/˚.Bs /wt D .B/.Bs /at .
Another way to write the regARIMA model (4.8) is
X
.1 B/d .1 Bs /D yt D ˇi .1 B/d .1 Bs /D xit C wt ; (4.9)
i
where wt follows the stationary ARMA model just given before. Equation (4.9)
emphasizes that the regression variables xit in the regARIMA model, as well as in
the series yt , are differenced by .1 B/d .1 Bs /D .
Notice that the regARIMA model as written in (4.8) assumes that the regression
variables xit ; i D 1; : : : ; r, affect the dependent series yt only at concurrent time
98 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
points, i.e., model (4.8) does not explicitly provide for lagged regression effects such
as ˇi xi;t1 . Lagged effects can be included by the X12ARIMA program, however,
by reading appropriate user-defined lagged regression variables.
The X12ARIMA program provides additional flexibility in the specification
of the ARIMA part of the regARIMA model by permitting (1) more than two
multiplicative ARIMA factors, (2) missing lags within the AR and MA polynomials,
(3) the fixing of individual AR and MA parameters at user-specified values when the
model is estimated, and (4) inclusion of a trend constant, which is a nonzero overall
mean for the differenced series .1 B/d .1 Bs /D yt .
The most basic regression variable is the constant term. If the ARIMA model
does not involve differencing, this is the usual regression intercept, which, if there
are no other regression variables in the model, represents the mean of the series.
If the ARIMA model does involve differencing, X12ARIMA uses a regression
variable such that, when it is differenced according to the ARIMA model [see
Eq. (4.9)], a column of ones is produced. The corresponding parameter is then
called a trend constant, since it provides for a polynomial trend of the same degree
as the order of differencing in the model. For example, with d > 0 and D D 0,
the (undifferenced) trend constant regression variable is proportional to td . Notice
that the lower order polynomial terms, tj for 0 j < d, are not included among
the regression variables because they would be differenced to zero by .1 B/d ,
hence their coefficients cannot be estimated. With or without the trend constant,
the model (4.8) (or 4.9) implicitly allows for these lower order polynomial terms
through the differencing. If seasonal differencing is requested (D > 0), the nature
of the undifferenced trend constant regression variable is more complicated, though
the trend constant can be thought of as allowing for a polynomial of degree d C sD.
Without a trend constant, model (4.8) implicitly allows for a polynomial of degree
d C sD 1.
Fixed seasonal effects in a monthly series can be modeled using 12 indicator
variables, one for each calendar month. Since these 12 variables always add to
one, they are confounded with an overall level effect. This leads to one of two
singularity problems: collinearity with the usual constant term in a model with no
differencing; or a singularity in a model with differencing since the 12 variables,
when differenced, always sum to 0. One appropriate reparameterization instead uses
11 contrasts in the 12 indicator variables. An alternative reparameterization uses
11 variables taken from the Fourier (trigonometric) series representation of a fixed
monthly pattern. X12ARIMA allows either of these options, and also allows spec-
ifying the trigonometric terms only for selected frequencies. For quarterly series,
or for series with other seasonal periods, X12ARIMA constructs the appropriate
versions of these variables. Notice that these variables cannot be used in a model
with seasonal differencing, as they would all be differenced to zero.
Trading day effects occur when a series is affected by the different day-of-the-
week compositions of the same calendar month in different years. Trading day
effects can be modeled with seven variables that represent (no. of Mondays), . . . ,
(no. of Sundays) in month t. Bell and Hillmer [1] proposed a better parameterization
of the same effects using six variables defined as (no. of Mondays/ .no. of
4.5 The X12ARIMA Method 99
and remains at the new level until the day before the holiday. For Thanksgiving the
model used assumes that the level of activity changes on the day that is a specified
number of days before or after Thanksgiving and remains at the new level until
December 24. The regression variable constructed for the holiday effect is, for a
given month t, the proportion of the affected time period that falls in month t.
Actually, these regressors are deseasonalized by subtracting their long-run monthly
means. Essentially the same Easter effect variable applies also to quarterly flow time
series, but Labor Day and Thanksgiving effects are not present in quarterly series.
X12ARIMA provides four other types of regression variables to deal with abrupt
changes in the level of a series of a temporary or permanent nature: additive outliers
(AO), level shifts (LS), temporary changes (TC), and ramps. AOs affect only one
observation in the time series, LSs increase or decrease all observations from a
certain time point onward by some constant amount, TCs allow for an abrupt
increase or decrease in the level of the series that returns to its previous level
exponentially rapidly, and ramps allow for a linear increase or decrease in the level
of the series over a specified time interval. LS regression variables are defined as
1 and then 0, in preference to an equivalent 0 and then 1, to make the overall
level of the regression mean function of any forecasts consistent with the most
recent level of the time series. Similar considerations dictate the definition of ramp
variables. Often, however, large seasonal movements make it difficult to identify
where such changes in level have occurred. Identifying the location and nature of
potential outliers is the object of the outlier detection methodology implemented .
This methodology can be used to detect AOs, TCs, and LSs (not ramps); any that are
detected are automatically added to the model as regression variables. Prespecified
AOs, LSs, TCs, and ramps are actually simple forms of interventions as discussed
by Box and Tiao [4]. While X12ARIMA does not provide the full range of dynamic
intervention effects discussed by Box and Tiao [4], often a short sequence of suitably
chosen AO, LS, TC, and/or ramp variables can produce reasonable approximations
to more complex dynamic intervention effects, although with one or two additional
parameters.
The main source of these new options is an extensive set of time series model
building facilities for fitting what are called regARIMA models, that is, regression
models with ARIMA errors. A fundamental use of regARIMA models is to
extend the series with forecasts (and backcasts) in order to improve the seasonal
adjustments of the most recent (and the earliest) data. Doing this mitigates problems
inherent in the trend estimation and asymmetric seasonal averaging processes of
the type used by the Census Method II-X11 near the ends of the series. The
introduction of the ARIMA extension was a major improvement introduced by
Estela Bee Dagum in the X11ARIMA. Its benefits, both theoretical and empirical,
have been documented in many publications, see Dagum [7], Bobbitt and Otto [2],
and references therein.
X12ARIMA is available as an executable program for computers running DOS
(version 3.0 or higher), Sun 4 UNIX workstations, and VAX/VMS computers.
When released, the X12ARIMA program is in the public domain, and can be
copied or transferred. The corresponding documentation and examples are available
online at http://www.census.gov/srd/www/x12a/. X12ARIMA is also implemented
in OxMetrics, in the open source software Gretl, and in the Eurostat software
Demetra+, and its recent extension JDemetra+. Since the 1990s, in the European
countries, Eurostat has been playing a leading role in the promotion, development,
and maintenance of source available software solution for seasonal adjustment of
time series in line with established best practices. In this regard, Eurostat developed
Demetra to provide a convenient and flexible tool for seasonal adjustment using
either TRAMO/SEATS and X12ARIMA methods. Subsequently, Demetra+ was
developed by the National Bank of Belgium, but due to constraints implied by the
use of FORTRAN components, Eurostat created a new seasonal adjustment software
which is platform independent and extensible, that is, JDemetra+ [13]. It is highly
influenced by the output of TRAMO/SEATS, structural models, and X12ARIMA.
JDemetra+ runs on operating systems that support the Java VM (Virtual Machine),
such as Microsoft Windows XP SP3/Vista SP1/Windows 7, Ubuntu 9.10, various
other Linux distributions, Solaris OS version 11 Express (SPARC and x86/x64
Platform Edition), and Macintosh OS X 10.6 Intel.
In this section, we make use of the X12ARIMA program as developed by the
US Bureau of Census to perform the seasonal adjustment of the US New Orders for
Durable Goods (NODG) series, observed from February 1992 to December 2013.
We do not provide a detailed discussion of all the tables and results produced by the
software, but we focus on the most important outputs for the interpretation of the
seasonal adjustment of the NODG series.
To apply X12ARIMA to any particular time series, an input file, called specification
file, must be created. This ASCII (or text) file contains a set of specifications that
X12ARIMA reads to obtain the information it needs about the time series data,
102 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
the time series model to be used, the analysis to be performed, and the output
desired. This specification file has the extension .spc. It is common practice to allow
the software to test automatically for the logarithmic or level specification, for the
presence of trading days effects as well as Easter and leap-year effects. Furthermore,
an automatic detection of several types of outliers [Additive Outliers (AO) and
Level Shift (LS)] is generally required by the software. The most appropriate
ARIMA(p; d; q)(P; D; Q/s model can be automatically selected by the software.
Based on these considerations, we have written the specification file presented in
Table 4.1. In detail, we have selected the following options:
• series describes the time series data (data), start date (start), seasonal period
(period), and the series title (title);
• transform illustrates that we allow the software to automatically select the most
appropriate transformation (level or log) for the data;
• regression specifies the regression variables used to form the regression part of
the regARIMA model. In particular, we allow the estimation of the trading day
effects (td), Easter effect over a span of 9 days before and after Easter (easter[9]),
as well as of the leap-year effect if the log transformation is selected;
• outlier allows for an automatic detection of additive outliers and/or level shifts
using the estimated model;
• pickmdl specifies an automatic selection procedure based on X11ARIMA [7]. It
differs from the option automdl according to which an automatic model selection
procedure is performed based on TRAMO [12]. The latter specification is imple-
mented by default in both Gretl and JDemetra+. However, it should be noticed
that we have found that this option generally selects less parsimonious models
than needed. The main reason be the fact that it is automatically used together
with the simplified trading day model with only one regressor. This generally
produces wrong estimates that are left in the residuals and must be picked up
by the ARIMA model which is then more complex than necessary. Using this
option the model chosen for the log transformed data of the US New Orders for
Durable Goods time series is the ARIMA(1,1,2)(0,1,1)12, but using the trading
Kruskal–Wallis test is the second test for stable seasonality provided by the
software. The test is calculated for the final estimation of the unmodified SI
component from which k samples are derived (k D 12 for monthly series and
k D 4 for quarterly series) of size n1 ; n2 ; : : : ; nk , respectively. The test is based
on the statistic:
12 X Sjk
WD 3.n C 1/;
n.n C 1/ jD1 nj
where Sj is the sum of the ranks of the observations from the j-th sample within
P
the whole sample of n D kjD1 nj observations. Under the null hypothesis, the test
statistic follows a chi-square distribution with k 1 degrees of freedom. Table 4.2
shows that, for the NODG series, the outcome of the test is stable seasonality
present.
A more detailed description of these two tests is provided in Sect. 4.4.3, and
we refer the reader to that section for the theoretical description of the moving
seasonality test as well as of the combined seasonality test. As shown in Table 4.2,
4.6 Illustrative Example: X12ARIMA Seasonal Adjustment of the US NODG. . . 105
all the tests confirm the presence of stable seasonality and the absence of moving
seasonality in the US NODG series.
4.6.3 Pre-processing
4.6.3.1 Regressors
Table 4.3 Automatic selection between level and logarithmic transformation of the NODG series
Likelihood statistics for model fit to untransformed series.
Likelihood statistics
Number of observations (nobs) 263
Effective number of observations (nefobs) 250
Number of parameters estimated (np) 11
AIC 5258.6176
AICC (F-corrected-AIC) 5259.7268
Hannan–Quinn 5274.2077
BIC 5297.3537
Likelihood statistics for model fit to log transformed series.
Likelihood statistics
Number of observations (nobs) 263
Effective number of observations (nefobs) 250
Number of parameters estimated (np) 10
Log likelihood 428.0586
Transformation Adjustment 3029.8684
Adjusted Log likelihood (L) 2601.8098
AIC 5223.6195
AICC (F-corrected-AIC) 5224.5400
Hannan–Quinn 5237.7924
BIC 5258.8342
AICC (with aicdiffD 2:00) prefers log transformation
Multiplicative seasonal adjustment will be performed.
For the log transformed NODG series under study, the following ARIMA(0,1,1)
(0,1,1)12 (airline) model has been automatically selected and estimated by the
software:
where at is assumed to follow a white noise process with zero mean and estimated
variance equal to 0.0124. All the coefficients result significantly different from zero
and satisfy the invertibility conditions (Table 4.5).
Similar to X11ARIMA the ARIMA model to be used in the context of the
X12ARIMA method must fulfill the double condition of fitting the data well
and generating “acceptable” projections for the last 3 years of observed data. By
“acceptable” projection is meant a projection with a mean absolute error smaller
than 5 % for well-behaved series (e.g., unemployment adult males) and smaller than
12 % for highly irregular series (e.g., unemployment teenage males). In this regard,
4.6 Illustrative Example: X12ARIMA Seasonal Adjustment of the US NODG. . . 107
4.6.3.3 Residuals
Beside the Portmanteau test previously introduced, a more complete analysis of the
residuals of the estimated ARIMA model has been carried out by specifying the
option check in the specification file. As suggested by the Portmanteau test, the
residuals satisfy the assumption of being generated by a purely random process
108 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
102
101
100
99
98
Fig. 4.1 Trading day estimates of the US New Orders for Durable Goods
101.5
100.5
99.5
98.5
Seasonal MA
Lag 12 0.7018 0.04671
Variance 0.12400E02
SE of Var 0.11091E03
Likelihood statistics
Number of observations (nobs) 263
Effective number of observations (nefobs) 250
Number of parameters estimated (np) 13
Log likelihood 477.6115
Transformation Adjustment 3029.8684
Adjusted Log likelihood (L) 2552.2569
AIC 5130.5139
AICC (F-corrected-AIC) 5132.0562
Hannan–Quinn 5148.9386
BIC 5176.2928
Table 4.6 Average absolute percentage error in forecasts and goodness of fit for the estimated
ARIMA(0,1,1)(0,1,1)12 model
Average absolute percentage error in within-sample forecasts:
Last year: 3.27 Last 1 year: 6.93
Last 2 years: 2.09 Last 3 years: 4.10
with constant mean and variance. However, to verify this assumption in more
detail, several statistics are computed on these residuals. Furthermore, the normality
assumption on the residual distribution is not rejected.
110 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
4.6.4 Decomposition
The Decomposition part includes tables with results from consecutive iterations
of the Census II-X11 filters and quality measures. As discussed in Sects. 4.4
and 4.5, to estimate the different components of the series, while taking into account
the possible presence of extreme observations, X12ARIMA proceeds iteratively
through four processing stages, denoted as A, B, C, and D, plus two stages, E,
and Quality measures, that propose statistics and charts and are not part of the
decomposition per se. Stage A has been widely discussed in the previous section.
On stages B, C, and D the basic algorithm is used. Specifically,
B. First automatic correction of the series. This stage consists in a first estimation
and downweighting of the extreme observations and a first estimation of the
working day effects. This stage is performed by applying the basic algorithm
detailed earlier. These operations lead to Table B20, adjustment values for
extreme observations, used to correct the unadjusted series and result in the
series from Table C1.
C. Second automatic correction of the series. Still applying basic algorithm once
again, this part leads to a more precise estimation of replacement values of the
extreme observations (Table C20). The series, finally “cleaned up”, is shown in
Table D1 of the printouts.
D. Seasonal adjustment. This part, at which the basic algorithm is applied for the
last time, is that of the seasonal adjustment per se, as it leads to final estimates:
• of the seasonal component (Table D10);
• of the seasonally adjusted series (Table D11);
• of the trend-cycle component (Table D12); and
• of the irregular component (Table D13).
X12ARIMA selects the filters automatically, taking into account the global
moving seasonality ratio, which is computed on preliminary estimates of the
irregular component and of the seasonal. For the NODG series, a 3 5 m.a.
has been selected to estimate the seasonal component, whereas the trend-cycle
has been estimated using a 13-term Henderson filter, being the signal-to-noise
ratio equal to 2.47. The seasonally adjusted NODG series is shown in Fig. 4.3
together with the original series, and the irregular component in Fig. 4.4.
E. Components modified for large extreme values. Part E includes: (1) components
modified for large extreme values, (2) a comparison between the annual totals of
the raw time series and seasonally adjusted time series, (3) changes in the final
seasonally adjusted series, (4) changes in the final trend-cycle, and (5) robust
estimation of the final seasonally adjusted series. The results from part E are
used in the subsequent part to calculate the quality measures.
F. Quality measures. The final part contains statistics for judging the quality of
the seasonal adjustment, as described in Sects. 4.4 and 4.5. The M statistics
are used to assess the quality of the seasonal adjustment. They account for the
relative contribution of the various components (irregular, trend-cycle, seasonal,
4.6 Illustrative Example: X12ARIMA Seasonal Adjustment of the US NODG. . . 111
NODG
Seasonally adjusted series
200000
100000
preliminary factors, trading days, and holidays effects) to the variance of the
stationary part of the original time series. These statistics vary between 0 and 3,
but only values smaller than 1 are acceptable. Table 4.7 illustrates these statistics
for the components estimated in the X12ARIMA seasonal adjustment of the
NODG series. It can be noticed that all the measures range from 0 to 3, indicating
a good seasonal adjustment for the series.
112
Table 4.7 Diagnostic checking on the residuals of the estimated ARIMA(0,1,1)(0,1,1)12 model
F3. Monitoring and quality assessment statistics
The measures below are between 0 and 3; acceptance region from 0 to 1.
1. The relative contribution of the irregular over 3 months span (from Table F2.B). M1 D 0.827
2. The relative contribution of the irregular component to the stationary portion of the variance (from Table F 2.F). M2 D 0.939
3. The amount of month-to-month change in the irregular component as compared to the amount of month-to-month
change in the trend-cycle (from Table F2.H). M3 D 0.736
4. The amount of autocorrelation in the irregular as described by the average duration of run (Table F 2.D). M4 D 0.285
5. The number of months it takes the change in the trend-cycle to surpass the amount of change in the irregular (from Table F 2.E). M5 D 0.625
6. The amount of year-to-year change in the irregular as compared to the amount of year-to-year change in the seasonal (from Table F 2.H). M6 D 0.295
7. The amount of moving seasonality present relative to the amount of stable seasonality (from Table F 2.I). M7 D 0.195
8. The size of the fluctuations in the seasonal component throughout the whole series. M8 D 0.518
9. The average linear movement in the seasonal component throughout the whole series. M9 D 0.236
10. Same as 8, calculated for recent years only. M10 D 0.454
11. Same as 9, calculated for recent years only. M11 D 0.332
14M1 C 15M2 C 10M3 C 8M4 C 11M5 C 10M6 C 32M7 C 0M8 C 0M9 C 0M10 C 0M11
QD :
100
The model has satisfactory quality if the Q statistic is less than 1. In this regard,
it can be noticed that the X12ARIMA seasonal adjustment of the NODG series
is satisfactory.
References
1. Bell, W. H., & Hillmer, S. C. (1984). Issues involved with the seasonal adjustment of economic
time series. Journal of Business and Economic Statistics, 2, 291–320.
2. Bobbit, L., & Otto, M. C. (1990). Effects of forecasts on the revisions of seasonally
adjusted values using the X11 seasonal adjustment procedure. In Proceedings of the American
Statistical Association Business and Economic Statistics Session (pp. 449–453).
3. Box, G. E. P., & Jenkins, G. M. (1970). Time series analysis: Forecasting and control. Holden-
Day: San Francisco (2nd ed., 1976).
4. Box, G. E. P., & Tiao, G. C. (1975). Intervention analysis with applications to economics and
environmental problems. Journal of the American Statistical Association, 70, 70–79.
5. Cholette, P. A. (1979). A comparison and assessment of various adjustment methods of sub-
annual series to yearly benchmarks. Research paper. Seasonal Adjustment and Time Series
Staff, Statistics Canada, Ottawa.
6. Dagum, E. B. (1978). Modelling, forecasting and seasonally adjusting economic time series
with the X-11 ARIMA method. The Statistician, 27(3), 203–216.
7. Dagum, E. B. (1988). The X-11 ARIMA/88 seasonal adjustment method-foundations and user’s
manual. Statistics Canada, Ottawa: Time Series Research and Analysis Centre.
8. Dagum, E. B., Chhab, N., & Chiu, K. (1990). Derivation and properties of the X11ARIMA
and census X11 linear filters. Journal of Official Statistics, 12(4), 329–347.
9. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M.C., & Chen B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
10. Findley, D. F., Monsell, B. C., Shulman, H. B., & Pugh, M. G. (1990). Sliding spans diagnostics
for seasonal and related adjustments. Journal of the American Statistical Association, 85, 345–
355.
11. Fischer, B., & Planas, C. (2000). Large scale fitting of regression models with ARIMA errors.
Journal of Official Statistics, 16(2), 173–184.
12. Gomez, V., & Maravall, A. (1996). Program TRAMO and SEATS. Madrid: Bank of Spain.
13. Grudkowska, S. (2015). JDemetra+ Reference manual. Warsaw: National Bank of Poland.
14. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
114 4 Linear Filters Seasonal Adjustment Methods: Census Method II and Its Variants
15. Higginson, J. (1975). An F-test for the presence of moving seasonality when using Census
Method II X-11 variant. Research paper, Seasonal Adjustment and Time Series Staff, Statistics
Canada, Ottawa.
16. Huot, G. (1975). Quadratic minimization adjustment of monthly or quarterly series to annual
totals. Research paper. Seasonal Adjustment and Time Series Staff, Statistics Canada, Ottawa.
17. Huot, G., & De Fontenay, A. (1973). General seasonal adjustment guidelines. Part 1, Statistics
Canada’s version of the U.S. Bureau of the Census monthly multiplicative and additive X-
11 program. Research paper. Seasonal Adjustment and Time Series Staff, Statistics Canada,
Ottawa
18. Ladiray, D., & Quenneville, B. (2001). Seasonal adjustment with the X11 method. Lecture
notes in statistics (Vol. 158). New York: Springer.
19. Ljung, G., & Box, G. (1979). The likelihood function of stationary autoregressive-moving
average models. Biometrika, 66, 265–270.
20. Lothian, J., & Morry, M. (1978a). A test for the presence of identifiable seasonality when
using the X-11 program. Research paper. Seasonal Adjustment and Time Series Analysis Staff,
Statistics Canada, Ottawa.
21. Lothian, J., & Morry, M. (1978b). A set of guidelines to assess the reliability of seasonal
adjustment by the X-11 program. Research paper. Seasonal Adjustment and Time Series Staff,
Statistics Canada, Ottawa.
22. Macaulay, F. R. (1931). The smoothing of time series. New York: National Bureau of Economic
Research.
23. Musgrave, J. (1964). A set of end weights to end all end weights. Working paper. U.S. Bureau
of Census, Washington, DC.
24. Newbold, P., & Granger, C.W. J. (1974). Experience with forecasting univariate time series and
the combination of forecasts (with discussion). Journal of the Royal Statistical Society. Series
A, 137, 131–165.
25. Reid, D. J. (1975). A review of short-term projection techniques. In H. A. Gorden (Ed.),
Practical aspects of forecasting (pp. 8–25). London: Operational Research Society.
26. Shiskin, J., Young, A. H. & Musgrave, J. C. (1967). The X-11 variant of the Census Method
II seasonal adjustment program. Technical Paper 15 (revised). US Department of Commerce,
Bureau of the Census, Washington, DC.
27. Wallis, W. A., & Moore, G. H. (1941). A significance test for time series analysis. Journal of
the American Statistical Association, 36, 401–409.
28. Young, A. H. (1968). A linear approximation to the Census and BLS seasonal adjustment
methods. Journal of the American Statistical Association, 63, 445–457.
Chapter 5
Seasonal Adjustment Based on ARIMA Model
Decomposition: TRAMO-SEATS
Peter Burman [4] was the first to develop a seasonal adjustment method based on
ARIMA model decomposition that he named SIGEX (Signal Extraction). Later,
working on the same topic, Hillmer and Tiao [16] developed what these authors
called ARIMA model-based seasonal adjustment, largely discussed in Bell and
Hillmer [2]. An ARIMA model is identified from the observed data and, by
imposing certain restrictions, models for each component are derived. Since the
components are unknown, to obtain a unique solution Hillmer and Tiao proposed
a canonical decomposition which has the property of maximizing the variance of
the irregulars and minimizing the variance of the estimated components. Because
ARIMA model identification and estimation are not robust to outliers or extreme
values and cannot deal with deterministic components such as trading days and
moving holidays, further changes were made by combining dummy variables
regression models with ARIMA models. In this regard, Gomez and Maravall [12]
produced at the Bank of Spain a seasonal adjustment software called TRAMO-
SEATS which is currently applied mainly by European statistical agencies. TRAMO
stands for “Time series Regression with ARIMA noise, Missing observations
and Outliers,” and SEATS for “Signal Extraction in ARIMA Time Series.” First,
TRAMO estimates via regression the deterministic components, which are later
removed from the input data. In a second round, SEATS estimates the seasonal
and trend-cycle components from the ARIMA model fitted to the data where the
deterministic components are removed. SEATS uses the filters derived from the
ARIMA model that describes the stochastic behavior of the linearized time series.
It should be mentioned that Eurostat, in collaboration with the National Bank
of Belgium, the US Bureau of the Census, the Bank of Spain, and the European
Central Bank, has developed an interface of TRAMO-SEATS and X12ARIMA
called DemetraC. In the Bank of Spain and Eurostat websites it is also possible
to find a considerable number of papers relevant to TRAMO-SEATS as well as in
the European Statistical System (ESS) Guidelines. Next sections briefly describe the
supporting methodology of TRAMO-SEATS.
TRAMO (Time Series Regression with ARIMA Noise, Missing Observations, and
Outliers) is a regression method that performs the estimation, forecasting, and
interpolation of missing observations and ARIMA errors, in the presence of possibly
several types of outliers. The ARIMA model can be identified automatically or by
the user. Given the vector of observations y D .y1 ; : : : ; yn /0 , the program fits the
regression model
yt D x0t ˇ C t ; (5.1)
where B is the backshift operator, .B/ and .B/ are finite polynomials in B, and
at is assumed to be normally identically distributed, that is, NID(0, a2 ).
The polynomial .B/ contains the unit roots associated with regular and
seasonal differencing, as well as the polynomial with stationary autoregressive roots
(and complex unit roots, if present). .B/ denotes the invertible moving average
polynomial. In TRAMO, these polynomials assume the following multiplicative
5.1 TRAMO: Time Series Regression with ARIMA Noise, Missing. . . 117
form:
where s denotes the number of observations per year, e.g., 12 for monthly and 4
for quarterly data. The model may contain a constant c related to the mean of the
differenced series .1 B/d .1 Bs /D yt . In practice, this parameter is estimated as
one of the regression parameters in (5.1). Initial estimates of the parameters can be
input by the user, set to the default values, or computed by the program.
The regression variables xt can be given by the user or generated by the program.
In the latter case, the variables correspond to trading day variations, Easter effects,
and outliers.
Trading day variations are linked to the calendar of each year. Traditionally,
six variables have been used to model the trading day effects. These are: (no. of
Mondays/ .no. of Sundays), . . . , (no. of Saturdays/ .no. of Sundays). The sum
of the effects of each day of the week cancels out. Mathematically, this can be
expressed by the requirement that the trading day coefficients ˇj ; j D 1; : : : ; 7,
P P
verify 7jD1 ˇj D 0, which implies ˇ7 D 6jD1 ˇj . There is the possibility
to consider a more parsimonious modeling of the trading day effects using one
variable instead of six. In this case, the days of the week are first divided into
two categories: working days and nonworking days. Then the variable is defined
as (no. of Mondays, Tuesdays, Wednesdays, Thursdays, and Fridays) – 5/2(no. of
Saturdays and Sundays). Again, the motivation P is that it is desirable that the trading
day coefficients ˇj ; j D 1; : : : ; 7, verify 7jD1 ˇj D 0. Since ˇ1 D ˇ2 D D ˇ5
and ˇ6 D ˇ7 , we have 5ˇ1 D 2ˇ6 . The set of trading day variables may also
include the leap-year effect. The leap-year variable assumes value 0 for months that
are not February, 0.25 for Februaries with 28 days, and 0.75 for Februaries with 29
days (leap year). Over a 4-year period, this effect cancels out. Trading day effects
typically characterize series reflecting economic activity, and if significant, most
often adding Easter effect variables improves significantly the results.
Easter effects. This variable models a constant change in the level of daily activity
during the d days before Easter, and its typical value ranges between three and
six. The value of d can also be supplied by the user. The variable has zeros for all
months different from March and April. The value assigned to March is equal to
pM mM , where pM is the proportion of the d days that fall on that month and mM
is the mean value of the proportions of the d days that fall on March over a long
period of time. The value assigned to April is pA mA , where pA and mA are defined
analogously. Usually, a value of mM D mA D 1=2 is a good approximation. Since
pA mA D 1 pM .1 mM / D .pm mM /, a desirable feature is that the sum
of the effects of both months, March and April, cancels out. If forecast of the series
118 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
is desired, the Easter effect variable has to be extended over the forecasting horizon
(2 years when TRAMO is used with SEATS). Furthermore, given that the holidays
vary among countries, it is strongly recommended that each country builds its own
(set of) holiday regression variables, extended over the forecast period.
Outliers reflect the effect of some special, nonregularly repetitive events, such
as implementation of a new regulation, major political or economic changes,
modifications in the way the variable is measured, occurrence of a strike or natural
disaster, etc. Consequently, discordant observations and various types of abrupt
changes are often present in times series data. The location and type of outliers are
“a priori” unknown. TRAMO uses an improved Chen and Liu [6] type procedure
for outlier detection and correction. The effect of an outlier is modeled by yt D
! .B/It .t0 / C t , where .B/ is a quotient of polynomials in B that models the type
of outlier (its effect over time), It .t0 / is an indicative variable of the occurrence of
the outlier, that is,
1 if t D t0
It .t0 / D
0 otherwise;
whereas ! represents the impact of the outlier at time t0 and t is the outlier free
series which follows the model specified in Eq. (5.2). In the automatic detection and
correction, by default, three types of outliers can be considered:
(1) Additive Outlier (AO): .B/ D 1;
(2) Transitory Change (TC): .B/ D 1=.1 ıB/, where, by default, ı D 0:7; and
(3) Level Shift (LS): .B/ D 1=.1 B/.
One can also include a fourth type of outlier, that is Innovational Outlier (IO) for
which .B/ D .B/=.B/ı.B/ that resembles a shock in the innovations at .
The procedure followed by TRAMO consists of:
1. Pretest for the log/level specification. A trimmed range mean regression test is
performed to select whether the original data will be transformed into log or
maintain the level. The former is selected if the slope is positive, whereas the
latter is chosen if it is negative. When the slope is close to zero, a selection
is made based on the Bayesian Information Criterion (BIC) when applied to
the default model using both specifications. A second test consists of direct
comparison of the BICs of the default model in levels and in logs (with a proper
correction).
2. Pretest for trading days and Easter effects. These pretests are made with
regressions using the default model for the noise and, if the model is subsequently
changed, the test is redone. Thus, the output file of TRAMO may say at the
beginning “Trading day is not significant,” but the final model estimated may
contain trading day variables (or vice versa).
3. Automatic detection and correction of outliers. The program has a facility for
detecting outliers and for removing their effects. The outliers can be entered by
the user or they can be automatically identified by the program, using an original
5.1 TRAMO: Time Series Regression with ARIMA Noise, Missing. . . 119
approach based on those of Tsay [29] and Chen and Liu [6]. The outliers are
identified one by one, as proposed by Tsay [29], and multiple regressions are
used, as in [6], to detect spurious outliers. The procedure used to incorporate or
reject outliers is similar to the stepwise regression procedure for selecting the
“best” regression equation. This results in a more robust procedure than that of
Chen and Liu [6], which uses “backward elimination” and may therefore detect
too many outliers in the first step of the procedure.
4. Automatic model selection. The program further performs an automatic identi-
fication of the ARIMA model. This is done in two steps. The first one yields
the nonstationary polynomial .1 B/d .1 Bs /D of model (5.2). This is done by
iterating on a sequence of AR and ARMA models (with constant) which have a
multiplicative structure when the data is seasonal. The procedure is based on the
results of Tiao and Tsay [27], and Tsay [28]. Regular and seasonal differences
are obtained, up to a maximum order of .1 B/2 .1 Bs /.
The second step identifies an ARMA model for the stationary series (modified
for outliers and regression-type effects) following the Hannan–Rissanen proce-
dure [15], with an improvement which consists of using the Kalman filter to
calculate the first residuals in the computation of the estimator of the variance of
the innovations of model (5.2). For the general multiplicative model
the search is made over the range 0 .p; q/ 3, and 0 .P; Q/ 2. This is
done sequentially (for fixed regular polynomials, the seasonal ones are obtained,
and vice versa), and the final orders of the polynomials are chosen according to
the BIC criterion, with some possible constraints aimed at increasing parsimony
and favoring “balanced” models (similar AR and MA orders).
Finally, the program combines the facilities for automatic identification and
correction of outliers and automatic ARIMA model identification just described
so that it has an option for automatic model identification of a nonstationary
series in the presence of outliers.
5. Estimation of the regARIMA model. The basic methodology is described by
Gomez and Maravall [11], and additional documentation is contained in [8, 9,
13]. Estimation of the regression parameters, including outliers, and the missing
observations among the initial values of the series, plus the ARIMA model
parameters, can be made by concentrating the former out of the likelihood, or
by joint estimation. Several algorithms are available for computing the nonlinear
sum of squares to be minimized. When the differenced series can be used, the
algorithm of Morf et al. [26], with a simplification similar to that of Mélard [25],
is used.
For the nonstationary series, it is possible to use the ordinary Kalman filter
(default option), or its square root version (see [1]). The latter is adequate when
numerical difficulties arise. By default, the exact maximum likelihood method
is employed, and the unconditional and conditional least squares methods are
available as options. Nonlinear maximization of the likelihood function and
120 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
computation of the parameter estimates standard errors are made using the
Marquardt’s method and first numerical derivatives.
Estimation of regression parameters is made by using first the Cholesky
decomposition of the inverse error covariance matrix to transform the regression
equation. In particular, the Kalman filter provides an efficient algorithm to
compute the variables in this transformed regression. Then, the resulting least
squares problem is solved by applying the QR algorithm, where the Householder
orthogonal transformation is used. This procedure yields an efficient and numer-
ically stable method to compute Generalized Least Square (GLS) estimators
of the regression parameters, which avoids matrix inversion. For forecasting,
the ordinary Kalman filter or the square root filter options are available. These
algorithms are applied to the original series. See Gomez and Maravall [10] for
a more detailed discussion on how to build initial conditions on a nonstationary
case.
Missing observations can be handled in two equivalent ways. The first one
is an extension to nonstationary models of the skipping approach by Jones [17].
In this case, interpolation of missing values is made by a simplified Fixed Point
Smoother, and yields identical results to those of Kohn and Ansley [19]. The
second one consists of assigning a tentative value and specifying an additive
outlier to each missing observation. If this option is used, the interpolator is the
difference between the tentative value and the estimated regression parameter. If
the ARIMA parameters are the same, it coincides with the interpolator obtained
with the skipping approach. Also, the likelihood can be corrected so that it
coincides with that of the skipping approach. When concentrating the regression
parameters out of the likelihood, the mean squared errors of the forecasts and
interpolations are obtained following the approach of Kohn and Ansley [18].
When some of the initial missing values cannot be estimated (free parameters),
the program detects them, and flags the forecasts or interpolations that depend on
these free parameters. The user can then assign arbitrary values (typically, very
large or very small) to the free parameters and rerun the program. Proceeding
in this way, all parameters of the ARIMA model can be estimated because the
function to minimize does not depend on the free parameters. Moreover, it will be
evident which forecasts and interpolations are affected by these arbitrary values
because they will strongly deviate from the rest of the estimates. However, if
all unknown parameters are jointly estimated, the program may not flag all free
parameters. It may happen that there is convergence to a valid arbitrary set
of solutions, i.e., some linear combinations of the initial missing observations,
including the free parameters, can be estimated.
In brief, regression parameters are initialized by OLS and the ARMA model
parameters are first estimated with two regressions, as in Hannan and Rissanen
[15]. Next, the Kalman filter and the QR algorithm provide new regression
parameter estimates and regression residuals. For each observation, t-tests are
computed for four types of outliers, namely additive, innovational, level shift,
and transitory. If there are outliers whose absolute t-values are greater than a
5.2 SEATS: Signal Extraction in ARIMA Time Series 121
preselected critical level C, the one with the greatest absolute t-value is selected.
Otherwise, the series is free from outlier effects and the algorithm stops.
If outliers are identified, the series is corrected and the ARMA model
parameters are reestimated. Then, a multiple regression is performed using the
Kalman filter and the QR algorithm. If some outliers have absolute t-values below
the critical value C, the one with the lowest absolute value is removed and the
multiple regression is reestimated. In the next step, using the residuals from the
last multiple regression model, t-tests are calculated for the four types of outliers
and for each observation. If there are outliers with absolute t-values greater than
the critical level C, the one with the greatest absolute t value is chosen and the
algorithm goes on to the estimation of the ARMA model to iterate, otherwise it
stops.
6. Diagnostic checks. The main diagnostic tests are on the residuals. The estimated
residuals aO t are analyzed to test the hypothesis that at are independent and
identically normally distributed with zero mean and constant variance a2 .
Besides inspection of the residual graph, the corresponding sample autocorre-
lation function is examined. The lack of residual autocorrelation is tested using
the Ljung–Box test statistics, and skewness and kurtosis tests are applied to
test for normality of the residuals. Specific out-of-sample forecast tests are also
performed to evaluate if forecasts behave in agreement with the model.
7. Optimal forecasts. If the diagnostics are satisfied, the model is used to compute
optimal forecasts for the series, together with their Mean Square Error (MSE).
These are obtained using the Kalman filter applied to the original series.
TRAMO is to be used as a preadjustment process to eliminate all the determinis-
tic components such that the residual is a linearized series that can be modeled with
an ARIMA process. The latter is decomposed by SEATS in stochastic trend-cycle,
seasonality, and irregulars. Both programs can handle routine applications to a large
number of series and provide a complete model-based solution to the problems of
forecasting, interpolation, and signal extraction for nonstationary time series.
SEATS stands for “Signal Extraction in ARIMA Time Series.” It belongs to the
class of procedures based on ARIMA models for the decomposition of time series
into unobserved components. The method was mainly developed for economic
series, and the corresponding computer program is based on a seasonal adjustment
method developed by Burman [4] at the Bank of England. In particular, the SEATS
procedure consists of the following steps.
1. ARIMA model estimation. SEATS starts by fitting an ARIMA model to a series
not affected by deterministic components, such as trading day variations, moving
holidays, and outliers. Let yt denote this linearized series, and consider an
additive decomposition model (multiplicative if applied to the log transformation
122 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
of yt ), such that
zt D .1 B/d .1 Bs /D yt ; (5.3)
represent the “differenced” series. The model for the differenced linearized series
zt can be written as
where c is equal to .B/Ny, being yN the mean of the linearized series yt . In words,
the model that SEATS assumes is that of a linear time series with Gaussian
innovations. When used with TRAMO, estimation of the ARIMA model is made
by the exact maximum likelihood method described in Gomez and Maravall
[11]. When used by itself, SEATS applies the quasi-maximum likelihood method
described by Burman [4]. In both cases, a (faster) least squares algorithm is also
available. SEATS contains an option where the automatic model identification
procedure is replaced by a faster one. This simplified procedure starts directly by
fitting the default model and, if found inappropriate, tries a sequence of models
according to the results.
The program starts with the estimation of the ARIMA model. The (inverse)
roots of the AR and MA polynomials are always constrained to remain in or
inside the unit circle. When the modulus of a root converges within a preset
interval around 1 (by default (0.98,1]), the program automatically fixes the root.
If it is an AR root, the modulus is made 1. If it is an MA root, it is fixed to
the lower limit. This simple feature makes the program very robust to over-
and underdifferencing. In the standard case in which SEATS and TRAMO are
used jointly, SEATS controls the AR and MA roots mentioned above, uses the
ARIMA model to filter the linearized series, obtains in this way new residuals,
and produces a detailed analysis of them.
2. Derivation of the ARIMA models for each component. The program proceeds
by decomposing the series that follows model (5.6) into several components.
The decomposition can be multiplicative or additive. Next, we shall discuss the
5.2 SEATS: Signal Extraction in ARIMA Time Series 123
additive model, since the multiplicative relation can be taken care with the log
transformation of the data. That is,
yt D Tt C Ct C St C It ; (5.7)
where Tt denotes the trend component, Ct the cycle, St represents the seasonal
component, and It the irregulars.
The decomposition is done in the frequency domain. The spectrum (or
pseudospectrum) is partitioned into additive spectra, associated with the different
components which are determined, mostly, from the AR roots of the model. The
trend component represents the long-term evolution of the series and displays
a spectral peak at frequency 0, whereas the seasonal component captures the
spectral peaks at seasonal frequencies (e.g., for monthly data these are 0.524,
1.047, 1.571, 2.094, 2.618, and 3.142). The cyclical component captures periodic
fluctuations with period longer than a year, associated with a spectral peak
for a frequency between 0 and (2=s), and short-term variation associated
with low order MA components and AR roots with small moduli. Finally, the
irregular component captures white noise behavior, and hence has a flat spectrum.
The components are determined and fully derived from the structure of the
(aggregate) ARIMA model (5.6) for the linearized series directly identified from
the data. The program is aimed at monthly or quarterly frequency data and the
maximum number of observations that can be processed is 600.
One important assumption is that of orthogonality among the components,
and each one will be described through an ARIMA model. In order to identify
the components, the canonical decomposition is used, such that the variance of
the irregulars is maximized, whereas the trend, seasonal, and cycle are as stable
as possible.
The canonical condition on the trend, seasonal, and cyclical components
identifies a unique decomposition, from which the ARIMA models for the
components are obtained (including the component innovation variances). This
is achieved as follows.
Let the total AR polynomial .B/ of the ARIMA model (5.6) be factorized as
• the negative real roots of p .B/ D 0 are assigned to the seasonal component
if s ¤ 1, and to the cycle if s D 1;
• the complex roots of p .B/ D 0 are assigned to the seasonal if ! 2 Œ a seasonal
frequency ˙
, being ! the frequency of the root. Otherwise, they are assigned
to the cycle; and
• the positive real roots of ˚P .Bs / D 0, if they are greater than or equal to k
they are assigned to the trend. On the other hand, when they are smaller than
k they are assigned to the cycle.
The parameters k and are automatically set equal to 0.5 and 0.4, respectively,
but they can be entered by the user as well.
The factorization of .B/ can be rewritten as
where T .B/; S .B/, and C .B/ are the AR polynomials with the trend, seasonal,
and cyclical roots, respectively. Let P and Q denote the orders of the polynomials
.B/ and .B/ in (5.6).
Considering the case in which P Q, a polynomial division of the spectrum
(or pseudospectrum) of model (5.6) yields a first decomposition of the type
.B/ .B/
Q
at D a1t C v1 ;
.B/ .B/
.B/
Q Q T .B/ Q S .B/ Q C .B/
a1t D aQ T C aQ S C aQ C ;
.B/ T .B/ t s .B/ t C .B/ t
gj .!/ D gQ j .!/ vj ; j D T; S; C;
X
v D v1 C vj ;
j
the spectrum of the final components are obtained, which give the models for the
components
T .B/Tt D T .B/aTt
S .B/St D S .B/aSt
5.2 SEATS: Signal Extraction in ARIMA Time Series 125
C .B/Ct D C .B/aCt
It WN.0; v/:
All components have balanced models, in the sense that the order of the AR
polynomial equals that of the MA one. On the other hand, when Q > P, the
decomposition proceeds as follows. First a decomposition is performed, whereby
The first component falls under the previous case of P Q, and hence can be
decomposed in the previous way. In general,
ARIMA.P; P 1/ D Tt C St C Ct C It ;
dt D yt yO t ;
dtjn D yt yO tjn ;
rtjn D yO t yO tjn :
for t D 1; : : : ; n, the final estimation error, the preliminary estimation error, and
the revision error in the preliminary estimator. The variances and autocorrelation
functions for dt ; dtjn , rtjn are displayed. The program shows how the variance of
the revision error in the concurrent estimator rnjn decreases as more observations
are added, and hence the time it takes in practice to converge to the final
estimator. Similarly, the program computes the decrease in precision as the
forecast moves away from the concurrent estimate. Finally, the SE of the
estimates of the linearized rates of growth for the concurrent estimate of the
rate and its successive revisions are presented, both for the trend and seasonally
adjusted series. All SE computed are exact given that the ARIMA model for the
observed series is correct. Further details can be found in [22, 23].
TRAMO-SEATS is a DOS program, but its Windows version with some modifi-
cations and additions, called TSW, has been developed by Caporello and Maravall
[5]. TRAMO-SEATS is also available, along with X12ARIMA, in the open source
software Gretl and in the commercial software Eviews. The two seasonal adjustment
procedures are also implemented in the Eurostat software Demetra+, recently
updated in JDemetra+ [14], both developed by the National Bank of Belgium.
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 127
The main feature of Demetra+ and JDemetra+ is to normalize the two different
methods and provide ease of use for national statistics agencies. This is also the aim
declared in the European Statistical System (ESS) Guidelines for Seasonal Adjust-
ment [7] that want to harmonize and provide guidelines for seasonal adjustment.
From the viewpoint of a typical user there is little difference between JDemetra+
and Demetra+. The actual added value of JDemetra+ to the development of the
seasonal adjustment tools lies in the rewriting of the original FORTRAN codes
of X12ARIMA and TRAMO-SEATS in Java. This solution is a crucial factor that
enables the long-term maintenance of the tool, integration of the libraries in the IT
environments of many institutions, and reuse of the modules and algorithms for
other purposes. It runs on operating systems that support the Java VM (Virtual
Machine), such as Microsoft Windows XP SP3/Vista SP1/Windows 7, Ubuntu
9.10, various other Linux distributions, Solaris OS version 11 Express (SPARC and
x86/x64 Platform Edition), and Macintosh OS X 10.6 Intel.
Eurostat uses the TRAMO-SEATS method in the JDemetra+ interface for
seasonal adjustment, and this is the common practice in several European countries.
Hence, in this section, JDemetra+ is applied to perform the seasonal adjustment of
the US New Orders for Durable Goods (NODG) series, observed from February
1992 to December 2013, using the TRAMO-SEATS method. The results of the
seasonal adjustment provided by JDemetra+ are divided into six parts: input, main
results, pre-processing, decomposition, benchmarking, and diagnostics. A detailed
discussion of all the tables, charts, and results produced by the software is not
provided. We focus on the most important outputs for the interpretation of the
seasonal adjustment of the NODG series performed using the TRAMO-SEATS
method.
The first part in the JDemetra+ output, called Input, presents information about
the specification used for seasonal adjustment including most important settings.
Several predefined specifications are available, but also user-defined specifications
can be applied. The predefined specifications refer to automatically test whether a
logarithmic transformation is made of the series or just left as is (log/level), to test for
the presence of working days (working days), or trading days effects (trading days),
as well as Easter (Easter) and leap-year effects. Furthermore, an automatic detection
of several types of outliers (AO, LS, and TC) can be performed by the software. For
some predefined specifications, the airline model ARIMA(0,1,1)(0,1,1)s is used as
default, but also there are cases in which JDemetra+ is required to automatically
identify and estimate the best ARIMA model for the series. In detail, the predefined
TRAMO-SEATS specifications available in JDemetra+ are the following:
• RSA0: level, airline model.
128 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
• RSA1: log/level, outliers detection (AO, TC, and LS), airline model.
• RSA2: log/level, working days, Easter, outliers detection, airline model.
• RSA3: log/level, outliers detection, automatic model identification.
• RSA4: log/level, working days, Easter, outliers detection, automatic model
identification.
• RSA5/RSAfull: log/level, trading days, Easter, outliers detection, automatic
model identification.
The default option in JDemetra+ is RSAfull, in agreement with recommendation
of Maravall and Perez [24], who suggested to start with the automatic option that
pretests for trading day and Easter effects, and for the presence of outliers. Hence,
we select this predefined specification in JDemetra+ for the seasonal adjustment of
the NODG series.
5.3.3 Pre-processing
The Main Results part includes basic information about pre-processing (TRAMO)
and the quality of the decomposition (SEATS). Information about the TRAMO
part of the process includes the estimation span, the number of observations,
transformation (if any), and deterministic effects. On the other hand, the second
part of the Main results summaries the results from the decomposition of the
variance of the white noise innovation for each component extracted by SEATS
from the linearized time series where the regression effects estimated by TRAMO
are removed. The general interest is on a rather stable seasonal component and
hence, when competing models are compared, the preference is given to the model
that minimizes the innovation variance of the seasonal component. For RSA5 and
RSAfull options the trading day and leap-year effects are pretested. If the trading
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 129
Friedman test
Kruskall–Wallis test
Value: 127.39417515948156
Distribution: F with 11 degrees of freedom in the numerator and 251 degrees of freedom in the
denominator
p-value: 0.0000
Seasonality present at the 1 % level
Value: 0.573282639857342
Distribution: F with 20 degrees of freedom in the numerator and 220 degrees of freedom in the
denominator
p-value: 0.9283
No evidence of moving seasonality at the 20 % level
day effects have been detected, the message “Trading days effect (6 regressors)” or
“Trading days effect (7 regressors)” is displayed, depending on whether the leap-
year effect has been detected or not. If the Easter effect is statistically significant,
Easter effect detected is displayed. The additional information, i.e., type, location,
and coefficients of every outlier, can be found in the Pre-processing node.
As shown in Table 5.2, it also includes information about the data, such as the
estimation span, the number of observations, the data transformation, the correction
for leap years, and the number of regression variables (calendar-related variables
and outliers). The notation of the estimation span varies according to the frequency
of the observations, in this case monthly, such that [2-1992:12-2013]. Based on
the predefined (default) specification RSAfull, the message “Series has been log-
transformed” is displayed since the logarithmic transformation has been applied
as result of the test performed by the software. Based on the selected predefined
specification, trading days effects, leap-year and the Easter effects are pretested and
estimated if present. It can be noticed that all these effects, except the leap year,
result significant in the series under analysis. Likelihood statistics related to the
Final model
Likelihood statistics
Number of effective observations = 250
Number of estimated parameters = 13
Loglikelihood = 460.5213592108896
Transformation adjustment = 3029.8684210569045
Adjusted loglikelihood = 2549.347061846015
final model are also shown. The number of effective observations is the number
of observations used to estimate the model, i.e., the number of observations of
regular and seasonal differenced series. The number of estimated parameters is
the sum of the regular and seasonal parameters for both AR and MA, mean effect,
trading/working days effect, outliers, regressors, and standard error of model. In the
pre-processing part, the model is estimated using the exact Maximum Likelihood
Estimation (MLE) method. The standard error of the regression (ML estimate) is
the standard error of the regression from MLE. JDemetra+ displays the maximized
value of the likelihood function, that is used for the selection of the proper ARIMA
model, that is, AIC, AICC, and BIC [14]. The model with the smallest values of
these three criteria is preferred.
5.3.3.1 Regressors
Easter [6]
Coefficients t-Stat p-value
0.0246 2.68 0.0078
Outliers
Coefficients t-Stat p-value
LS (1-2009) 0.2379 8.32 0.0000
LS (10-2008) 0.1788 6.26 0.0000
AO (6-2000) 0.1528 5.22 0.0000
132 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
1.010
1.000
0.990
Fig. 5.1 Trading day estimates of the US New Orders for Durable Goods
p-value associated to the F-test statistic is lower than 5 %). Figure 5.1 displays the
monthly trading days component that ranges from 0.988 to 1.016 over the period
February 1992–December 2013.
The Easter effect is also estimated, and the value of the corresponding coefficient
is reported with its standard error, t-statistics, and the related p-value. As shown in
Table 5.3, there is a significant effect lasting over 6 days around Easter. In particular,
Fig. 5.2 displays that the Easter effect is felt 6 days before Easter and on Easter
Sunday but not after Easter.
JDemetra+ presents also the results of the outliers detection. This table includes
the type of outlier, its date, the value of the coefficient, and corresponding t-statistics
and p-values. For the series under study, the software automatically detects the
presence of three outliers: an additive outlier at June 2000, and two level shift
outliers on October 2008 and January 2009, respectively.
The Arima section shows that, for the log transformed NODG series under study, the
ARIMA(0,1,1)(0,1,1)12 model is chosen and estimated. It should be noticed that we
have reversed the sign of the parameters given by JDemetra+ following the notation
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 133
1.010
1.005
1.000
0.995
0.990
5.3.3.3 Residuals
The examination of residuals from the fitted model is a crucial step of model
validation. JDemetra+ produces several residual diagnostics and graphs for model
checking. They are presented in the Residuals section and its subcategories.
Specifically, Fig. 5.3 illustrates the residuals from the ARIMA(0,1,1)(0,1,1)12 fitted
to the NODG series, after deterministic effects have been removed.
Their pattern seems to be in agreement with the assumption that they have
been generated by a purely random process with constant mean and variance.
However, to verify this assumption in more detail, several tests are computed on
these residuals, and the results are presented both in summary and detail. Summary
statistics report for each test the corresponding p-value. Bold p-values (ranging from
0.1 to 1) mean “good,” italics (p-value ranging from 0.05 to 0.1) means “uncertain,”
and red (p-value ranging from 0 to 0.05) means “bad” (the latter are not present
in our illustrative example). The four tests considered evaluate the normality, no
correlation, randomness, and linearity of the residuals. Table 5.5 shows that, for the
NODG series, all the tests do not reject the null hypothesis.
The Details sections, not reported here, correspond to the appropriate tables from
the Summary diagnostics. Specifically, detailed information are provided for the
0.10
0.05
0.00
-0.05
-0.10
Fig. 5.3 Residuals from the ARIMA(0,1,1)(0,1,1)12 fitted to the NODG series
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 135
Table 5.5 Summary statistics for residuals from the regARIMA(0,1,1)(0,1,1)12 model
Analysis of the residuals
Summary
p-value
Mean 0.9231
Skewness 0.2516
Kurtosis 0.0948
Normality 0.1151
p-value
Ljung-Box(24) 0.3738
Box-Pierce(24) 0.4324
Ljung-Box on seasonality (2) 0.8426
Box-Pierce on seasonality (2) 0.8507
P-value
Runs around the mean: number 0.0992
Runs around the mean: length 1.0000
Up and Down runs: number 1.0000
Up and Down runs: length 1.0000
P-value
Ljung-Box on squared residuals(24) 0.4171
Box-Pierce on squared residuals(24) 0.4742
analysis of the:
1. Distribution of the residuals. For the NODG series, the estimated mean of the
residuals results not significantly different from zero. Furthermore, the normality
assumption is tested in terms of skewness and kurtosis, separately, as well as
jointly using the Doornik–Hansen test which combines both the skewness and
kurtosis tests [14]. In this specific case, the residuals can be considered as
realization of a normal random process.
2. The independence test provided by JDemetra+ is the Ljung–Box test [20].
Denoting with j the sample autocorrelation at rank j of the n residuals, the
136 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
statistics is given by
X
k
j2
LB.k/ D n.n 2/
jD1
.n j/
5.3.4 Decomposition
SEATS processes the linearized series obtained with TRAMO. The decomposition
made by SEATS assumes that all the components in the time series, trend, seasonal,
transitory effects (if present), and irregular, are orthogonal and can be expressed by
ARIMA models. The identification of the components requires that only irregular
components include noise.
peak at frequency 0. By contrast, the seasonal component picks up the spectral peaks
at seasonal frequencies and the irregular component captures a white noise behavior.
In some cases, the ARIMA model chosen by TRAMO is changed by SEATS.
This is done when it is nonadmissible for its decomposition. The decomposition is
admissible when the variance of the seasonal component V.aSt / D S2 ranges in a
given interval ŒNS 2 ; QS 2
(see for details Bell and Hillmer [2]), under the assumption
that the series can be decomposed into the sum of a seasonal and nonseasonal
component, that are independent of each other. Furthermore, the decomposition is
defined to be canonical when this variance S2 is exactly equal to NS 2 .
For the NODG series, the ARIMA(0,1,1)(0,1,1)12 model estimated by TRAMO
is kept by SEATS, and, from that, ARIMA representations of the several com-
ponents are derived as described in Sect. 5.3. In the Decomposition part of the
JDemetra+ output, each model is presented in closed form (i.e., using the backshift
operator B) as illustrated in Table 5.6.
Specifically, the theoretical model for the trend results
Table 5.6 The ARIMA models for original series, seasonally adjusted series and components
Model
D: 1.00000 B B^ 12 + B^ 13
MA: 1.00000 0.523460 B 0.694005 B^ 12 + 0.363284 B^ 13
Sa
D: 1.00000 2.00000 B + B^ 2
MA: 1.00000 1.49745 B + 0.511615 B^ 2
Innovation variance: 0.73560
Trend
D: 1.00000 2.00000 B + B^ 2
MA: 1.00000 + 0.0299450 B 0.970055 B^ 2
Innovation variance: 0.04117
Seasonal
D: 1.00000 + B + B^ 2 + B^ 3 + B^ 4 + B^ 5 + B^ 6 + B^ 7 + B^ 8 + B ^ 9 + B^ 10 + B^ 11
MA: 1.00000 + 1.23274 B + 1.17869 B^ 2 + 1.02437 B^ 3 + 0.790409 B^ 4 + 0.539837 B^ 5
+ 0.294997 B^ 6 + 0.0730214 B^ 7 0.100683 B^ 8 0.252590 B^ 9 0.346686 B^ 10
0.571634 B^ 11
Innovation variance: 0.02285
Irregular
Innovation variance: 0.41628
138 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
Fig. 5.4 The theoretical spectra of the components and seasonally adjusted function
where aSt WN.0; 0:023/, and 11 .B/ is a polynomial of order 11 in the backshift
operator B.
Expressions (5.9) and (5.10) theoretically describe the components and at these
equations JDemetra+ refers with the term Components. The corresponding spectra,
as well as the one of the seasonally adjusted series, are presented in the section
WK analysis in the subsection Components. The sum of the components spectra
should be equal to the spectrum for the linearized time series, presented in the Arima
subsection in the Pre-processing part. Figure 5.4 shows the standard spectra for
trend (green), seasonal (blue), and irregular (grey) components defined in Table 5.6.
The seasonally adjusted series spectrum (yellow) is the sum of trend component
spectrum (green) and the irregular component spectrum (grey).
As discussed in Sect. 5.3, the estimation of the components is performed using
symmetric and bi-infinite Wiener–Kolmogorov (WK) filters that define Minimum
Mean Square Error (MMSE) estimators. As example, in the case of the seasonal
component, the estimator is given by
1
X
SO t D WS .B/yt D wS;j Bj yt t D 1; : : : ; n;
jD1
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 139
S .B/ S .F/
WS .B/ D S2 yt ;
.B/ .F/
This equation defines what JDemetra+ indicates as Final estimator, or simply Esti-
mator, for the seasonal component, from which the variance and autocorrelations
of the (theoretical) MMSE estimator can be obtained. Similarly, we can derive the
theoretical estimator for the trend component defined in Eq. (5.9).
The behavior of the WK filters is generally studied in the frequency domain, such
that the spectrum of the seasonal estimator is derived as
1
X fS ./
WS .ei / D wS;j eij D ;
jD1
fY ./
where fS ./ and fY ./ are the spectral densities of the theoretical seasonal compo-
nent and of the original series, respectively. These spectral densities are generally
unknown, but they are approximated based on the ARIMA specifications given in
Eqs. (5.8) and (5.10), respectively. The transfer functions of the WK estimators of
the other components are similarly derived. They are represented graphically in
the subsection Final estimators in the WK analysis part of the JDemetra+ output.
Beyond the spectra of the theoretical estimators, JDemetra+ also provides the
corresponding squared gain functions and WK filter weights, for a maximum of
71 weights. In particular, the spectra of the final estimators are shown in Fig. 5.5,
and they result similar to those of the theoretical components shown in Fig. 5.4,
although the spectra of the final estimators show spectral zeros at the frequencies
where the corresponding theoretical components have spectral peaks. However, the
estimator adapts to the structure of the analyzed series, that is, the spectral holes in
seasonally adjusted series (yellow line) are wider as more stochastic is the seasonal
component.
140 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
The Stochastic series part presents the table containing the estimated components
produced by SEATS, that is, the seasonally adjusted series, the trend, the seasonal
component, and the irregular component. The seasonally adjusted NODG series is
shown in Fig. 5.6 together with the original series.
The estimated components, denoted by JDemetra+ as Estimates, are derived by
applying the WK filter to the observations extended at both ends with forecasts
and backcasts from the ARIMA model specified for the linearized NODG series.
Alternatively, the spectrum of the component is estimated by applying the spectrum
(transfer function) of the filter to the spectral density of the ARIMA model (5.8) [2].
The Model based tests part analyzes the theoretical distribution of estimators making
use of the estimated components. Specifically, the tests concern the variance and the
autocorrelation function of each estimator, as well as the cross-correlation among
them. Hence, the output is organized in three main sections.
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 141
NODG
Table 5.7 Cross-correlation between the components estimators and the components sample
estimates
Cross-correlation
Estimator Estimate p-value
Trend/seasonal 0.1011 0.1708 0.3719
Trend/irregular 0.2041 0.1173 0.2919
Seasona/irregular 0.0513 0.1131 0.2324
Table 5.8 Variance of the components, of their theoretical WK estimators and of empirical
estimates
Variance
SEATS identifies the components assuming that, except for the irregular, they
are clean of noise. This implies that the variance of the irregular is maximized in
contrast to that of the trend-cycle and seasonal components which are as stable
as possible. In this regard, JDemetra+ compares the variance expected for the
theoretical estimator with the variance empirically computed on the corresponding
estimated component. In the NODG example, the variance of the WK estimator of
the seasonal component is derived using its theoretical specification (5.11). For all
the components, Table 5.8 shows that, for the NODG series, there are no statistically
significant differences between the variances of the theoretical WK estimators and
their estimated values.
Finally, for each component, JDemetra+ exhibits autocorrelations corresponding
to the ARIMA model selected for each component, the corresponding WK estima-
tors and the estimated components. They are calculated from the first lag up to the
seasonal lag.
For each table, the user should check whether the empirical estimates agree with
the model, i.e., if their ACFs are close to those of the WK estimator. Special attention
should be given to first and/or seasonal order autocorrelations. The autocorrelation
functions for the trend and seasonal components of the NODG series are provided
in Tables 5.9 and 5.10, respectively. It can be noticed that for both components there
is evidence of difference in the theoretical and estimated autocorrelations in the first
lags, but not at the seasonal lag.
It has to be noticed that the coefficients of the autocorrelation function of the
irregular component are always null (“Component” column) as shown in Table 5.11
since the theoretical model for irregular component is always a white noise. As
outlined by the p-values column, the estimated irregular component can be thought
as realization of an uncorrelated process.
5.3 Illustrative Example: TRAMO-SEATS Seasonal Adjustment of the US. . . 143
Table 5.9 Autocorrelations of the stationary transformation of trend, the estimators and the
sample estimates
Trend
Lag Component Estimator Estimate p-value
1 0.0005 0.4622 0.3619 0.0078
2 0.4995 0.2542 0,3984 0.0261
3 0.0000 0.3292 0.3102 0.7656
4 0.0000 0.1708 0.0062 0.0525
5 0.0000 0.0865 0.1200 0.0216
6 0.0000 0.0399 0.1204 0.0397
7 0.0000 0.0108 0.0038 0.8717
8 0.0000 0.0137 0.0983 0.2171
9 0.0000 0.0440 0.0451 0.3054
10 0.0000 0.0357 0.0515 0.8594
11 0.0000 0.0723 0.0125 0.4977
12 0.0000 0.1536 0.1440 0.9009
Table 5.10 Autocorrelations of the stationary transformation of the seasonal component, the
estimators and the sample estimates
Seasonal
Lag Component Estimator Estimate p-value
1 0.8654 0.7040 0.5605 0.0188
2 0.6674 0.3311 0.1542 0. 0216
3 0.4467 0.0428 0.1614 0.0374
4 0.2340 0.3588 0.3517 0.8991
5 0.0510 0.5750 0.3705 0.0491
6 0.0886 0.6661 0.4018 0.0452
7 0.1781 0.6233 0.3492 0.0073
8 0.2168 0.4537 0.3495 0.1107
9 0.2087 0.1798 0.1658 0.8765
10 0.1621 0.1604 0.1012 0.6234
11 0.0881 0.5139 0.4329 0.4704
12 0.0000 0.8129 0.7992 0.8198
144 5 Seasonal Adjustment Based on ARIMA Model Decomposition: TRAMO-SEATS
Table 5.11 Autocorrelations of the stationary transformation of the irregular component, the
estimators and the sample estimates
Irregular
Lag Component Estimator Estimate p-value
1 0.0000 0.2382 0.2980 0.2870
2 0.0000 0.1246 0.1918 0.3210
3 0.0000 0.0651 0.0111 0.2715
4 0.0000 0.0338 0.0292 0.9476
5 0.0000 0.0171 0.0672 0.2300
6 0.0000 0.0079 0.0110 0.7568
7 0.0000 0.0022 0.0511 0.4486
8 0.0000 0.0027 0.1111 0.1004
9 0.0000 0.0086 0.0321 0.7347
10 0.0000 0.0184 0.0040 0.7465
11 0.0000 0.0361 0.0732 0.5936
12 0.0000 0.1532 0.1196 0.5818
References
1. Anderson, B., & Moore, J. (1979). Optimal filtering. New Jersey: Prentice Hall.
2. Bell, W. R., & Hillmer, S. C. (1984). Issues involved with the seasonal adjustment of economic
time series. Journal of Business and Economic Statistics, 2, 291–320.
3. Box, G. E. P., & Jenkins, G. M. (1970). Time series analysis: Forecasting and control. San
Francisco: Holden-Day (Second Edition 1976).
4. Burman, J. P. (1980). Seasonal adjustment by signal extraction. Journal of the Royal Statistical
Society A, 143, 321–337.
5. Caporello, G., Maravall, A., & Sanchez, F. J. (2001). Program TSW reference manual. Working
Paper 0112, Servicio de Estudios, Banco de España.
6. Chen, C., & Liu, L. M. (1993). Joint estimation of model parameters and outlier effects in time
series. Journal of the American Statistical Association, 88, 284–297.
7. Eurostat. (2009). ESS guidelines on seasonal adjustment. Luxembourg: Office for Official
Publications of the European Communities.
8. Gomez, V. (1997). Automatic model identification in the presence of missing observations
and outliers. Mimeo. Ministerio de Economia y Hacienda, Direccion General de Analisis y
Programacion Presupuestaria, June 1997
9. Gomez, V., & Maravall. A. (1992). Time series regression with ARIMA noise and missing
observations - Program TRAM. EUI Working Paper Eco No. 92/81, Department of Economics,
European University Institute.
10. Gomez, V., & Maravall, A. (1993). Initializing the Kalman filter with incompletely specified
initial conditions. In G. R. Chen (Ed.), Approximate Kalman filtering. Series on approximation
and decomposition. London: World Scientific Publishing Company.
11. Gomez, V., & Maravall, A. (1994). Estimation, prediction and interpolation for nonstationary
series with the Kalman filter. Journal of the American Statistical Association, 89, 611–624.
12. Gomez, V., & Maravall, A. (1996). Program TRAMO and SEATS: Instructions for users.
Working Paper 9628, Service de Estudios, Banco de Espana.
13. Gomez, V., Maravall, A., & Pena, D. (1999). Missing observations in ARIMA models:
Skipping strategy versus additive outlier approach. Journal of Econometrics, 88, 341–364.
14. Grudkowska, S. (2015). JDemetra+ reference manual. Warsaw: National Bank of Poland.
References 145
15. Hannan, E. J., & Rissanen, J. (1982). Recursive estimation of mixed autoregressive moving
average order. Biometrika, 69, 81–94.
16. Hillmer, S. C., & Tiao, G. C. (1982). An ARIMA-model based approach to seasonal
adjustment. Journal of the American Statistical Association, 77, 63–70.
17. Jones, R. (1980). Maximum likelihood fitting of ARMA models to time series with missing
observations. Technometrics, 22, 389–395.
18. Kohn, R., & Ansley, C. F. (1985). Efficient estimation and prediction in time series regression
models. Biometrika, 72, 694–697.
19. Kohn, R., & Ansley, C. F. (1986). Estimation, prediction and interpolation for ARIMA models
with missing data. Journal of the American Statistical Association, 81, 751–761.
20. Ljung, G., & Box, G. (1979). The likelihood function of stationary autoregressive-moving
average models. Biometrika, 66, 265–270.
21. Maravall, A. (1987). On minimum mean squared error estimation of the noise in unobserved
component models. Journal of Business and Economic Statistics, 5, 115–120.
22. Maravall, A. (1988). The use of ARIMA models in unobserved components estimation.
In W. Barnett, E. Berndt, & H. White (Eds.), Dynamic econometric modeling. Cambridge:
Cambridge University Press.
23. Maravall, A. (1995). Unobserved components in economic time series. In H. Pesaran, P.
Schmidt, & M. Wickens (Eds.), The handbook of applied econometrics (Vol. 1). Oxford: Basil
Blackwell.
24. Maravall, A., & Pérez, D. (2012). Applying and interpreting model-based seasonal adjustment
- The Euro-area industrial production series. In W. R. Bell, S. H. Holan, & T. S. McElroy
(Eds.), Economic time series: Modeling and seasonality. London: Chapman and Hall.
25. Melard, G. (1984). A fast algorithm for the exact likelihood of autoregressive moving average
models. Applied Statistics, 35, 104–114.
26. Morf, M., Sidhu, G. S., & Kailath, T. (1974). Some new algorithms for recursive estimation on
constant, linear, discrete-time systems. IEEE Transactions on Automatic Control, 19, 315–323.
27. Tiao, G. C., & Tsay, R. S. (1983). Consistency properties of least squares estimates of
autoregressive parameters in ARMA Models. The Annals of Statistics, 11, 856–871.
28. Tsay, R. S. (1984). Regression models with time series errors. Journal of the American
Statistical Association, 79, 118–124.
29. Tsay, R. S. (1986). Time series model specification in the presence of outliers. Journal of the
American Statistical Association, 81, 132–141.
Chapter 6
Seasonal Adjustment Based on Structural Time
Series Models
The structural time series approach involves decomposing a series into components
which have a direct interpretation. A structural model, also called unobserved
component (UC) model, consists of a number of stochastic linear processes that
stand for the trend, cycle, seasonality, and remaining stationary dynamic features in
an observed time series. The trend component typically represents the longer term
movement of the series and is often specified as a smooth function of time. The
recurring but persistently changing patterns within the years are captured by the
seasonal component. In economic time series, the cycle component can represent
the dynamic features associated with the business cycle.
The key to handling structural time series models is the state space form with
the state of the system representing the various unobserved components such
as trend, cycle, and seasonality. The estimate of the unobservable state can be
updated by means of a filtering procedure as new observations become available.
Predictions are made by extrapolating these estimated components into the future,
while smoothing algorithms give the best estimate of the state at any point within
the sample. The statistical treatment can therefore be based on the Kalman filter
and its related methods. A detailed discussion of the methodological and technical
yt D Tt C Ct C St C It ; (6.1)
where Tt stands for the trend, Ct for the cycle, St for seasonality, and It is the irregular
component generally assumed to be NID.0; I2 /. All four components are stochastic
and the disturbances driving them are assumed to be mutually uncorrelated. The
definitions of the components are given below, but a full explanation of the
underlying rationale can be found in Harvey [11]. The effectiveness of structural
time series models compared to ARIMA-type models is discussed in Harvey et al.
[12].
The trend component can be specified in many different ways. The most common
specifications are listed below.
Local Level Model The trend component can be simply modeled as a random walk
process, that is
where the disturbance series aTt is normally independently distributed with mean
zero and variance T2 , that is NID.0; T2 /. It is also mutually independent on all other
disturbance series related to yt in (6.1). The initial trend T1 is generally treated as an
unknown coefficient that has to be estimated together with the variance T2 .
6.1 Structural Time Series Models 149
where the disturbance series aTt is as in (6.2). The initial values T1 and ˇ1 are treated
as unknown coefficients. When Tt is given by (6.3), Harvey [10] defines yt D Tt C It
as the local linear trend model. In case
2 D 0, the trend (6.3) reduces to an I.1/
process given by TtC1 D Tt C ˇ1 C aTt , where the drift ˇ1 is fixed. This specification
is referred as a random walk plus drift process. If in addition T2 D 0, the trend
reduces to the deterministic linear trend TtC1 D T1 Cˇ1 t. When T2 D 0 and
2 > 0,
the trend Tt in (6.3) remains an I.2/ process.
Trend with Stationary Drift An extension of the previous model is obtained by
including a stationary stochastic drift component:
with the autoregressive coefficient 0 < 'ˇ < 1, and where the disturbance series
aTt and
t are as in Eqs. (6.3) and (6.4). The restriction for 'ˇ is necessary to have a
stationary process for the drift ˇt . In this case, the initial variable T1 is treated as an
unknown coefficient, while the initial drift is specified as ˇ1 N.0;
2 =1'ˇ2 /. The
stationary drift process for ˇt can be generalized to a higher order autoregressive
process and can include moving average terms. However, in practice it may be
difficult to empirically identify such drift processes without very large data samples.
Higher Order Smooth Trend The local linear trend (6.3) with T2 D 0 is a smooth
I.2/ process. It can be alternatively specified as 2 TtC2 D
t , where the initial
variables T1 and T2 are treated as unknown coefficients. To enforce more smoothness
in the trend component, it can be generalized as k TtCk D
t , where the initial
variables T1 ; : : : ; Tk are treated as unknown coefficients for k D 1; 2; : : :. In the
.k/
usual way, we can specify the higher order smooth trend component by Tt D Tt ,
where
. j/ . j/ . j1/ .0/
TtC1 D Tt C Tt ; Tt D
t ; (6.6)
To capture the business cycle features of a time series, various stochastic specifica-
tions of the cycle component can be considered.
Autoregressive Moving Average Process The cycle component Ct can be formu-
lated as a stationary autoregressive moving average (ARMA) process as
where 'C .B/ is the autoregressive polynomial of order p in the lag operator B with
coefficients 'C;1 ; : : : ; 'C;p and C .B/ is the moving average polynomial of order
q with coefficients C;1 ; : : : ; C;q . The requirement of stationarity applies to the
autoregressive polynomial 'C .B/, and states that the roots j'C .B/j D 0 lie outside
the unit circle. The theoretical autocorrelation function of an ARMA process has
cyclical properties when the roots of j'C .B/j D 0 are within the complex range. It
requires p > 1. In this case, the autocorrelations converge to zero with increasing
lags, but the convergence pattern is cyclical. It implies that the time series itself
has cyclical dynamic properties. Once the autoregressive coefficients are estimated,
it can be established whether the empirical model with Ct as in (6.8) has detected
cyclical dynamics in the time series.
Time-Varying Trigonometric Cycle Another stochastic formulation of the cycle
component is based on a time-varying trigonometric process as follows:
CtC1 cos c sin c Ct a Ct
D C t D 1; : : : ; n; (6.9)
CtC1 sin c cos c Ct aCt
with frequency c associated with the typical length of a business cycle, say between
1.5 and 8 years according to Burns and Mitchell [4]. The factor 0 < < 1 is
6.1 Structural Time Series Models 151
introduced to enforce a stationary process for the stochastic cycle component. The
disturbance and the initial conditions for the cycle variables are given by
a Ct
C1 2
NID 0; C2 I2 NID 0; C2 I2 (6.10)
aCt C1
where the cyclical disturbance series aCt and aCt are serially and mutually inde-
pendent, also with respect to all other disturbance series. The coefficients ; c ,
and C2 are unknown and need to be estimated together with the other parameters.
This stochastic cycle specification is discussed by Harvey [11], that shows that the
process (6.9) to be the same as and ARMA process (6.8) with complex roots where
p D 2 and q D 1.
Smooth Time-Varying Trigonometric Cycle To enforce smoothness on the cycle
.m/
component, Harvey and Trimbur [13] propose the specification Ct D Ct , where
! ! !
. j/ . j/ . j1/
CtC1 cos c sin c Ct Ct
. j/ D . j/ C . j1/ ; (6.11)
CtC1 sin c cos c Ct Ct
for j D m; m 1; : : : ; 1, and
!
.0/
CtC1 a Ct
.0/ D NID.0; C2 I2 /; (6.12)
CtC1 aCt
for t D 1; : : : ; n. The initial conditions for this stationary process are provided by
Trimbur [27].
6.1.3 Seasonality
As discussed in Chap. 2, to account for the seasonal variation in a time series, the
component St is included in model (6.1). Common specifications for the seasonal
component St are provided in the following:
Dummy Seasonal Since the seasonal effects should sum to zero over a year, a basic
model for this component is given by
X
s1
St D Stj C !t t D s; : : : ; n; (6.13)
jD1
where s denotes the number of “seasons” in a year. In words, the seasonal effects are
allowed to change over time by letting their sum over the previous year be equal to a
random disturbance term !t with mean zero and variance !2 . Writing out Eq. (6.13)
152 6 Seasonal Adjustment Based on Structural Time Series Models
When s is even, the sine term disappears for j D s=2, and so the number of
trigonometric parameters, the Sj and Sj , is always .s 1/=2, which is the same as
the number of coefficients in the seasonal dummy formulation. A seasonal pattern
based on (6.15) is the sum of Œs=2
cyclical components, each with D 1, and it may
be allowed to evolve over time in exactly the same way as the cycle was allowed to
move in Eq. (6.10). The model is
Œs=2
X
St D Sj;t ; (6.16)
jD1
with !j;t and !j;t ; j D 1; : : : ; Œs=2
, being white noise processes which are uncorre-
lated with each other and have a common variance !2 . Sj;t
appears as a matter of
construction, and its interpretation is not particularly important. Note that, when s
is even, the component at j D s=2 collapses to
The model (6.1) may provide a good description of the time series, although it
may sometimes be necessary to include additional components. As discussed in the
previous chapters, seasonal economic time series is often affected by trading day
effects and holiday effects which can influence the dynamic behavior of the series.
Hence, a set of explanatory variables need to be included in the model to capture
specific (dynamic) variations in the time series, as well as outliers and breaks.
Therefore, Koopman and Ooms [20] suggest to extend model (6.1) as follows:
The statistical treatment of structural time series models is based on the correspond-
ing state space representation, according to which the observations are assumed to
depend linearly on a state vector that is unobserved and generated by a stochastic
time-varying process. The observations are further assumed to be subject to a
measurement error that is independent on the state vector. The state vector can be
estimated or identified once a sufficient set of observations becomes available.
The state space form provides a unified representation of a wide range of linear
time series models, as discussed by Harvey [11], Kitagawa and Gersch [17], and
Durbin and Koopman [8].
The general linear state space model for a sequence of n observations, y1 ; : : : ; yn ,
is specified as follows:
yt D zTt ˛t C It ; t D 1; : : : ; n (6.18)
˛tC1 D t ˛t C Rt "t : (6.19)
Here the system matrices t and zt do not depend on time. The variances of the
disturbances are unknown and need to be estimated. They are I2 ; T2 ;
2 , and !2 .
For the trend component, the initial variables T1 and ˇ1 are treated as unknown,
whereas for the seasonal component, with s D 4, the unknown coefficients are the
initial variables S1 ; S0 ; S1 . All these coefficients form part of the initial state vector
˛1 . This is the basic structural time series model. A typical application of this basic
model is for the seasonal adjustment of time series, being the seasonally adjusted
series given by yt St D Tt C It for t D 1; : : : ; n.
Local Trend Plus Cycle Model Another model is given by yt D Tt C Ct C It with
Tt specified as in Eq. (6.2) and Ct as in Eq. (6.9). Defining
0 1 0 1 0 1
Tt aTt 1 0 0
˛t D @ Ct A ; "t D @ aCt A ; t D @ 0 cos.c / sin.c / A ; zTt D .1; 1; 0/;
Ct aCt 0 sin.c / cos.c /
0 2 1 0 1
0 0 100
˙ "t D @ 0 C2 .1 2 / 0 A ; Rt D @ 0 1 0 A
2 2 001
0 0 C .1 /
in (6.18) and (6.19), we obtain the following local level plus cycle model:
yt D Tt C Ct C It ; It N.0; I2 /
TtC1 D Tt C t ; t N.0; 2 /
CtC1 D Œcos.c /Ct C sin.c /Ct
C aCt ; aCt NID.0; C2 .1 2 //
CtC1 D Œ sin.c /Ct C cos.c /Ct
C aCt ; aCt NID.0; C2 .1 2 //
for t D 1; : : : ; n, where 0 < < 1 is the damping factor and c is the frequency of
the cycle in radians so that 2=c is the period of the cycle. In case D 1, the cycle
reduces to a fixed sine–cosine wave, but the component is still stochastic since the
initial values C1 and C1 are stochastic variables with mean zero and variance C2 . A
typical application of this model is for the signal extraction of business cycles from
macroeconomic time series.
Given the values of all system matrices and initial conditions a1 and P1 , the
predictive estimator of the state vector ˛tC1 is based on observations y1 ; : : : ; yn . The
Kalman filter [15] computes the minimum mean square linear estimator (MMSLE)
of the state vector ˛tC1 conditional on the observations y1 ; : : : ; yt , denoted by atC1jt ,
together with its mean square error (MSE) matrix, denoted by PtC1jt . atC1jt is
156 6 Seasonal Adjustment Based on Structural Time Series Models
generally called the state prediction estimate, whereas PtC1jt is its state prediction
error variance matrix. Specifically, the one-step ahead prediction error vt D yt
E.yt jy1 ; : : : ; yt1 / and its variance Var.vt / D ft are derived as
such that the innovations have mean zero and are serially independent by construc-
tion, that is, E.vt vs0 / D 0 for t ¤ s, and t; s D 1; : : : ; n. For particular initial values
a1j0 and P1j0 , the MMSLE and corresponding MSE matrix are then determined as
follows:
atC1jt D t atjt1 C kt vt
PtC1 D t Pt T
t C Rt ˙ "t RTt kt MTt
atjt D atjt1 C Ptjt1 zt ft1 vt ; Ptjt D Ptjt1 Ptjt1 zt ft1 zTt Ptjt1 :
It follows that
Formal proofs of the Kalman filter can be found in Anderson and Moore [1],
Harvey [11], and Durbin and Koopman [8].
The Kalman filter can be used to evaluate the likelihood function via the prediction
error decomposition [11, 14, 26], according to which the joint density of y1 ; : : : ; yn
is specified as follows:
Y
n
p.y1 ; : : : ; yn / D p.y1 / p.yt jy1 ; : : : ; yt1 /:
tD2
The predictive density p.yt jy1 ; : : : ; yt1 / is Gaussian with mean E.yt jy1 ; : : : ; yt1 / D
zTt atjt1 and variance V.yt jy1 ; : : : ; yt1 / D zTt Ptjt1 zt CI2 D ft . Hence, for a realized
6.2 Linear State Space Models 157
time series y1 ; : : : ; yn , and denoting with the vector of model parameters, the log-
likelihood function is given by
X
n
`./ D log p.y1 ; : : : ; yn / D log p.yt jy1 ; : : : ; yt1 / (6.20)
tD1
1X 1 X 2 1
n n
n
D log.2/ log j ft j v f :
2 2 tD1 2 tD1 t t
The one-step ahead prediction errors vt and their variances ft are computed by the
Kalman filter for a given value of the parameter vector .
The log likelihood (6.20) has not closed form and has to be numerically
maximized with respect to . This can be done using a numerical quasi-Newton
method, such as the Broyden–Fletcher–Goldfarb–Shanno (BFGS) method. It is
generally regarded as computationally efficient in terms of convergence speed and
numerical stability [24]. This iterative optimization method is based on information
from the gradient (or score). Analytical and computationally fast methods for
computing the score for a current value of in a state space analysis are developed
by Koopman and Shepard [21].
An alternative method for maximum likelihood estimation is the EM-algorithm.
It is based on the joint density p.y1 ; : : : ; yn ; ˛1 ; : : : ; ˛n / and consists of two step.
The expectation (E) step takes the expectation of the components of the joint
density conditional on y1 ; : : : ; yn . Hence, in the E-step, the estimated state space
vector is evaluated using a smoothing algorithm related to the Kalman filter. In the
maximization (M) step the resulting expression of the joint density is maximized
with respect to . It is usually done analytically, and is simpler than maximizing the
full likelihood function directly. Given the updated estimate of from the M-step,
the algorithm returns to the E-step and evaluate the smoothed estimates based on
the new parameter estimates. This iterative procedure converges to the maximum
likelihood estimate of . Details on the properties of the EM algorithm for state
space models can be found in Koopman [18].
The nonstationary trend and seasonal components, as discussed in Sect. 6.2, rely
on initial variables that are treated as fixed unknown coefficients. A straightforward
approach to the estimation of ˛1 is to estimate it jointly with by the method of
maximum likelihood as previously discussed. However, numerical problems may
arise when the likelihood function is maximized with respect to a parameter vector
of high dimension. Several methods have been used in the literature, and most of
them can be embedded in a unified treatment for the initialization of the Kalman
filter with respect to the initial variables, and we refer the reader to Ansley and Kohn
[2], de Jong [6], and Koopman [18] for a detailed discussion. The latter references
also provide a detailed treatment in presence of stationary variables, where in this
case the initial conditions for ˛1 can be obtained from the theoretical autocovariance
function.
158 6 Seasonal Adjustment Based on Structural Time Series Models
The assumptions underlying model (6.1), and all the component specifications given
in Sect. 6.2, are that all disturbances related to the unobserved components are
normally distributed and serially and mutually independent with constant variances.
Under these assumptions the standardized one-step ahead prediction errors (or
prediction residuals) are given by
vt
et D p ; t D 1; : : : ; n;
ft
being also normally distributed and serially independent with unit variance.
The Normality assumption is analyzed looking at the first four moments of the
standardized forecast errors, that are given by
1X 1X
n n
m1 D et ; mq D .et m1 /q ; q D 2; 3; 4:
n tD1 n tD1
When the model assumptions are valid, skewness M3 and kurtosis M4 are normally
distributed as [3]
m3 6 m4 24
M3 D q N 0; ; M4 D 2 N 3; :
m32 n m2 n
Standard statistical tests can be used to check whether the observed values of M3
and M4 are consistent with their asymptotic densities.
A simple test for heteroscedasticity is obtained by comparing the sum of squares
of two exclusive subsets of the sample. For example, the statistic
Pn 2
tDnhC1 et
Ph 2
tD1 et
X
q
c2j
Q.q/ D n.n C 2/ ;
jD1
nj
for some positive integer q, that depends on the frequency of the seriesP (generally
q D 24 for monthly series), where cj is the j-th correlation cj D nm1 2 ntDjC1 .et
m1 /.etj m1 /. The statistic Q is distributed as a chi-square variable with q p
6.3 Illustrative Example: Analysis of the US Unemployment Rate for Males 159
The implementation and estimation of the structural time series models discussed in
the previous sections are made simpler by the availability of a specific software
called STAMP (Structural Time series Analyzer, Modeller, and Predictor). It is
a graphical user interface (GUI)-based package designed to model and forecast
time series using the structural time series approach developed by Koopman et al.
[19]. STAMP runs on Windows, Macintosh, and Linux operating systems as part
of OxMetrics, a software system for (econometric) data analysis and forecasting
developed by Doornik [7]. It provides a powerful, flexible, easy to use, and up-to-
date environment to perform state space analysis of time series data.
STAMP enables a large variety of models to be analyzed, including all the
structural time series models discussed in the previous sections. The software also
provides an extensive array of test statistics, residual, and auxiliary residual graphics
after estimation for assessing the goodness of the model fit. Furthermore, OxMetrics
integrates STAMP into a larger class of related models, and provides a broad suite
of graphics, data transformation, and arithmetic options.
STAMP default model is the basic structural model discussed in Sect. 6.3 which
includes a stochastic level and slope specified through Eqs. (6.3) and (6.4), as well
as a stochastic seasonal component given in (6.13). By default, STAMP does not
allow for the treatment of trading day and holiday effects. Hence, it is appropriate
for the analysis of stock series. In this regard, we analyze the monthly series of US
Unemployment Rate for Males (16 years and over) observed from January 1992
to December 2013. In general, we recommend to select the automatic detection of
outliers in the Select components window among the several possible options. Once
the estimation is complete, some basic information appears in the Results window
as illustrated in Table 6.1.
The “Estimation report” tells us that the convergence was very strong. Maximum
likelihood estimation has been carried out by numerical optimization.
The “Diagnostic summary report” provides some of the basic diagnostic and
goodness of fit statistics discussed in Sect. 6.2.3. In Table 6.1, T denotes the sample
size (indicated by n in our notation), p the number of model parameters to estimate,
and N is the number of analyzed time series. The Ljung–Box test for residual serial
correlation, Q.24; 21/, is based on q D 24 residual autocorrelations and should
be tested against a chi-square distribution with q p D 21 degrees of freedom.
Based on the observed value of Q equal to 24.628, we can conclude that there is no
serial correlations in the residuals, a strong indication that the model is adequately
capturing the dynamic structure of the series. The classical Durbin–Watson (DW)
160 6 Seasonal Adjustment Based on Structural Time Series Models
Summary statistics
UNEMP_RATE
T 264.00
p 3.0000
std.error 0.2042
Normality 0.7291
H(82) 1.2800
DW 2.0335
r(1) -0.0248
q 24.000
r(q) -0.0316
Q(q,q-p) 24.628
Rsˆ2 0.2933
test is also given. Since the DW statistic is approximately equal to 2.1 /,O where O
is the sample first order autocorrelation of the residuals, a value of the DW statistic
equal to 2 indicates no autocorrelation. The value of the DW statistics always lies
between 0 and 4. If the Durbin–Watson statistic is substantially less than 2, there is
evidence of positive serial correlation. As a rough rule of thumb, if Durbin/Watson
is less than 1, there may be cause for alarm. Small values of the DW statistic
indicate that successive error terms are, on average, close in value to one another, or
positively correlated. If the statistic is greater than 2, successive error terms are, on
average, much different in value from one another.
The heteroscedasticity test statistic H.82/, illustrated in Sect. 6.2.3, presents a
small value indicating that the residuals are characterized by constant variance
over time. Finally, the Normality test statistic of Bowman–Shenton discussed in
6.3 Illustrative Example: Analysis of the US Unemployment Rate for Males 161
1.0
UNEMP_RATE-Standardised Residuals +/- 2SE ACF-Standardised Residuals
2
0.5
0.0
0
-0.5
-2
2
0.3
0.2 0
0.1
-2
-3 -2 -1 0 1 2 3 4 -2 -1 0 1 2 3
Sect. 6.2.3, and based on third and fourth moments of the residuals, has a 2
distribution with 2 degrees of freedom when the model is correctly specified. The
5 % critical value is 5.99. The small value of the Normality test statistic together
with the results shown for the other statistics in Table 6.1 clearly indicates that the
basic structural model is adequate for the US Unemployment Rate series. This is
also confirmed by the standardized residuals graphs illustrated in Fig. 6.1.
STAMP also provides further information for the estimated model. As shown in
Table 6.2, in the Print parameter section it reports on the parameters, in this example
estimated by maximum likelihood, where the variances and all the parameters are
organized by each component. Concerning the variances of the disturbances, a zero
parameter estimate would suggest that the corresponding component is fixed. In our
specific case, this is true for the seasonal component, whereas the trend appears to
be stochastic as assumed. The q-ratios in parenthesis are the ratio of each standard
deviation to the standard deviation associated with the largest variance, being the
162 6 Seasonal Adjustment Based on Structural Time Series Models
12.5
UNEMP_RATE
UNEMP_RATE Level+Intv
Level+Intv
10.0
7.5
5.0
1.0
UNEMP_RATE-Seasonal
UNEMP_RATE-Seasonal
0.5
0.0
0.2
UNEMP_RATE-Irregular
UNEMP_RATE-Irregular
0.1
0.0
-0.1
latter related to the level of the trend component. The original series together with
the estimated components are shown in Fig. 6.2.
The final state vector contains information on the values taken by the various
components at the end of the sample. Following Koopman et al. [22], the final state
is the filtered estimate at time n, that is anjn as defined in Sect. 6.3. The square
roots of the diagonal elements of the corresponding MSE matrix, PO njn , are the root
mean square errors (RMSEs) of the corresponding state elements. The t-value is the
estimate of the state divided by the RMSE, and the corresponding p-value is also
reported. The null hypothesis of a zero value is rejected (at 5 % level) for both the
level and slope in the trend component, and for the seasonal dummies, except for
those related to June and December (at 1 % level).
STAMP automatically detected three outliers. In particular, level breaks are
identified in January 1994, December 2008, and April 2010.
References 163
References
1. Anderson, B. D. O., & Moore, J. B. (1979). Optimal filtering. Englewood Cliffs: Prentice-Hall.
2. Ansley, C. F., & Kohn, R. (1985). Estimation, filtering and smoothing in state space models
with incompletely specified initial conditions. Annals of Statistics, 13, 1286–1316.
3. Bowman,p K. O., & Shenton, L. R. (1975). Omnibus test contours for departures from normality
based on b1 and b2 . Biometrika, 62, 243–250.
4. Burns, A., & Mitchell, W. (1946). Measuring business cycles. Working paper. New York:
NBER.
164 6 Seasonal Adjustment Based on Structural Time Series Models
5. Commandeur, J. J. F., & Koopman, S. J. (2007). An introduction to state space time series
analysis. Oxford: Oxford University Press.
6. de Jong, P. (1991). The diffuse Kalman filter. Annals of Statistics, 19, 1073–1083.
7. Doornik, J. A. (2009). An introduction to OxMetrics 6: A software system for data analysis and
forecasting. London: Timberlake Consultants Ltd.
8. Durbin, J., & Koopman, S. J. (2001). Time series analysis by state space methods. Oxford:
Oxford University Press.
9. Gomez, V. (2001). The use of Butterworth filters for trend and cycle estimation in economic
time series. Journal of Business and Economic Statistics, 19, 365–373.
10. Harvey, A. C. (1985). Trends and cycles in macroeconomic time series. Journal of Business
and Economic Statistics, 3, 216–227.
11. Harvey, A. C. (1989). Forecasting, structural time series models and the Kalman filter.
Cambridge: Cambridge University Press.
12. Harvey, A. C., Koopman, S. J., & Penzer, J. (1998). Messy time series: A unified approach. In
T. B. Fomby & R. Carter Hill (Eds.), Advances in econometrics (Vol. 13, pp. 103–143). New
York, NY, USA: JAI Press.
13. Harvey, A. C., & Trimbur, T. M. (2003). General model-based filters for extracting trends and
cycles in economic time series. Review of Economics and Statistics, 85, 244–255.
14. Jones, R. H. (1980). Maximum likelihood fitting of ARIMA models to time series with missing
observations. Technometrics, 22, 389–395.
15. Kalman, R. E. (1960). A new approach to linear filtering and prediction problems. Journal of
Basic Engineering, Transactions ASMA, Series D, 82, 35–45.
16. Kim, C. J., & Nelson, C. R. (1999). State space models with regime switching. Cambridge,
MA: MIT Press.
17. Kitagawa, G., & Gersch, W. (1996). Smoothness priors analysis of time series. New York:
Springer.
18. Koopman, S. J. (1997). Exact initial Kalman filtering and smoothing for non-stationary time
series models. Journal of the American Statistical Association, 92, 1630–1638.
19. Koopman, S. J, Harvey, A. C., Doornik, J. A., & Shepard, N. (2000). STAMP 8.2: Structural
time series analyser, modeller and predictor. London: Timberlake Consultants Ltd.
20. Koopman, S. J., & Ooms, M. (2011). Forecasting economic time series using unobserved
components time series models. In M. Clements & D. F. Hendry (Eds.), Oxford handbook
of economic forecasting. New York: Oxford University Press.
21. Koopman, S. J., & Shephard, N. (1992). Exact score for time series models in state space form.
Biometrika, 79, 823–826.
22. Koopman, S. J., Shephard, N., & Doornik, J. A. (1999). Statistical algorithms for models in
state space form using SsfPack 2.2. Econometrics Journal, 2, 113–166.
23. Ljung, G. M., & Box, G. E. P. (1978). On a measure of lack of fit in time series models.
Biometrika, 66, 67–72.
24. Nocedal, J., & Wright, S. J. (1999). Numerical optimization. New York: Springer.
25. Proietti, T. (2000). Comparing seasonal components for structural time series models. Interna-
tional Journal of Forecasting, 16, 247–260.
26. Schweppe, F. (1965). Evaluation of likelihood functions for Gaussian signals. IEEE Transac-
tions on Information Theory, 11, 61–70.
27. Trimbur, T. M. (2006). Properties of higher-order stochastic cycles. Journal of Time Series
Analysis, 27, 1–17.
Part II
Trend-Cycle Estimation
Chapter 7
Trend-Cycle Estimation
Abstract Economists and statisticians are often interested in the short-term trend
of socioeconomic time series. The short-term trend generally includes cyclical
fluctuations and is referred to as trend-cycle. In recent years, there has been an
increased interest to use trend-cycle estimates. Among other reasons, this interest
originated from major economic and financial changes of global nature which have
introduced more variability in the data, and, consequently, in the seasonally adjusted
numbers. This makes very difficult to determine the direction of the short-term trend,
particularly to assess the presence or the upcoming of a turning point. This chapter
discusses in detail stochastic and deterministic trend-cycle models formulated for
global and local estimation with special reference to the TRAMO-SEATS and
STAMP software as well as the Henderson filters used in the Census II-X11 and its
variants X11/X12ARIMA. It includes also trend-cycle models obtained with locally
weighted averages as the Gaussian kernel, locally weighted regression smoother
(LOESS), and cubic smoothing spline. Many illustrative examples are exhibited.
The trend corresponds to sustained and systematic variations over a long period
of time. It is associated with the structural causes of the phenomenon in question,
for example, population growth, technological progress, new ways of organization,
or capital accumulation. For the majority of socioeconomic time series, the trend
is very important because it dominates the total variation of the series. The
identification of the trend component has always posed a serious statistical problem.
The problem is not of mathematical or analytical complexity, but of conceptual
complexity. This problem exists because the trend and the remaining components
of a time series are latent (not directly observable) variables and, therefore,
assumptions must be made on their behavioral pattern. The trend is generally
thought of as a smooth and slow movement over a long term. The concept of “long”
in this connection is relative and what is identified as trend for a given series span
might well be part of a long cycle once the series is considerably augmented, such as
the Kondratieff economic cycle. Kondratieff [35] estimated the length of this cycle
to be between 47 and 60 years. Often, a long cycle is treated as a trend because the
length of the observed time series is shorter than one complete cycle.
To avoid the complexity of the problem posed by a statistically vague definition,
statisticians have resorted to two simple solutions. One consists of estimating trend
and cyclical fluctuations together calling this combined movement trend-cycle. The
other consists of defining the trend in terms of the series length, denoting it as the
longest no periodic movement. The estimation of the trend-cycle can be done via a
specified model applied to the whole data, called the global trend-cycle, or by fitting
a local polynomial function in such a way that, at any time point, its estimates
only depend on the observations at that point and some specified neighboring
observations.
Local polynomial fitting has a long history in the smoothing of noisy data.
Henderson [30], Whittaker and Robinson[49], and Macauley [37] are some of the
earliest classical references. These authors were very much concerned with the
smoothing properties of linear estimators, being Henderson [30] the first to show
that the smoothing power of a linear filter depends on the shape and values of its
weighting system.
On the other hand, more recent contributions (among others, [8, 18–21, 25, 26,
44, 48]) concentrated on the asymptotic statistical properties of optimally estimated
smoothing parameters. Optimality is being defined in terms of minimizing a given
loss function, usually the mean square error of the estimator or the prediction risk.
In this chapter, we will discuss stochastic and deterministic trend-cycle models
formulated for global and local estimation with special reference to those available
in TRAMO-SEATS and STAMP, as well as the Henderson filters used in the Census
II-X11 and its variants X11/X12ARIMA.
Deterministic global trend models are based on the assumption that the trend or
nonstationary mean of a time series can be approximated closely by simple functions
of time over the entire span of the series.
The most common representation of a deterministic trend is by means of
polynomial functions. The time series from which the trend has to be identified
is assumed to be generated by a nonstationary process where the nonstationarity
results from a deterministic trend. A classical model is the regression or error model
[3] where the observed series is treated as the sum of a systematic part or trend and
a random part. This model can be written as
yt D Tt C et (7.1)
where fet g is a purely random process, that is, fet g i:i:d:.0; e2 / (independent and
identically distributed with expected value 0 and variance, e2 ).
In the case of a polynomial trend-cycle,
Tt D a0 C a1 t C a2 t2 C C ap tp (7.2)
7.1 Deterministic Global Trend Models 169
When b > 0 and c > 1, and so ˛ > 0, model (7.4) represents a trend that increases
at a constant relative rate ˛. For 0 < c < 1, the trend decreases at the rate ˛.
Models (7.3) and (7.4) are solutions of the differential equation
dTt
D ˛.Tt a/; ˛ D log c (7.5)
dt
which specifies the simple assumption of no inhibited growth.
Several economic variables during periods of sustained growth or of rapid
inflation, as well as population growths measured in relative short periods of time,
can be well approximated by trend models (7.3) and (7.4). But in the long run,
socioeconomic and demographic time series are often subject to obstacles that
slow their time path, and if there are no structural changes, their growth tends to
a stationary state. Quetelet [40] made this observation with respect to population
growth and Verhulst [46] seems to have been the first to formalize it by deducing
the logistic model. Adding to Eq. (7.5) an inhibit factor proportional to Tt2 , the
result is
dTt
D ˛Tt ˇTt2 D ˛Tt .1 Tt =k/ (7.6)
dt
k D ˛=ˇ; ˛; ˇ > 0
Model (7.7) belongs to a family of S-shaped curves generated from the differen-
tial equation (see Dagum [10]):
dTt
D Tt .t/˚.Tt =k/; ˚.1/ D 0: (7.8)
dt
Solving Eq. (7.8) for D log c and ˚ D log.Tt =k/, we obtain the Gompertz curve
used to fit mortality table data. That is,
t
Tt D kbc ; b > 0; b ¤ 1; 0 < c < 1; (7.9)
The second major class of global trend models is the one that assumes the trend
to be a stochastic process, most commonly that the series from which the trend
will be identified has been generated by a homogeneous nonstationary stochastic
process [51]. Processes of this kind are nonstationary, but applying a homogeneous
filter, usually the difference filter, we obtain a stationary process in the differences
of a finite order. In empirical applications, the nonstationarity is often present in the
level and/or slope of the series. Hence, the order of the difference is low.
An important class of homogeneous linear nonstationary processes are the
Autoregressive Integrated Moving Average (ARIMA) processes which can be
written as [4]
where B is the backshift operator, such that Bn Tt D Ttn , p .B/ and q .B/
are polynomials in B of order p and q, respectively, satisfying the conditions of
stationarity and invertibility. .1 B/d is the difference operator of order d, and faTt g
is a purely random process. Model (7.10) is also known as an ARIMA process of
order (p; d; q). If p D 0, the process follows an IMA.0; q/ model.
7.2 Stochastic Global Trend Models 171
Three common stochastic trend models are the IMA.1; 1/, IMA.2; 2/, and
ARIMA.2; 1; 2/ which take the following forms
• IMA.1; 1/
or, equivalently,
• IMA.2; 2/
or, equivalently,
• ARIMA.2; 1; 2/
or equivalently
The aTt ’s may be regarded as a series of random shocks that drive the trend and can
be interpreted as measuring the extent to which the random shocks or innovations
incorporate themselves into the subsequent history of the trend. For example, in
model (7.11), the smaller the value of , the more flexible the trend; the higher the
value of , the more rigid the trend (less sensitive to new innovations). For D 1,
model (7.11) reduces to one type of random walk model which has been used mainly
for economic time series, such as stock market price data [23]. In such models, as
time increases the random variables tend to oscillate about their mean value with an
ever increasing amplitude.
172 7 Trend-Cycle Estimation
Stochastic models for the trend have become widespread in economics (see Stock
and Watson [45]), and several modeling approaches have been suggested. Within the
context of linear stochastic processes, there are two general ones. First, the so-called
ARIMA Model-Based (AMB) approach, adopted by TRAMO-SEATS, in which the
model for the trend is derived from the ARIMA model identified for the observed
series (see Burman [5]). The other approach starts by directly specifying a model for
the trend, as implemented in STAMP. Basic references are Harvey [27], and Harvey
and Todd [29].
yt D Tt C et ; (7.16)
where the two components are uncorrelated and each generated by an ARIMA
model as follows
where 'T .B/; 'e .B/; T .B/ and e .B/ are finite polynomials in B. In particular,
'T .B/ D T .B/ıT .B/ and 'e .B/ D e .B/ıe .B/ where both ıT .B/ and ıe .B/ contain
unit autoregressive roots. aTt and aet are orthogonal white noise variables, with
variances T2 and e2 , respectively. It is generally assumed that the roots of the AR
polynomials T .B/ and e .B/ are different, since AR roots for the same frequency
should be associated with the same component. The two polynomials have also to
be prime.
Combining Eqs. (7.16)–(7.18), yt is assumed to follow an ARIMA model of the
type
where '.B/ D .B/ı.B/, being .B/ D T .B/e .B/, and ı.B/ D .1 B/d .1 Bs /D
where s is the number of observations for year. It can be easily derived that the
polynomial .1 Bs / can be factorized as .1 Bs / D .1 B/.1 C B C C Bs1 /,
where .1 C B C C Bs1 / denotes the annual aggregation operator. .B/at is a
Moving Average (MA) process such that
with at WN.0; a2 /. It has to be noticed that in order to allow model (7.19) to be
invertible it is assumed that T .B/ and e .B/ do not have common unit root.
In the AMB approach '.B/, .B/ and a2 are assumed known, and the parameters
of 'T .B/, 'e .B/, T .B/ and e .B/ as well as the variances T2 and e2 have to be
derived from them. This is done in the frequency domain. Based on Eq. (7.16),
the pseudospectrum of yt is partitioned into additive spectra associated with the
terms in the RHS of Eq. (7.16). Identification of a unique decomposition is achieved
by imposing the canonical condition that the minimum of the trend component
pseudospectrum be zero (see Burman [5]).
The Wiener–Kolmogorov (WK) filter for estimating Tt is given by the ratio
of the Tt and yt pseudospectra, and yields the Minimum Mean Square Error
(MMSE) estimator (also conditional mean) of Tt given yt . The filter is centered,
symmetric, and convergent. Its derivation requires an infinite realization of yt in
the direction of the past and of the future. Hence, to apply the filter to a finite
realization, model (7.19) is used to extend yt with forecasts and backcasts. As new
observations become available, forecasts will be updated and eventually replaced
by observations. As a consequence, the estimator of Tt near the end of the series is
preliminary and will be revised. The duration of the revision process of a preliminary
estimator depends on the ARIMA model identified for the series. The spectral
factorization provides the time domain expression of the components given in
Eqs. (7.17) and (7.18).
Letting F denote the forward shift operator (F D B1 ) and replacing the ratio
of pseudospectra by the ratio of Autocovariance Generating Functions (ACGF), the
time domain expression of the WK filter is
where the order q of the polynomial .B/ is equal to max.qT C s 1; qe C 2/, being
qT and qe the orders of the polynomials T .B/ and e .B/, respectively.
Following Maravall [38], necessary and sufficient conditions for the decomposi-
tion (7.20) to be underidentified are that qT > 1 and qe > s 2.
The AMB decomposition restricts the order of the AR and MA polynomial in
model (7.17) for the trend to be of the same order. Hence, the trend model becomes
an IMA.2; 2/ defined in Eq. (7.13), that is,
which contains two parameters 1 and T2 . The number of the unknowns in the
system of covariance equations is .qe C 3/ and the number of equations is greater
than .qe C 3/. The decomposition becomes identifiable and there will be a unique
model (7.24) which will represent the trend component contained in model (7.22).
As an illustrative example, Fig. 7.1 shows the trend estimates, derived using
the TRAMO-SEATS software, of the series of the US New Orders for Durable
Goods (NODG) observed for the period February 1992–December 2013. The
trend estimates have been obtained through canonical decomposition of the
ARIMA.0; 1; 1/.0; 1; 1/12 identified by TRAMO-SEATS to be the stochastic
generating process of the original series.
7.2 Stochastic Global Trend Models 175
NODG series
200000
100000 TRAMO-SEATS trend estimate
Fig. 7.1 The US New Orders for Durable Goods trend-cycle estimates obtained using the
TRAMO-SEATS software
Structural models can be seen as the sum of unobservable ARIMA models specified
to describe the behavior of each time series component separately, and random walk
processes are basic elements in many of these models.
The simplest structural model consists of a trend component Tt plus a random
disturbance term It , where the level Tt , representing the long-term movements in
the series, is allowed to fluctuate up and down according to a random walk process.
In particular, if the irregular component is assumed to be normally distributed, this
local level model is specified as
for t D 1; 2; : : : ; n, where the It ’s and aTt ’s are mutually independent. The model is
incomplete without a specification of the trend at T1 , which generally is assumed
to be normally distributed T1 N.1 ; P1 /, where 1 and P1 are known, and
independent of It and aTt .
When I2 is equal to 0, the observations follow a random walk and the forecast
of future observations is the last observation yT . If T2 is equal to 0, then the level is
constant and the best forecast of future observation is the sample mean. The effect
of aTt is to allow the trend to shift up and down. The larger the variances, the greater
the stochastic movements in the trend. Although it is simple, the local level model
176 7 Trend-Cycle Estimation
provides the basis for the treatment of important series in practice. The properties
of series generated by this model are studied in detail by Harvey [28].
An extension of the local level model is represented by the local linear trend
model which is characterized by a stochastic linear trend whose level and slope vary
over time according to random walk processes as follows
yt D Tt C It ; (7.26)
Tt D Tt1 C ˇt1 C aTt ;
ˇt D ˇt1 C
t ;
NID.0;
2 /;
Equation (7.27) is called the observation equation and Eq. (7.28) is called the state
equation, being the unobserved vector ˛t defined as the state. The vector zt and the
matrices t ; Rt , and " are assumed known as well as I2 , whereas the disturbances
It and "t are independent sequences of normal random variables. The matrix Rt ,
when it is not the identity, is usually a selection matrix whose columns are a subset
of the columns of the identity matrix. It is needed when the dimensionality of ˛t is
7.2 Stochastic Global Trend Models 177
250000
NODGseries
200000
150000
100000 Local linear trend estimate
Fig. 7.2 The US New Orders for Durable Goods trend estimates obtained using the STAMP
software
greater than that of the disturbance vector "t . The structure of model defined through
Eqs. (7.27) and (7.28) is a natural one for representing the behavior of many time
series as a first approximation. The first equation is a standard multivariate linear
regression model whose coefficient vector ˛t varies over time. The development
over time of ˛t is determined by the first order autoregressive vector given in the
second equation. The Markovian nature of the model has a remarkably wide range
of applications to problems in practical time series analysis where the local level
and local linear trend models are special cases. The local level model admits a state
space representation by setting in Eq. (7.27) ˛t D Tt , zt D 1 and in Eq. (7.28)
t D Rt D 1, and ˙ " D T2 .
On the other hand, the local linear trend model defined in Eq. (7.26) can be
rewritten in matrix form as follows
Tt
yt D 1 0 C It ;
ˇt
TtC1 11 Tt aTt
D C ;
ˇtC1 01 ˇt
t
T1 N.1 ; T21 /:
178 7 Trend-Cycle Estimation
For the measurement of the short-term trend, the business cycle component has to
be included in the model. Hence, the trend plus cycle model is specified as follows
yt D Tt C Ct C It ; (7.29)
Tt D Tt1 C ˇt1 C aTt ;
ˇt D ˇt1 C
t ;
being a damping factor on the amplitude such that 0 1, and both aCt and
aCt are NID.0; C2 /. The irregular component It is assumed to follow an NID.0; I2 /
process, and the disturbances in all the three components are taken to be independent
of each other. It can be seen that cycle and trend are simply added to each other.
When is zero, then this model is not identifiable and, since Ct is assumed to be a
stationary process, is strictly less than one. In particular, Ct follows an ARMA(2,1)
process in which both the MA and AR parts are subject to restrictions [27]. The
period of the cycle, which is the time taken to go through its complete sequence of
values, is 2=c . When T2 is equal to zero and C2 is small, the trend plus cycle
model allocates most of the short run variation in the series to the nonstationary
trend component. When both
2 and T2 are small, then the trend becomes much
smoother and most of the variation in the series is allocated to the stationary cyclical
component. Finally, the trend plus cycle model reduces to an AR(1) process when
c is equal to zero or .
Estimation of the parameters I2 , T2 ,
2 , ; c , and C2 can be carried out by
maximum likelihood either in the time domain or in the frequency domain. Once
this has been done, estimates of the trend, cyclical, and irregular components are
obtained from a smoothing algorithm. These calculations may be carried out using
the STAMP software.
Economists and statisticians are often interested in the short-term trend of socioeco-
nomic time series. The short-term trend generally includes cyclical fluctuations and
is referred to as trend-cycle. In recent years, there has been an increased interest
to use trend-cycle estimates or smoothed seasonally adjusted data to facilitate
recession and recovery analysis. Among other reasons, this interest originated from
major economic and financial changes of global nature which have introduced more
7.3 Stochastic Local Trend-Cycle Models 179
variability in the data, and, consequently, in the seasonally adjusted numbers. This
makes very difficult to determine the direction of the short-term trend, particularly
to assess the presence or the upcoming of a turning point. The local polynomial
regression predictors developed by Henderson [30] and LOESS due to Cleveland
[7] are widely applied to estimate the short-term trend of seasonally adjusted
economic indicators. Particularly, the former is available in nonparametric seasonal
adjustment software such as the US Bureau of the Census II-X11 method [43] and
its variants the X11ARIMA ([11] and [12]) and X12ARIMA [22], the latter, in the
STL (Seasonal-Trend decomposition procedure based on LOESS) [6].
The basic assumption is that the input series fyt ; t D 1; : : : ; ng, that is seasonally
adjusted or without seasonality, can be decomposed into the sum of a systematic
component called the signal (or nonstationary mean) TCt , that represents the trend
and cyclical components, usually referred to as trend-cycle for they are estimated
jointly, plus an erratic component, called the noise It , such that
yt D TCt C It : (7.31)
The irregular component It is assumed to be either a white noise, WN.0; I2 /, or,
more generally, to follow a stationary and invertible ARMA process. The trend-
cycle can be represented locally by a polynomial of degree of the time distance j
between yt and the neighboring observations ytCj . Hence, it is possible to find a
local polynomial trend-cycle estimator
where b0 ; b1 ; : : : ; bp are real and aTCt is assumed to be purely random and mutually
uncorrelated with It . The coefficients, b0 ; b1 ; : : : ; bp , can be estimated by ordinary or
b
weighted least squares or by summation formulae. The solution for TCt .0/ provides
the trend-cycle estimate, which equivalently is a weighted average applied in a
moving manner [32], such that
b b X
m
TCt .0/ D TCt D wj ytj ; (7.33)
jDm
b X
m
TCt D wj Bj yt D W.B/yt ; t D 1; 2; : : : ; n; (7.34)
jDm
X
m X
m
wj D 1; jwj D 0; (7.35)
jDm jDm
such that it preserves a constant and a linear trend. On the other hand, W.B/ is a
higher order estimator if
X
m X
m
wj D 1; ji wj D 0 i D 1; 2; : : : ; p: (7.36)
jDm jDm
c ti D t0i bi
TC (7.37)
7.3 Stochastic Local Trend-Cycle Models 181
The Henderson smoothing filters are derived from the graduation theory and used in
Census II-X11 method and its variants X11/ X12ARIMA. The basic principle of the
graduation theory is the combination of operations of differencing and summation
in such a manner that, when differencing above a certain order is ignored, they
will reproduce the functions operated on. The merit of these procedures is that the
smoothed values thus obtained are functions of a large number of observations
whose errors, to a considerable extent, cancel out. These smoothers have the
properties that when fitted to second or third degree parabolas, their output will
fall exactly on those parabolas and when fitted to stochastic data, they will give
smoother results than can be obtained from weights that give the middle point to a
second degree parabola fitted by ordinary least squares.
Recognition of the fact that the smoothness of the resulting graduation depends
directly on the smoothness of the weight diagram led Henderson [30] to develop a
formula which makes the sum of squares of the third differences of the smoothed
series a minimum for any number of terms.
182 7 Trend-Cycle Estimation
Henderson’s starting point was the requirement that the filter should reproduce a
cubic polynomial trend without distortion. Henderson showed that three alternative
smoothing criteria lead to the same formula:
1. minimization of the variance of the third differences of the series defined by the
application of the moving average;
2. minimization of the sum of squares of the third differences of the coefficients of
the moving average formula; and
3. fitting a cubic polynomial by weighted least squares, where the weights are
chosen as to minimize the sum of squares of their third differences.
Kenny and Durbin [33] and Gray and Thomson [24] showed the equivalence of
these three criteria. The problem is one of locally fitting a cubic trend by weighted
least squares to the observations where the weights are chosen to minimize the sum
of squares of their third differences (smoothing criterion). The objective function to
be minimized is
X
m
2
Wj yatCj a0 a1 j a2 j2 a3 j3 ; (7.40)
jDm
c t , Wj D
where the solution for the constant term aO 0 is the smoothed observation TC
Wj , and the filter length is 2m C 1. The solution is a local cubic smoother with
weights
Wj / .m C 1/2 j2 .m C 2/2 j2 .m C 3/2 j2 ; (7.41)
and the weight diagram known as Henderson’s ideal formula is obtained, for a filter
length equal to 2m 3,
315 .m 1/2 j2 .m2 j2 / .m C 1/2 j2 .3m2 16 11j2 /
wj D :
8m.m2 1/.4m2 1/.4m2 9/.4m2 25/
(7.42)
By making m D 8 the wj values are obtained for each j of the 13-term filter (m D 6
for the 9-term Henderson filter, and m D 13 for the 23-term Henderson).
On the contrary, the weights of the usually known as asymmetric Henderson
filters developed by Musgrave [39] are based on the minimization of the mean
squared revision between the final estimates and the preliminary estimates subject
to the constraint that the sum of the weights is equal to one. The assumption made
is that the most recent values of the series (where seasonality has been removed if
present in the original observations) follow a constant linear trend plus an erratic
7.3 Stochastic Local Trend-Cycle Models 183
component (see Laniel [36] and Doherty [17]). The equation used is
0 12
h i2 X
m X
m
.i;m/
E rt D c21 @t wa;j .t j/A C I2 .wj wa;j /2 ; (7.43)
jDa jDm
where w and wa are the weights of the symmetric (central) filter and the asymmetric
filters, respectively; wa;j D 0 for j D m; : : : ; .a C 1/, c1 is the slope of the line
and I2 denotes the noise variance. There is a relation between c1 and I such that
the noise to signal ratio, I=C, is given by
The noise to signal ratio I=C (7.44) determines the length of the Henderson trend-
cycle filter to be applied. Thus, setting t D 0 and m D 6 for the end weights of the
13-term Henderson filter, we have
0 12
.i;6/ 6
X X6
EŒr0
2 4
D @ wa;j A C .wj wa;j /2 : (7.45)
I2 .I=C/2 jDa jD6
Making I=C D 3:5 (the most noisy situation where the 13-term Henderson filter is
applied), Eq. (7.45) gives the same set of end weights of Census II-X11 variant [43].
The end weights for the remaining monthly Henderson filters are calculated using
I=C D 0:99 for the 9-term filter and I=C D 4:5 for the 23-term filter.
The estimated final trend-cycle is obtained by cascade filtering that results from
the convolution of various linear trend and seasonal filters. In fact, if the output from
the filtering operation H is the input to the filtering operation Q, the coefficients of
the cascade filter C result from the convolution of H G. For symmetric filters,
H G D G H, but this is not valid for asymmetric filters. Assuming an input
series yt ; t D 1; 2; : : : ; n, we can define a matrix W D Œwa;j
; a D 1; 2; : : : ; mI j D
1; 2; : : : ; 2m C 1, where each row is a filter and m is the half length of the symmetric
filter. w0; denotes an asymmetric filter where the first m coefficients are zeroes,
whereas wm; denotes the symmetric filter.
Given data up to time n, the m C 1 most recent values of the output (filtered
series) are given by
2mC1
X
c nC1a D
TC wa;j yn2m1Cj a D 1; 2; : : : ; m C 1: (7.46)
jDmaC2
184 7 Trend-Cycle Estimation
The 13-term Henderson filter, which is the most often applied, can then be put in
matrix form as follows:
2 3
0:019 0:028 0:000 0:066 0:147 0:214 0:240 0:214 0:147 0:066 0:000 0:028 0:019
6 7
6 0 0:017 0:025 0:001 0:066 0:147 0:213 0:238 0:212 0:144 0:061 0:006 0:034 7
6 7
6 7
6 0 0 0:011 0:022 0:003 0:067 0:145 0:210 0:235 0:205 0:136 0:050 0:018 7
6 7
6 7
6 0:145 0:208 0:230 0:201 0:131 0:046 7
6 0 0 0 0:009 0:022 0:004 0:066 7:
6 7
6 0 0 0 0 0:016 0:025 0:003 0:068 0:149 0:216 0:241 0:216 0:148 7
6 7
6 7
6 0 0 0 0 0 0:043 0:038 0:002 0:080 0:174 0:254 0:292 0:279 7
4 5
0 0 0 0 0 0 0:092 0:058 0:012 0:120 0:244 0:353 0:421
(7.47)
Kernel type smoothers are locally weighted averages. Given a series of seasonally
adjusted observations and corresponding target points .ytj ; tj /; j D 1; : : : ; n, a kernel
smoothing gives, at time th ; 1 h n, the smoothed estimate
X
n
c th D
TC whj ytj ; (7.48)
jD1
where
th tj
K
whj D Pn th tj
b
(7.49)
jD1 K b
are the weights from a parametric kernel which is a nonnegative symmetric function
that integrated over its domain gives unity, and b ¤ 0 is the smoothing parameter.
Kernel smoothers are local functions since their weighting systems are local. In
fact, for any observed value yth , a weighted average is computed. Each weight is
obtained as a function of the distance between the target point th and the tj ’s close
to th that belong to an interval whose amplitude is established by the smoothing
parameter b, otherwise said bandwidth parameter. Increasing the distance between
th and tj ; j D 1; : : : ; n, the weights assigned to the corresponding observations
decrease till a certain point when they become zero. Such a point depends on the
bandwidth parameter that, in practice, determines the lengths of the smoother, i.e.,
the number of observations near yth that have non-null weight.
In matrix form the action of kernel smoother can be represented by the relation
c D Wb ya
TC (7.50)
7.3 Stochastic Local Trend-Cycle Models 185
where Wb is the smoothing matrix of generic element whj as defined in (7.49) for
h; j D 1; : : : ; n.
Once the smoothing parameter has been selected, the whj ’s corresponding to the
target points falling out of the bandwidth of any th turn out to be null, where the
number of no null weights depends on both the value of the bandwidth parameter
and on the number of decimals chosen for each weight. Notice that as long as the
kernel is a symmetric function, the number of no null weights turns out to be odd.
For instance, for a standard Gaussian kernel function
t t
h j 1 1 th tj 2
K D p exp (7.51)
b 2 2 b
with b = 5 and weights whj taken with three decimals, the smoothing matrix W5 of
no null weights is of size 31 31. For further details, we refer the reader to Dagum
and Luati [15, 16].
In kernel smoothing, the bandwidth parameter is of great importance. Similar to
LOESS, increasing the smoothing parameter from the extreme case of interpolation
(b ! 0) to that of oversmoothing (b ! 1) produces an increase in bias and a
decrease in variance.
we want to minimize
X
n X
n
2
.yt TCt /2 C 2 .1 B/k TCt ; (7.53)
tD1 tDkC1
where .1 B/k denotes the k-th order difference of TCt , e.g., .1 B/TCt D TCt
TCt1 , .1 B/2 TCt D .1 B/..1 B/TCt /, and so on. The smoothing trade-off
parameter must be appropriately chosen.
Following this direction, Schoenberg [42] extended Whittaker smoothing method
to the fitting of a continuous function to observed data, not necessarily evenly
spaced. In this case, the model is
and Wahba [9] who developed the Generalized Cross Validation (GCV) method
currently available in most computer packages. GCV can be obtained from OCV by
rotating the system to a standard coordinate system, doing OCV, and rotating back.
The GCV estimate of is obtained by minimizing
c that is,
where W./ is the influential matrix associated with TC,
2c 3 2 3
TCt1 yt1
c t2 7
6 TC 6 yt 7
6 7 6 27
6 : 7 D W./ 6 : 7 (7.57)
4 :: 5 4 :: 5
c tn
TC ytn
or in compact form
c D W./y:
TC
The trace of W./ can be viewed as the degrees of freedom for the signal and
so, (7.56) can be interpreted as minimizing the standardized sum of squares of the
residuals. Since is usually estimated by cross validation, cubic smoothing splines
are considered nonlinear smoothers. However, Dagum and Capitanio [13] showed
that if is fixed, i.e., D 0 , then W.0 / D W0 has the form
" 1 #
1
W0 D In DT B C DDT D ; (7.58)
0
and
2 1 1 1 1
3
t2 t1
t2 t1
t3 t2 t3 t2
0
6 :: 7
6 0 1 1
t3 t 1
: 0 7
6 t3 t2 t4 t3 7
6 :: ::
2
:: 7
DD6
6 : : :
7
7 (7.60)
6 :: :: :: 7
6 7
4 : : : 5
1 1 1 1
0 tn1 tn2 tn1 tn2
tn tn1 tn tn2
22 3 2 3
1
0 0 1 2 1 0
3 6
6 6 : 7
6
1
6
2
3
1
6
0 7 7 6 0 1 2 :: 0 7
6 :: :: 7 6 7
6 7
BD6
60 : : 77 and D D 6 ::: ::: ::: 7 : (7.61)
6: :: :: 7 6 7
4 :: 6: : : 7
: : 5 4 :: :: :: 5
0 16 23 0 0 1
The theoretical properties of the symmetric and asymmetric linear filters can be
studied by analyzing their frequency response functions. It is defined by
X
m
H.!/ D wj ei!j ; 0 ! 1=2; (7.62)
jDm
where wj are the weights of the filter and ! is the frequency in cycles per unit of
time. In general, the frequency response functions can be expressed in polar form as
follows:
function is zero. For a better interpretation the phase shifts will be here given in
months instead of radians, that is, given by .!/=2! for ! ¤ 0.
The gain function should be interpreted as relating the spectrum of the original
series to the spectrum of the output obtained with a linear time invariant filter. For
example, let yO t be the estimated seasonally adjusted observation for the current
period based on data y1 ; y2 ; : : : ; yn by application of the concurrent linear time-
invariant filter w0;j . Thus, the gain function shown in the below figures relates the
spectrum of fyt ; t D 1; : : : ; ng to the spectrum of yO t , and not to the spectrum of
the complete seasonally adjusted series produced at time n (which includes yt , a
first revision of time t 1, a second revision at time t 2, and so on). Dagum
et al. [14] derived the gain functions of standard, short, and long convolutions
corresponding to the 13-term (H13), 9-term (H9), and 23-term (H23) symmetric
Henderson filters, respectively. These authors showed how cycles of 9 and 10
months periodicity (in the 0.08–0.18 frequency band) are not suppressed by any
of the cascade filters, particularly, those using H13 and H9 filters. In fact, about
90 %, 72 %, and 21 % of the power of these short cycles are left in the output by
the 9-, 13-, and 23-term Henderson filters, respectively. Figure 7.3 shows the gain
functions of four symmetric trend-cycle filters of 13-term each. In the context of
trend-cycle estimation, it is useful to divide the total range of ! 2 Œ0; 0:50
into
two major intervals, one, for the signal, and another for the noise. There are no-
fixed rules on defining the cut-off frequency, but for monthly data the intervals are
1.2
Cubic spline
Gaussian kernel
Henderson filter
LOESS
0.8
G( )
0.4
0.0
Fig. 7.3 Gain functions G.!/; 0 ! 0:5, of several symmetric 13-term filters
190 7 Trend-Cycle Estimation
Cubic spline
0.4
Gaussian kernel
Henderson filter
LOESS
0.0
Fig. 7.4 Gain functions G.!/; 0 ! 0:5, of last point asymmetric filters
7.4 Illustrative Results 191
1.5
Cubic spline
Gaussian kernel
Henderson filter
LOESS
0.5
( )
−0.5
−1.5
Fig. 7.5 Phase shift functions .!/; 0 < ! 0:5, of last point asymmetric filters
We now illustrate with a real series the various trend-cycle values obtained with the
different filters discussed before. To select the appropriate lengths of the Henderson,
LOESS, and Gaussian kernel filters we use the irregular/trend-cycle (I=C) ratio,
whereas for the cubic spline we look at the smoothing parameter estimated by means
of the generalized cross validation criterion. In particular, we consider the US New
Orders for Durable Goods (NODG) observed from February 1992 till December
2013.
Figures 7.6, 7.7, and 7.8 exhibit the various trend-cycle estimates for the US
New Orders for Durable Goods series, where 13-term filters have been chosen for
the Henderson, the LOESS, and the Gaussian kernel and the smoothing parameter
for the cubic spline is 16.438 and the equivalent degrees of freedom is 47.289.
Figure 7.6 shows that the LOESS and Henderson filter give values close to each
other whereas the Gaussian kernel (Fig. 7.7) produces the smoothest trend, and the
cubic spline (Fig. 7.8) gives the most volatile. The difference is more noticeable at
cyclical turning points where the Gaussian kernel cuts them more whereas the cubic
spline cuts them the least.
192 7 Trend-Cycle Estimation
250000
NODG series
Henderson
LOESS
150000
Fig. 7.6 US New Orders for Durable Goods trend-cycle estimates based on LOESS and Hender-
son 13-term filters
250000
NODG series
Gaussian kernel
150000
Fig. 7.7 The US New Orders for Durable Goods trend-cycle estimates based on the 13-term
Gaussian kernel
References 193
250000
NODG series
200000
150000
100000 Cubic smoothing spline
Fig. 7.8 The US New Orders for Durable Goods trend-cycle estimates based on cubic smoothing
spline
References
1. Akaike, H. (1980). Likelihood and the Bayes procedure (with discussion). In J. M. Bernardo,
M. H. De Groot, D. V. Lindley, & A. F. M. Smith (Eds.), Bayesian statistics (pp. 143–203).
Valencia, Spain: University Press.
2. Akaike, H. (1981). Likelihood of a model and information criteria. Journal of Econometrics,
16, 3–14.
3. Anderson, T. W. (1971). The statistical analysis of time series. New York: Wiley.
4. Box, G. E. P., & Jenkins, G. M. (1970). Time series analysis: Forecasting and control. San
Francisco, CA, USA: Holden-Day (Second Edition 1976).
5. Burman, J. P. (1980). Seasonal adjustment by signal extraction. Journal of the Royal Statistical
Society, Series A, 143, 321–337.
6. Cleveland, R. B., Cleveland, W. S., McRae, J. E., & Terpenning, I. (1990). STL: A seasonal
trend decomposition procedure based on LOESS. Journal of Official Statistics, 6(1), 3–33.
7. Cleveland, W. S. (1979). Robust locally regression and smoothing scatterplots. Journal of the
American Statistical Association, 74, 829–836.
8. Cleveland, W. S., & Devlin, S. J. (1988). Locally weighted regression: An approach to
regression analysis by local fitting. Journal of the American Statistical Association, 83(403),
596–610.
9. Craven, P., & Wahba, G. (1979). Smoothing noisy data with spline functions. Numerical
Mathematics, 31, 377–403.
10. Dagum, C. (1985). Analyses of income distribution and inequality by education and sex in
Canada. In R. L. Basmann & G. F. Rhodes Jr. (Eds.), Advances in econometrics (Vol. IV, pp.
167–227). Greenwich, CN: JAI Press.
11. Dagum, E. B. (1980). The X11ARIMA seasonal adjustment method. Ottawa, Canada: Statistics
Canada. Catalogue No. 12-564.
194 7 Trend-Cycle Estimation
12. Dagum, E. B. (1988). The X11ARIMA/88 seasonal adjustment method-foundations and user’s
manual. Ottawa, Canada: Time Series Research and Analysis Centre, Statistics Canada.
13. Dagum, E. B., & Capitanio, A. (1998). Smoothing methods for short-term trend analysis: Cubic
splines and Henderson filters. Statistica, LVIII(1), 5–24.
14. Dagum, E. B., Chhab, N., & Chiu, K. (1996). Derivation and analysis of the X11ARIMA and
Census X11 linear filters. Journal of Official Statistics, 12(4), 329–347.
15. Dagum, E. B., & Luati, A. (2000). Predictive performance of some nonparametric linear and
nonlinear smoothers for noisy data. Statistica, LX(4), 635–654.
16. Dagum, E. B., & Luati, A. (2001). A study of asymmetric and symmetric weights of kernel
smoothers and their spectral properties. Estadistica, Journal of the Inter-American Statistical
Institute, 53(160–161), 215–258.
17. Doherty, M. (2001). The surrogate Henderson filters in X-11. Australian and New Zealand
Journal of Statistics, 43(4), 385–392.
18. Eubank, R. L. (1999). Spline smoothing and nonparametric regression. New York: Marcel
Dekker.
19. Fan, J. (1992). Design-adaptive nonparametric regression. Journal of the American Statistical
Association, 87, 998–1004.
20. Fan, J. (1993). Local linear regression smoothers and their minimax efficiencies. Annals of
Statistics, 21, 196–216.
21. Fan, J., & Gijbels, I. (1996). Local polynomial modelling and its applications. New York:
Chapman and Hall.
22. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., & Chen, B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
23. Granger, C. W., & Morgenstern, O. (1970). Predictability of stock market prices. Lexington,
MA: D.C. Heath.
24. Gray, A., & Thomson, P. (1996). Design of moving-average trend filters using fidelity and
smoothness criteria. In P. M. Robinson & M. Rosenblatt (Eds.), Time series analysis (in
memory of E.J. Hannan), Vol. II. Springer Lecture notes in statistics (Vol. 115, pp. 205–219).
New York: Springer.
25. Green, P. J., & Silverman, B. W. (1994). Nonparametric regression and generalized linear
models. London: Chapman and Hall.
26. Hardle, W. (1990). Applied nonparametric regression. Cambridge: Cambridge University
Press.
27. Harvey, A. C. (1985). Trends and cycles in macroeconomic time series. Journal of Business
and Economic Statistics, 3, 216–227.
28. Harvey, A. C. (1989). Forecasting, structural time series models and the Kalman filter.
Cambridge: Cambridge University Press.
29. Harvey, A. C., & Todd, P. H. J. (1983). Forecasting economic time series with structural and
Box-Jenkins models: A case study. Journal of Business and Economic Statistics, 1, 299–306.
30. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
31. Hillmer, S. C., & Tiao, G. C. (1982). An ARIMA-model-based approach to seasonal adjust-
ment. Journal of the American Statistical Association, 77, 63–70.
32. Kendall, M. G., Stuart, A., & Ord, J. (1983). The advanced theory of statistics (Vol. 3).
New York: Wiley.
33. Kenny, P., & Durbin, J. (1982). Local trend estimation and seasonal adjustment of economic
and social time series. Journal of the Royal Statistical Society A, 145, 1–41.
34. Kitagawa, G., & Gersch, W. (1996). Smoothness priors analysis of time series. Lecture notes
in statistics (Vol. 116). New York: Springer.
35. Kondratieff, N. (1925). Long economic cycles. Voprosy Konyuktury (Vol. 1, No. (1)) (English
translation: The long wave cycle. Richardson and Snyder: New York, 1984).
36. Laniel, N. (1985). Design criteria for the 13-term Henderson end weights. Working paper.
Methodology Branch. Statistics Canada, Ottawa.
References 195
37. Macauley, F. (1931). The smoothing of time series. New York: National Bureau of Economic
Research.
38. Maravall, A. (1993). Stochastic linear trends: Models and estimators. Journal of Econometrics,
56, 5–37.
39. Musgrave, J. (1964). A set of end weights to end all end weights. Working paper. U.S. Bureau
of Census, Washington, DC.
40. Quetelet, A. (1835). Sur l’homme et le développement de ses facultés. Paris: Bachelier.
41. Sbrana, G. (2011). Structural time series models and aggregation: Some analytical results.
Journal of Time Series Analysis, 32, 315–316.
42. Schoenberg, I. (1964). Monosplines and quadrature formulae. In T. Greville (Ed.), Theory and
applications of spline functions. Madison, WI: University of Wisconsin Press.
43. Shiskin, J., Young, A. H., & Musgrave, J. C. (1967). The X-11 variant of the Census Method
II seasonal adjustment program. Technical Paper 15 (revised). US Department of Commerce,
Bureau of the Census, Washington, DC.
44. Simonoff, J. S. (1995). Smoothing methods in statistics. New York: Springer.
45. Stock, J. H., & Watson, M. W. (1988). Variable trends in economic time series. Journal of
Economic Perspectives, 2, 147–174.
46. Verhulst, P. F. (1838). Notice sur la loi que la population suit dans son accroissement. In A.
Quetelet (Ed.), Correspondance Mathematique et Physique (Tome X, pp. 113–121).
47. Wahba, G. (1990). Spline models for observational data. Philadelphia: SIAM.
48. Wand, M., & Jones, M. (1995). Kernel smoothing. Monographs on statistics and applied
probability (Vol. 60). London: Chapman and Hall.
49. Whittaker, E., & Robinson, G. (1924). Calculus of observations: A treasure on numerical
calculations. London: Blackie and Son.
50. Whittaker, E. T. (1923). On a new method of graduation. In Proceedings of the Edinburg
Mathematical Association (Vol. 78, pp. 81–89).
51. Yaglom, A. M. (1962). An introduction to the theory of stationary random functions.
Englewood Cliffs, NJ: Prentice-Hall.
Chapter 8
Further Developments on the Henderson
Trend-Cycle Filter
Abstract The linear filter developed by Henderson is the most widely applied to
estimate the trend-cycle component in seasonal adjustment software such as the US
Bureau of Census II-X11 and its variants, the X11/X12ARIMA. Major studies have
been done on trend-cycle estimation during the last 20 years by making changes
to the Henderson filters. The emphasis has been on determining the direction of
the short-term trend for an early detection of a true turning point. This chapter
introduces in detail three major contributions: (1) a nonlinear trend-cycle estimator
also known as Nonlinear Dagum Filter (NLDF), (2) a Cascade Linear Filter (CLF)
that closely approximates the NLDF, and (3) an approximation to the Henderson
filter via the Reproducing Kernel Hilbert Space (RKHS) methodology.
The linear filter developed by Henderson [22] is the most widely applied to estimate
the trend-cycle component in seasonal adjustment software such as the US Bureau
of Census II-X11 method [30] and its variants, the X11ARIMA [8] and X12ARIMA
[20]. Major studies have been done on trend-cycle estimation during the last 20
years by making changes to the Henderson filters. The emphasis has been made
on determining the direction of the short-term trend for an early detection of a
true turning point. Major changes of global character in the financial and economic
sectors have introduced high levels of variability in time series making difficult to
detect the direction of the short-term trend by simply looking at seasonally adjusted
data and the use of trend-cycle data has been supported.
In 1996, Dagum [9] developed a nonlinear trend-cycle estimator to improve on
the classical 13-term Henderson filter [22]. The Nonlinear Dagum Filter (NLDF)
results from applying the 13-term symmetric Henderson filter (H13) to seasonally
adjusted series where outliers and extreme observations have been replaced and
which have been extended with extrapolations from an ARIMA model [3]. Later
on, in 2009, Dagum and Luati [16] developed a linear approximation to the NLDF
to facilitate a broader application since the new linear approximation does not need
an ARIMA model identification.
From another perspective, in 2008 Dagum and Bianconcini [11] developed an
approximation to the Henderson filter via reproducing kernel methodology that was
shown to produce better results for real time trend-cycle estimation. All three papers
make explicit reference to the Henderson filters applied to monthly time series
but the results can be easily extended to quarterly series. It should be noticed that
following the tradition of statistical agencies, in these papers it is called asymmetric
Henderson filter the one that corresponds to the asymmetric weights developed
by Musgrave [27]. In the next sections, we will discuss the major assumptions
and modifications introduced in each of these studies and refer the reader to the
published papers for more details.
The modified Henderson filter developed by Dagum [9] is nonlinear and basically
consists of:
(a) extending the seasonally adjusted series with ARIMA extrapolated values, and
(b) applying the 13-term Henderson filter to the extended series where extreme
values have been modified using very strict sigma limits.
To facilitate the identification and fitting of simple ARIMA models, Dagum [9]
recommends, at step (a), to modify the input series for the presence of extreme
values using the standard ˙2:5 limits of X11/X12ARIMA and X13. These
computer programs are those that can be used to implement the Henderson filter.
In this way, a simple and very parsimonious ARIMA model, the ARIMA(0,1,1), is
often found to fit a large number of series.
Concerning step (b), it is recommended to use very strict sigma limits, such as
˙0:7 and ˙1:0. The main purpose of the ARIMA extrapolations is to reduce the
size of the revisions of the most recent estimates, whereas that of extreme values
replacement is to reduce the number of unwanted ripples produced by H13. An
unwanted ripple is a 10-month cycle (identified by the presence of high power at
! D 0:10 in the frequency domain) which, due to its periodicity, often leads to
the wrong identification of a true turning point. In fact, it falls in the neighborhood
between the fundamental seasonal frequency and its first harmonic. On the other
hand, a high frequency cycle is generally assumed to be part of the noise pertaining
to the frequency band 0:10 ! < 0:50. The problem of the unwanted ripples is
specific of H13 when applied to seasonally adjusted series.
The NLDF can be formally described in matrix notation as follows. Let y 2 Rn
be the n-dimensional seasonally adjusted time series to be smoothed, which consists
of a nonstationary trend-cycle TC plus an erratic component e, that is,
y D TC C e: (8.1)
8.1 The Nonlinear Dagum Filter (NLDF) 199
It is assumed that the trend-cycle is smooth and can be well estimated by means of
the 13-term Henderson filter applied to y. Hence,
c D Hy;
TC (8.2)
y D Hy C e; (8.3)
or,
where W0 is a zero-one diagonal matrix, being the diagonal element wtt equal to
zero when the corresponding element yt of the vector y is identified as an outlier.
For instance, if in the series y the only extreme value is y2 , then the weight matrix
for the residuals will be
2 3
1 0 0 0
6 0 0 0 0 7
6 7
6 7
W0 D 6 0 0 1 0 7 : (8.6)
6 7
4 5
0 0 0 1
Denoting by
e0 D W0 e (8.7)
the vector of the modified residuals, then the series modified by extreme values with
zero weights becomes
y0 D TC C e0 ; (8.8)
200 8 Further Developments on the Henderson Trend-Cycle Filter
Using (8.9), one year of ARIMA extrapolations is obtained in order to extend the
series modified by extreme values. Denoting with yE0 the extended series, that is,
the n C 12 vector whose first n elements are given by y0 while the last 12 are the
extrapolated ones, in block-matrix notation we have
ŒH C W0 .In H/
y
yE0 D ; (8.10)
y12
and
C;12 y
y D 12 ; (8.12)
y
yE0 becomes
Finally, the NLDF estimates are given by applying a 13-term Henderson filter to
Eq. (8.15), that is,
Given the excellent performance of the NLDF according to the three criteria
mentioned above, Dagum and Luati [16] developed a cascade linear filter that
closely approximates it. The cascading is done via the convolution of several filters
chosen for noise suppression, trend estimation, and extrapolation.
A linear filter offers many advantages over a nonlinear one. First, its application
is direct and hence, does not require knowledge of ARIMA model identification.
202 8 Further Developments on the Henderson Trend-Cycle Filter
Furthermore, linear filtering preserves the crucial additive constraint by which the
trend of an aggregated variable should be equal to the algebraic addition of its
component trends, thus avoiding the selection problem of direct versus indirect
adjustments. Finally, the properties of a linear filter concerning signal passing and
noise suppression can always be compared to those of other linear filters by means
of spectral analysis.
The symmetric filter is the one applied to all central observations. In this case,
the purpose is to offer a linear solution to the unwanted ripples problem. To avoid
the latter, the NLDF largely suppresses the noise in the frequency band between
the fundamental seasonal and first harmonic. In this regard, a cascade linear filter
is derived by double smoothing the residuals obtained from a sequential application
of H13 to the input data. The residuals smoothing is done by the convolution of two
short smoothers, a weighted 5-term and a simple 7-term linear filters. The linear
approximation for the symmetric part of the NLDF is truncated to 13 terms with
weights normalized to add to one.
On the other hand, the asymmetric filter is applied to the last six data points
which are crucial for current analysis. It is obtained by means of the convolution
between the symmetric filter and the linear extrapolation filters for the last six data
points. The extrapolation filters are linearized by fixing both the ARIMA model and
its parameter values. The latter are chosen so as to minimize the size of revisions and
phase shifts. The model was selected among some parsimonious processes found to
fit and extrapolate well a large number of seasonally adjusted series. Such model
is the ARIMA(0,1,1) with D 0:40. A simple linear transformation [15] allows to
apply the asymmetric filter to the first six observations.
The new filter is called the Cascade Linear Filter (CLF) and there is a distinction
between the Symmetric (SLF) and the Asymmetric Linear Filter (ALF).
The smoothing matrix associated with the symmetric linear filter results
H H C M7.0:14/ .In H/ H C M5.0:25/ .In H/ ; (8.17)
where M5.0:25/ is the matrix representative of a 5-term moving average with weights
.0:250; 0:250; 0:000; 0:250; 0:250/, and M7.0:14/ is the matrix representative of a 7-
term filter with all weights equal to 0:143.
5- and 7-term filters are chosen following the standard filters length selected
in Census II-X11 and X11/X12ARIMA software for the replacement of extreme
Seasonal-Irregular (SI) values. In these computer packages, a first iteration is made
by means of a short smoother, a 5-term (weighted) moving average, and a second
one by a 7-term (weighted) average. In this case, 5- and 7-term filters are applied
to the residuals from a first pass of the H13 filter. These two filters have the
good property of suppressing large amounts of power at the frequency ! D 0:10,
8.2 The Cascade Linear Filter 203
Fig. 8.1 Gain function of the symmetric weights of the modified irregulars filter
denoted by SLF, are given by 0:027; 0:007; 0:031; 0:067; 0:136; 0:188; 0:224 .
Table 8.1 Empirical RMSE and the number of false turning points (ftp) for SLF and H13 filters
applied to real time series (mean values standardized by those of H13)
Empirical measures of fitting and smoothing SLF H13
RMSE=RMSEH13 1.05 1
ftp=ftpH13 0.80 1
8.2 The Cascade Linear Filter 205
The smoothing matrix associated with the asymmetric linear filter for the last six
data points is obtained in two steps:
1. A linear extrapolation filter for six data points is applied to the input series. This
filter is represented by a .n C 6/ n matrix A
" #
In
A D ;
O6n12 ˘ 612
where
1
X
. / . h/
j D jC 1 C h j ;
hD1
206 8 Further Developments on the Henderson Trend-Cycle Filter
.1/
for j D 1; 2; : : : and j D j . Strictly speaking, the yt ’s go back to infinite
past, but since the power series is convergent, their dependence on ytj can be
. /
ignored after some time has elapsed. For this reason, j D 0 for j > 12.
Furthermore, to generate one year of ARIMA extrapolation D 12 is fixed. In
terms of signal passing and noise suppression the model chosen is D 0:40.
For an ARIMA(0,1,1) model, the coefficients are j D .1 / j1 , and it can
. /
be shown that j D j , 8. Hence, ˘ 12n is the matrix whose generic row is
(0, . . . , 0, 0.001, 0.002, 0.006, 0.015, 0.038, 0.096, 0.24, 0.6).
Because only six observations are needed to be extrapolated, the centered
12-term filter is truncated and the weights are uniformly normalized to obtain
the 6 12 matrix ˘ 612 given by
–0,023 –0,004 0,004 0,086 0,143 0,187 0,200 0,180 0,130 0,071 0,021 –0,024
0,000 –0,023 –0,004 0,033 0,088 0,148 0,186 0,195 0,160 0,116 0,085 0,017
0,000 0,000 –0,023 –0,005 0,035 0,093 0,146 0,180 0,166 0,155 0,165 0,085
0,000 0,000 0,000 –0,024 –0,003 0,037 0,090 0,141 0,148 0,182 0,255 0,173
0,000 0,000 0,000 0,000 –0,021 –0,006 0,034 0,089 0,114 0,196 0,331 0,264
0,000 0,000 0,000 0,000 0,000 –0,032 –0,009 0,039 0,075 0,200 0,386 0,342
y D SA y
b
The convolution SA produces 12-term asymmetric filters for the last six
observations, that are truncated and uniformly normalized in order to obtain the
following final Asymmetric Linear Filters (ALF) for the last observations:
–0,026 –0,007 0,030 0,065 0,132 0,183 0,219 0,183 0,132 0,066 0,030 –0,006
0,000 –0,026 –0,007 0,030 0,064 0,131 0,182 0,218 0,183 0,132 0,065 0,031
0,000 0,000 –0,025 –0,004 0,034 0,069 0,137 0,187 0,222 0,185 0,131 0,064
0,000 0,000 0,000 –0,020 0,006 0,046 0,083 0,149 0,196 0,226 0,184 0,130
0,000 0,000 0,000 0,000 0,001 0,033 0,075 0,108 0,167 0,205 0,229 0,182
0,000 0,000 0,000 0,000 0,000 0,045 0,076 0,114 0,134 0,182 0,218 0,230
8.2 The Cascade Linear Filter 207
Hence, the asymmetric filters for the last six data points result from the convolution
of:
1. the asymmetric weights of an ARIMA(0,1,1) model with D 0:40,
2. the weights of M7.0:14/ and M5.0:25/ filters repeatedly used for noise suppression,
and
3. the weights of the final linear symmetric filter SLF.
The convergence pattern of the asymmetric filters corresponding to H13 and ALF
is shown in Figs. 8.2 and 8.3, respectively. It is evident that the ALF asymmetric
filters are very close one another, and converge faster to the SLF relative to H13.
The distance of each asymmetric filter with respect to the symmetric one gives an
indication of the size of the revisions due to filter changes, when new observations
are added to the series. Figure 8.4 shows that the gain function of the last point
ALF does not amplify the signal as H13 and suppresses significantly the power
at the frequency ! < 0:10. From the viewpoint of the gain function, the ALF
is superior to H13 concerning the unwanted ripples problem as well as in terms
of faster convergence to the corresponding symmetric filter which implies smaller
revisions.
On the other hand, the phase shift of the last point ALF is much greater (near
2 months at very low frequencies) relative to H13 as exhibited in Fig. 8.5. For the
Fig. 8.2 Gain functions showing the convergence pattern of H13 asymmetric weights to the
symmetric H13
208 8 Further Developments on the Henderson Trend-Cycle Filter
Fig. 8.3 Gain function showing the convergence pattern of the ALF asymmetric weights to the
symmetric SLF
Fig. 8.4 Gain functions of the last point filter of H13 and the last point ALF
remaining asymmetric filters, the differences are much smaller. Nevertheless, the
impact of the phase shift in a given input cannot be studied in isolation of the
corresponding gain function. It is well-known that a small phase shift associated
8.2 The Cascade Linear Filter 209
Fig. 8.5 Phase shifts of the last point filter of H13 and last point ALF
with frequency gain amplifications may produce as poor results as a much larger
phase shift without frequency gain amplifications.
The performance of last and previous to the last point ALF was evaluated empir-
ically using a large sample characterized by different degrees of variability. The
purpose is to reduce the size of the revision of these ALF with respect to the
asymmetric H13 filter. We denote with jDs;l j the mean absolute revision between
the final estimates obtained with the symmetric filter, yO sk , and the last point estimate
obtained with the asymmetric ˇ filter
ˇ for the last point, yO lk , calculated over the whole
ˇ ˇ
sample. We denote with Ds;p the mean absolute error between the previous yO k
p
and the last point yO k . We also calculate the corresponding mean square errors. The
l
results are shown in Table 8.2 standardized by last point H13.It is evident that the
size of revisions for the most recent estimates is smaller for ALF relative to last
point asymmetric H13. This indicates a faster convergence to the symmetric filter,
which was also shown theoretically by means of spectral analysis. Similarly, the
210 8 Further Developments on the Henderson Trend-Cycle Filter
distance between the last and previous to the last filters is smaller for ALF relative
to H13.
In summary, the symmetric and asymmetric weights of the cascade filter are:
• Symmetric weights
0:027 0:007 0:031 0:067 0:136 0:188 0:224 0:188 0:136 0:067 0:031 0:007 0:027
• Asymmetric weights
0:026 0:007 0:030 0:065 0:132 0:183 0:219 0:183 0:132 0:065 0:030 0:006
0:000 0:026 0:007 0:030 0:064 0:131 0:182 0:218 0:183 0:132 0:065 0:031
0:000 0:000 0:025 0:004 0:034 0:069 0:137 0:187 0:222 0:185 0:131 0:064
0:000 0:000 0:000 0:020 0:005 0:046 0:083 0:149 0:196 0:226 0:184 0:130
0:000 0:000 0:000 0:000 0:001 0:033 0:075 0:108 0:167 0:205 0:229 0:182
0:000 0:000 0:000 0:000 0:000 0:045 0:076 0:114 0:134 0:182 0:218 0:230
The study of the properties and limitations of the Henderson filter has been done
in different contexts and attracted the attention of a large number of authors,
among them, Cholette [6], Kenny and Durbin [23], Castles [4], Dagum and Laniel
[13], Cleveland et al. [7], Dagum [9], Gray and Thomson [21], Loader [25],
Dalton and Keogh [17], Ladiray and Quenneville [24], Quenneville et al. [29],
Findley and Martin [19], and Dagum and Luati [15, 16]. However, none of these
studies have approached the Henderson smoother from a RKHS perspective. In
this regard, in 2008 Dagum and Bianconcini [11] introduced a Reproducing Kernel
Hilbert Space (RKHS) representation of the Henderson filter. The Henderson kernel
representation enabled the construction of a hierarchy of kernels with varying
smoothing properties. Furthermore, for each kernel order, the asymmetric filters
could be derived coherently with the corresponding symmetric weights or from
a lower or higher order kernel within the hierarchy, if more appropriate. In the
particular case of the currently applied asymmetric Henderson filters, those obtained
by means of the RKHS, coherent to the symmetric smoother, were shown to have
superior properties from the viewpoint of signal passing, noise suppression, and
revisions. A comparison was made with real life series.
8.3 The Henderson Filter in the Reproducing Hilbert Space (RKHS) 211
Let fyt ; t D 1; 2; : : : ; ng denote the input series. In this study, we work under the
following (basic) specification for the input time series.
Assumption 1 The input time series fyt ; t D 1; 2; : : : ; ng, that is seasonally
adjusted or without seasonality, can be decomposed into the sum of a systematic
component, called the signal (or nonstationary mean or trend-cycle) TCt , plus an
irregular component It , called the noise, such that
yt D TC t C It : (8.19)
The noise It is assumed to be either a white noise, WN.0; I2 /, or, more generally, to
follow a stationary and invertible AutoRegressive Moving Average (ARMA) process.
The trend-cycle can be deterministic or stochastic, and has a global or a local
representation. It can be represented locally by a polynomial of degree p of a
variable j which measures the distance between yt and the neighboring observations
ytCj .
Assumption 2 Given "t for some time point t, it is possible to find a local
polynomial trend estimator
X
m
c t .0/ D TC
TC ct D wj ytCj ; (8.21)
jDm
where wj , j < n, denotes the weights to be applied to the observations ytCj to get the
estimate TCc t for each point in time t D 1; 2; : : : ; n.
The weights depend on: (1) the degree of the fitted polynomial, (2) the amplitude
of the neighborhood, and (3) the shape of the function used to average the
observations in each neighborhood.
Once a (symmetric) span 2m C 1 of the neighborhood has been selected, the
wj ’s for the observations corresponding to points falling out of the neighborhood
of any target point are null or approximately null, such that the estimates of the
n 2m central observations are obtained by applying 2m C 1 symmetric weights
to the observations neighboring the target point. The missing estimates for the first
212 8 Further Developments on the Henderson Trend-Cycle Filter
X
m
ct D
TC wj Bj yt D W.B/yt ; t D 1; 2; : : : ; n; (8.22)
jDm
X
m
wj D 1; (8.23)
jDm
and
X
m
ji wj D 0; (8.24)
jDm
where the weighting function f0 depends on the distance between the target point t
and each observation within a symmetric neighborhood of points around t.
Theorem 8.1 Under the Assumptions 1–3, the minimization problem (8.27) has a
unique and explicit solution.
Proof By the projection theorem (see, e.g., Priestley [28]), each element ytCj of the
Hilbert space L2 . f0 / can be decomposed into the sum of its projection in a Hilbert
subspace of L2 . f0 /, such as the space Pp , plus its orthogonal complement as follows
where ˘Pp ŒytCj
denotes the projection of the observations ytCj ; j D m; : : : ; m; on
Pp . By orthogonality, for every j 2 T
c t .0/ D TC
< ytCj ; Rp .j; 0/ >D< ˘Pp ŒytCj
; Rp .j; 0/ >D ˘Pp Œyt
D TC c t: (8.29)
c t .0/ is given by
Thus, TC
Z
c t .0/ D
TC y.t C s/Rp .s; 0/f0 .s/ds (8.30)
T
Z
D ˘Pp Œy.t C s/
Rp .s; 0/f0 .s/ds; (8.31)
T
order (p C 1),
X
p
Rp .t; 0/ D Pi .t/Pi .0/: (8.33)
iD0
X
p
KpC1 .t/ D Pi .t/Pi .0/f0 .t/: (8.34)
iD0
An important outcome of the RKHS theory is that linear filters can be grouped
into hierarchies fKp ; p D 2; 3; 4; : : :g with the following property: each hierarchy is
identified by a density f0 and contains kernels of order 2, 3, 4, . . . which are products
of orthonormal polynomials by f0 .
The weight system of a hierarchy is completely determined by specifying: (a) the
bandwidth or smoothing parameter, (b) the maximum order of the estimator in the
family, and (c) the density f0 .
There are several procedures to determine the bandwidth or smoothing parameter
(for a detailed discussion see, e.g., Berlinet and Devroye [2]). In this study, however,
the smoothing parameter is not derived by data dependent optimization criteria,
but it is fixed with the aim to obtain a kernel representation of the most often
applied Henderson smoothers. Kernels of any length, including infinite ones, can be
obtained with the above approach. Consequently, the results discussed can be easily
extended to any filter length as long as the density function and its orthonormal
polynomials are specified. The identification and specification of the density is one
of the most crucial task for smoothers based on local polynomial fitting by weighted
least squares, as LOESS and the Henderson smoothers. The density is related to the
weighting penalizing function of the minimization problem.
It should be noticed that the RKHS approach can be applied to any linear filter
characterized by varying degrees of fidelity and smoothness as described by Gray
8.3 The Henderson Filter in the Reproducing Hilbert Space (RKHS) 215
Recognition of the fact that the smoothness of the estimated trend-cycle curve
depends directly on the smoothness of the weight diagram led Henderson to develop
a formula which makes the sum of squares of the third differences of the smoothed
series a minimum for any number of terms.
Henderson’s starting point was the requirement that the filter should reproduce a
cubic polynomial trend without distortion. Henderson proved that three alternative
smoothing criteria give the same formula, as shown explicitly by Kenny and Durbin
[23] and Gray and Thomson [21]: (1) minimization of the variance of the third
differences of the series defined by the application of the moving average; (2)
minimization of the sum of squares of the third differences of the coefficients of
the moving average formula; and (3) fitting a cubic polynomial by weighted least
squares, where the weights are chosen as to minimize the sum of squares of their
third differences.
The problem is one of fitting a cubic trend by weighted least squares to the
observations ytCj ; j D m; : : : ; m, the value of the fitted function at j D 0 being
taken as the smoothed observation TC c t . Representing the weight assigned to the
residuals from the local polynomial regression by Wj ; j D m; : : : ; m , where
Wj D Wj , the problem is the minimization of
X
m
Wj ŒytCj a0 a1 j a2 j2 a3 j3
2 ; (8.35)
jDm
c t.
where the solution for the constant term aO 0 is the smoothed observation TC
Henderson showed that TC c t is given by
X
m X
m
ct D
TC .j/Wj ytCj D wj ytCj ; (8.36)
jDm jDm
where .j/ is a cubic polynomial whose coefficients have the property that the filter
reproduces the data if they follow a cubic. Henderson also proved the converse: if
the coefficients of a cubic-reproducing summation formula fwj ; j D m; : : : ; mg do
not change their sign more than three times within the filter span, then the formula
can be represented as a local cubic smoother with weights Wj > 0 and a cubic
polynomial .j/, such that .j/Wj D wj .
216 8 Further Developments on the Henderson Trend-Cycle Filter
P
Henderson measured the amount of smoothing of the input series by ntD3 ..1
B/3 yt /2 or equivalently
P by the sum of squares of the third differences of the
3 2
weight diagram, m jDmC3 ..1 B/ wj / . The solution is that resulting from the
minimization of a cubic polynomial function by weighted least squares with
as the weighting penalty function of criterion (3) above. The weight diagram
fwj ; j D m; : : : ; mg corresponding to Eq. (8.37), known as Henderson’s ideal
formula, is obtained, for a filter length equal to 2m0 3, by
315Œ.m0 1/2 j2
.m02 j2 /Œ.m0 C 1/2 j2
.3m02 16 11j2 /
wj D : (8.38)
8m0 .m0 2 1/.4m02 1/.4m0 2 9/.4m02 25/
This optimality result has been rediscovered several times in the modern literature,
usually for asymptotic variants. Loader [25] showed that the Henderson’s ideal
formula (8.38) is a finite sample variant of a kernel with second order vanishing
moments which minimizes the third derivative of the function given by Muller [26].
In particular, Loader showed that for large m, the weights of Henderson’s ideal
penalty function Wj are approximately m6 W.j=m/, where W.j=m/ is the triweight
density function. He concluded that, for very large m, the weight diagram is
approximately .315=512/ W.j=m/.3 11.j=m/2 / equivalent to the kernel given
by Muller [26].
To derive the Henderson kernel hierarchy by means of the RKHS methodology,
the density corresponding to Wj and its orthonormal polynomials have to be deter-
mined. The triweight density function gives very poor results when the Henderson
smoother spans are of short or medium lengths, as in most application cases, ranging
from 5 to 23 terms. Hence, the exact density function corresponding to Wj was
derived.
Theorem 8.2 The exact probability density corresponding to the Henderson’s ideal
weighting penalty function (8.37) is given by
.m C 1/
f0H .t/ D W..m C 1/t/; t 2 Œ1; 1
(8.39)
k
R mC1
where k D m1 W.j/dj, and j D .m C 1/t.
Proof The weighting function W.j/ in Eq. (8.37), is nonnegative in the intervals
.m 1; m C 1/, .m 3; m 2/, .m C 2; m C 3/ and negative otherwise. W.j/ is
also equal to zero if j D ˙.m C 1/; ˙.m C 2/; ˙.m C 3/, and on Œm 1; m C 1
,
W.j/ is increasing in Œm 1; 0/, decreasing in .0; m C 1
, and reaches its maximum
at j D 0.
Therefore, the support chosen is Œm 1; m C 1
to satisfy the positive definite
R mC1
condition. The integral k D m1 W.j/dj is different from 1 and represents the
integration constant on this support. It follows that the density corresponding to
8.3 The Henderson Filter in the Reproducing Hilbert Space (RKHS) 217
f0 .j/ D W.j/=k:
where t.j/ D j
mC1
, and t1 .j/ D .m C 1/t. The result follows by substitution.
The density f0H .t/ is symmetric, i.e., f0H .t/ D f0H .t/, nonnegative on Œ1; 1
, and
is equal to zero when t D 1 or t D 1. Furthermore, f0H .t/ is increasing on Œ1; 0/,
decreasing on .0; 1
, and reaches its maximum at t D 0.
For m D 6, the filter is the classical 13-term Henderson and the corresponding
probability function results
15
f0H .t/ D .5184 12289t2 C 9506t4 2401t6/; t 2 Œ1; 1
: (8.40)
79376
To obtain higher order kernels the corresponding orthonormal polynomials have to
be computed for the density (8.39). The polynomials can be derived by the Gram–
Schmidt orthonormalization procedure or by solving the Hankel system based on
the moments of the density f0H . This latter is the approach followed in this study.
The hierarchy corresponding to the 13-term Henderson kernel is shown in Table 8.3,
where for p D 3 it provides a representation of the classical Henderson filter.
Since the triweight density function gives a poor approximation of the Henderson
weights for small m (5–23 terms), another density function with well-known
theoretical properties was searched. The main reason is that the exact density (8.39)
is function of the bandwidth and need to be calculated any time that m changes
together with its corresponding orthonormal polynomials. The biweight density was
found to give almost equivalent results to those obtained with the exact density
function without the need to be calculated any time that the Henderson smoother
length changes.
Another important advantage is that the biweight density function belongs to the
well-known Beta distribution family, that is,
r
f .t/ D 2B.sC1; 1r / .1 jtjr /s IŒ1;1
.t/; (8.41)
R1
where B.a; b/ D 0 ta1 .1 t/b1 dt with a; b > 0 is the Beta function. The
orthonormal polynomials needed for the reproducing kernel associated with the
biweight function are the Jacobi polynomials, for which explicit expressions for
computation are available and their properties have been widely studied in the
literature.
Therefore, another Henderson kernel hierarchy is obtained using the biweight
density
15
f0B .t/ D .1 t2 /2 IŒ1;1
.t/; (8.42)
16
combined with the Jacobi orthonormal polynomials. These latter are characterized
by the following explicit expression:
n
1 X nC˛ nCˇ
P˛;ˇ
n .t/ D n .t 1/nm .t C 1/m ; (8.43)
2 mD0 m nm
where ˛ D 2 and ˇ D 2. The Henderson second order kernel is given by the density
function f0B , since the reproducing kernel R1 .j; 0/ of the space of polynomials of
degree at most one is always equal to one.
On the other hand, the third order kernel is given by
15
.1 jtj2 /2 74 21 2
4
t : (8.44)
16
Table 8.4 shows the classical and the two kernels (exact and biweight) Henderson
symmetric weights for spans of 9, 13, and 23 terms, where the central weight values
are given in bold.
The small discrepancy of the two kernel functions relative to the classical
Henderson smoother is due to the fact that the exact density is obtained by
interpolation from a finite small number of points of the weighting penalty function
Wj .
On the other hand, the biweight is already a density function which is made
discrete
Pm by choosing selected points to produce the weights. The smoothing measure
3 2
jDmC3 ..1 B/ wj / for each filter span is calculated as shown in Table 8.5.
The smoothing power of the filters is very close except for the exact 9-term
Henderson kernel which gives the smoothest curve. Given the equivalence for
symmetric weights, the RKHS methodology is used to generate the correspondent
asymmetric filters for the m first and last points.
Table 8.4 Weight systems for 9-, 13-, and 23-term Henderson smoothers
Length Henderson filter
9 Classical 0:041 0:010 0:118 0:267 0:331
Exact Kernel 0:038 0:005 0:119 0:262 0:325
Biweight Kernel 0:039 0:011 0:120 0:266 0:328
13 Classical 0:019 0:028 0:000 0:065 0:147 0:214 0:240
Exact Kernel 0:019 0:027 0:001 0:066 0:147 0:213 0:238
Biweight Kernel 0:020 0:030 0:002 0:070 0:149 0:211 0:234
23 Classical 0:004 0:011 0:016 0:015 0:005 0:013 0:039 0:068 0:097 0:122 0:138 0:144
Exact Kernel 0:004 0:011 0:016 0:014 0:005 0:014 0:039 0:068 0:097 0:122 0:138 0:144
Biweight Kernel 0:005 0:014 0:018 0:014 0:001 0:019 0:045 0:072 0:098 0:118 0:132 0:137
8.3 The Henderson Filter in the Reproducing Hilbert Space (RKHS)
219
220 8 Further Developments on the Henderson Trend-Cycle Filter
Table 8.5 Sum of squares of the third differences of the weights for the classical and kernel
Henderson smoother
Filters 9-term 13-term 23-term
Classical Henderson smoother 0:053 0:01 0:00
Exact Henderson kernel 0:048 0:01 0:00
Biweight Henderson kernel 0:052 0:01 0:00
K.j=b/
wq;j D Pq ; j D m; : : : ; qI q D 0; : : : ; m 1; (8.45)
jDm K.j=b/
where K./ denotes the third order Henderson kernel, j the distance to the target point
t, b the bandwidth parameter equal to (m C 1), and (m C q C 1) is the asymmetric
filter length. For example, the asymmetric weights of the 13-term Henderson kernel
for the last point (q D 0) are given by
K.j=7/
w0;j D P0 ; j D 6; : : : ; 0: (8.46)
jD6 K.i=7/
Figure 8.6 shows the gain functions of the symmetric Henderson filter together
with the classical and kernel representation for the last point superposed on the
X11/X12ARIMA Seasonal Adjustment (SA) symmetric filter. This latter results
from the convolution of: (1) 12-term centered moving average, (2) 3 3 m.a., (3)
35 m.a., and (4) 13-term Henderson filter. Its property has been widely discussed in
Dagum et al. [12]. The gain function of the SA filter should be interpreted as relating
the spectrum of the original series to the spectrum of the estimated seasonally
adjusted series.
It is apparent that the asymmetric Henderson kernel filter does not amplify
the signal as the classical asymmetric one and converges faster to the final one.
Furthermore, the asymmetric kernel suppresses more noise relative to the Musgrave
filter. Since the weights of the biweight and exact Henderson kernels are equal up to
the third digit, no visible differences are seen in the corresponding gain and phase
shift functions. Hence, we only show those corresponding to the exact Henderson
kernel in Figs. 8.6 and 8.7. There is an increase of the phase shift for the low
frequencies relative to that of the classical H13 but both are less than a month, as
exhibited in Fig. 8.7.
8.3 The Henderson Filter in the Reproducing Hilbert Space (RKHS) 221
Fig. 8.6 Gain functions of the (last point) asymmetric 13-term filter (continuous line) and kernel
(gray continuous line) with the corresponding symmetric (dotted line) and SA cascade filter (gray
dotted line)
Fig. 8.7 Phase shifts of the asymmetric (end point) weights of the Henderson kernel (dotted line)
and of the classical H13 filter (continuous line)
222 8 Further Developments on the Henderson Trend-Cycle Filter
References
1. Berlinet, A. (1993). Hierarchies of higher order kernels. Probability Theory and Related Fields,
94, 489–504.
2. Berlinet, A., & Devroye, L. (1994). A comparison of kernel density estimates. Publications de
l’Institut de Statistique de l’Université de Paris, 38, 3–59.
3. Box, G. E. P., & Jenkins, G. M. (1970). Time series analysis: Forecasting and control. San
Francisco, CA, USA: Holden-Day (Second Edition 1976).
4. Castles, I. (1987). A guide to smoothing time series estimates of trend. Catalogue No. 1316.
Australian Bureau.
5. Chhab, N., Morry, M., & Dagum, E. B. (1999). Results on alternative trend-cycle estimators
for current economic analysis. Estadistica, 49–51, 231–257.
6. Cholette, P. A. (1981). A comparison of various trend-cycle estimators. In O. D. Anderson &
M. R. Perryman (Eds.), Time series analysis (pp. 77–87). Amsterdam: North Holland.
7. Cleveland, R. B., Cleveland, W. S., McRae, J. E., & Terpenning, I. (1990). STL: A seasonal
trend decomposition procedure based on LOESS. Journal of Official Statistics, 6(1), 3–33.
8. Dagum, E. B. (1988). The X11ARIMA/88 seasonal adjustment method-foundations and user’s
manual. Ottawa, Canada: Time Series Research and Analysis Centre, Statistics Canada.
9. Dagum, E. B. (1996). A new method to reduce unwanted ripples and revisions in trend-cycle
estimates from X11ARIMA. Survey Methodology, 22, 77–83.
10. Dagum, E. B., & Bianconcini, S. (2006). Local polynomial trend-cycle predictors in reproduc-
ing kernel Hilbert spaces for current economic analysis. Anales de Economia Aplicada, 1–22.
11. Dagum, E. B., & Bianconcini, S. (2008). The Henderson smoother in reproducing kernel
Hilbert space. Journal of Business and Economic Statistics, 26(4), 536–545.
12. Dagum, E. B., Chhab, N., & Chiu, K. (1996). Derivation and analysis of the X11ARIMA and
Census X11 linear filters. Journal of Official Statistics, 12(4), 329–347.
13. Dagum, E. B., & Laniel, N. (1987). Revisions of trend-cycle estimators of moving average
seasonal adjustment methods. Journal of Business and Economic Statistics, 5, 177–189.
14. Dagum, E. B., & Luati, A. (2000). Predictive performance of some nonparametric linear and
nonlinear smoothers for noisy data. Statistica, LX(4), 635–654.
15. Dagum, E. B., & Luati, A. (2004). A linear transformation and its properties with special
applications in time series filtering. Linear Algebra and its Applications, 338, 107–117.
16. Dagum, E. B., & Luati, A. (2009). A cascade linear filter to reduce revisions and turning points
for real time trend-cycle estimation. Econometric Reviews, 28(1–3), 40–59.
17. Dalton, P., & Keogh, G. (2000). An experimental indicator to forecast turning points in the Irish
business cycle. Journal of the Statistical and Social Inquiry Society of Ireland, XXIX, 117–176.
18. Darnè, O. (2002). La dèsaisonnalisation des chroniques économiques: analyse des conditions
conjuncturelles (Ph.D. dissertation, Department of Economics, University of Montpellier 1,
France).
19. Findley, D. F., & Martin, D. E. K. (2006). Frequency domain analysis of SEATS and X-11/X-
12-ARIMA seasonal adjustment filters for short and moderate length time series. Journal of
Official Statistics, 22, 1–34.
20. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., & Chen, B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
21. Gray, A., & Thomson, P. (1996). Design of moving-average trend filters using fidelity and
smoothness criteria. In P. M. Robinson & M. Rosenblatt (Eds.), Time series analysis (in
memory of E.J. Hannan). Vol. II. Lecture notes in statistics (Vol. 115, pp. 205–219). New York:
Springer.
22. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
23. Kenny, P., & Durbin, J. (1982). Local trend estimation and seasonal adjustment of economic
and social time series. Journal of the Royal Statistical Society A, 145, 1–41.
References 223
24. Ladiray, D., & Quenneville, B. (2001). Seasonal adjustment with the X-11 method. Lecture
notes in statistics (Vol. 158). New York: Springer.
25. Loader, C. (1999). Local regression and likelihood. New York: Springer.
26. Müller, H. G. (1984). Smooth optimum kernel estimators of regression curves, densities and
modes. Annals of Statistics, 12, 766–774.
27. Musgrave, J. (1964). A set of end weights to end all end weights. Working Paper. U.S. Bureau
of Census, Washington DC.
28. Priestley, M. B. (1981). Spectral analysis and time series. Probability and mathematical
statistics. New York: Academic Press.
29. Quenneville, B., Ladiray, D., & Lefrancois, B. (2003). A note on Musgrave asymmetrical trend-
cycle filters. International Journal of Forecasting, 19(4), 727–734.
30. Shiskin, J., Young, A. H., & Musgrave, J. C. (1967). The X-11 variant of the Census Method
II seasonal adjustment program. Technical Paper 15 (revised). US Department of Commerce,
Bureau of the Census, Washington, DC.
31. Wahba, G. (1990). Spline models for observational data. Philadelphia: SIAM.
32. Zellner, A., Hong, C., & Min, C. (1991). Forecasting turning points in international output
growth rates using Bayesian exponentially weighted autoregression, time-varying parameter
and pooling techniques. Journal of Econometrics, 48, 275–304.
Chapter 9
A Unified View of Trend-Cycle Predictors
in Reproducing Kernel Hilbert Spaces (RKHS)
In many scientific fields the observations are measured with error and in such
cases the nonstationary mean of the time series is of great interest. Based on the
assumption that a time series can be decomposed as a sum of a signal plus an erratic
component, signal extraction deals with the problem of finding the “best” estimate
of the nonstationary mean given the observations corrupted by noise.
If the time series is without seasonality or seasonally adjusted, the signal
represents the trend and cyclical components, usually referred to as trend-cycle for
they are estimated jointly in medium and short length series. The determination of
a suitable inferential methodology will hinge on the assumptions made about the
signal.
A common approach is to use nonparametric smoothing estimators which can
be finite or infinite in length. They are based on different criteria of fitting and
smoothing and are: (1) density functions, (2) local polynomial fitting, (3) graduation
theory, and (4) smoothing spline regression.
A unified approach for all of these different nonparametric estimators is provided
by means of the Reproducing Kernel Hilbert Space (RKHS) methodology.
yt D TCt C It ; (9.1)
In other words, fY.t/; t 2 Tg belong to the Hilbert space L2 . f0 / endowed with the
inner product defined by
Z
< Y1 ; Y2 >L2 . f0 / D Y1 .t/Y2 .t/f0 .t/dt 8Y1 ; Y2 2 L2 . f0 /: (9.2)
T
where kk2Pp denotes the Pp -norm. The weighting function f0 depends on the distance
between the target point t and each observation within a symmetric neighborhood
of points around t.
c t , for which a general charac-
The solution for aO 0 provides the estimate TC
terization and explicit representation can be provided by means of the RKHS
methodology.
Definition 9.1 Given a Hilbert space H , the reproducing kernel R is a function
R W T T ! R, satisfying the following properties:
1. R.t; :/ 2 H , 8t 2 T;
2. < g.:/; R.t; :/ >D g.t/, 8t 2 T and g 2 H .
228 9 A Unified View of Trend-Cycle Predictors in Reproducing Kernel Hilbert. . .
Condition 2 is called the reproducing property: the value of the function TC at the
point t is reproduced by the inner product of TC with R.t; :/. R.t; :/ is called the
reproducing kernel since
The proof easily follows by the fact that any finite dimensional Hilbert space has a
reproducing kernel [3].
Theorem 9.1 Under the Assumptions 1–3, the minimization problem (9.4) has a
unique and explicit solution given by
Z
c
TC.t/ D y.t s/KpC1 .s/ds; (9.6)
T
c D ˘Pp Œ y.t/
D< ˘Pp Œ y.t s/
; Rp .s; 0/ >D< y.t s/; Rp .s; 0/ > :
TC.t/ (9.9)
c is given by
Thus, TC.t/
Z
c D
TC.t/ ˘Pp Œ y.t s/
Rp .s; 0/f0 .s/ds (9.10)
T
Z
D y.t s/Rp .s; 0/f0 .s/ds; (9.11)
T
Second, using the Christoffel–Darboux formula, for any sequence .Pi /0ip of
.p C 1/ orthonormal polynomials in L2 . f0 /, we have
X
p
Rp .t; 0/ D Pi .t/Pi .0/: (9.13)
iD0
X
p
KpC1 .t/ D Pi .t/Pi .0/f0 .t/: (9.14)
iD0
Applied to seasonally adjusted data or data without seasonality, the kernel acts as a
locally weighted average or linear filter that for each target point t gives the estimate
X
n
ct D
TC wt;i yi ; t D 1; 2; : : : ; n; (9.15)
iD1
where wt;i denotes the weights to be applied to yi to get the estimate TC c t for each
point in time t. Once a symmetric span 2mC1 of the neighborhood has been selected,
the weights for the data corresponding to points falling out the neighborhood of any
target point are null or approximately null. Hence, the estimates of the n2m central
values are obtained by applying 2mC1 symmetric weights to the values neighboring
the target point. That is,
X
m
ct D
TC wj ytCj t D m C 1; : : : ; n m:
jDm
On the other hand, the missing estimates for the first and last m data points can be
obtained by applying asymmetric linear filters of variable lengths.
Several nonparametric estimators have been developed in the literature for
time series smoothing. One approach based on least squares includes: (1) kernel
estimators, (2) local polynomial fitting, and (3) graduation theory. A second
approach corresponds to smoothing spline regression. Smoothing splines introduce
a roughness penalty term in the minimization problem (9.4), searching for an
optimal solution between both fitting and smoothing of the data. This requires an
adapted RKHS. Here it is shown how an equivalent kernel representation for the
smoothing spline can be derived in the polynomial space Pp .
A unified perspective is provided for all of these different nonparametric esti-
mators according to which they are transformed into kernel functions. This enables
comparisons among smoothers of different order within the same hierarchy as well
as kernels of the same order, but belonging to different hierarchies. Therefore, for
every estimator the density function f0 and the corresponding reproducing kernel are
identified.
Local kernel regression deals with the problem of fitting a polynomial trend to
ytCj ; j D m; : : : ; m, where the value of the fitted function at j D 0 defines the
smoothed observation TC c t . Representing the weights assigned to the residuals from
the local polynomial regression by a symmetric and nonnegative function K0 , the
problem is the minimization of
X
ti 2
n
K0 yt a0 a1 .t i/ ap .t i/p ; (9.17)
iD1
b
Table 9.1 Second and third order kernel within the Gaussian, tricube (LOESS) and biweight
(Henderson) hierarchies
Estimator Density Third order kernel 2
2 2
Gaussian f0G .t/ D p1 exp t2 t 2 .1; 1/ K3G .t/ D p12 exp t2 3t
2 2
kernel
70
3 70
3 539
LOESS f0T .t/ D 1 jtj3 IŒ1;1
.t/; K3T .t/ D 1 jtj3 3719 t2
81
15
81
15
293 638
the Gaussian kernel, the LOESS estimator [5], and the Henderson filter [19]. The
density function and third order kernel for each estimator are shown in Table 9.1.
We refer the reader to [4, 9, 10], for a detailed discussion on how these density
functions and reproducing kernels have been derived.
Extending previous work by Whittaker [35] and Whittaker and Robinson [36],
Schoenberg [30] showed that a natural smoothing spline estimator of order ` arises
by minimizing the loss function
Z Z
.`/
2
min ky TCk2W ` D .y.t/ TC.t//2 dt C TC .t/ dt; (9.18)
TC2W2` .T/ 2 T T
where TC.`/ denotes the `-th derivative of the function TC, and W2` is the Sobolev
space of functions TC satisfying the following properties: (1) TC.i/ ; i D 1; : : : ; `1,
are absolutely continuous, (2) TC.`/ 2 L2 .T/.
The parameter regulates the balance between the goodness of fit and the
smoothness of the curve. As ! 0 the solution approaches an interpolating spline,
whereas as ! 1, the solution tends to the least squares line.
Kimeldorf and Wahba [21, 22] showed that the minimizer of Eq. (9.18) is a
polynomial spline of degree p D 2`1 with knots at the data points. It is determined
by the unique Green’s function G .t; s/, such that
Z
c D
TC.t/ G .t; s/y.s/ds: (9.19)
T
1 t s
R`; .t; s/ D R`; .t s/ D R`;1 : (9.20)
A general formula for R`;1 , denoted as R` , is given by [32],
c D A./y:
TC (9.22)
Table 9.2 2-Norm distances between classical and reproducing kernel smoothing splines
Filter length
Third order kernel 9-term 13-term 23-term
Weights Gain function Weights Gain function Weights Gain function
Sobolev space R2 0:144 3:213 0:428 1:925 0:482 1:709
Standard Laplace 0:049 1:088 0:608 3:863 0:583 2:916
Logistic ˇ D 0:2 0:021 0:480 0:271 0:588 0:260 0:471
where ! denotes the frequency in cycles per unit of time and G.!/ is the gain of the
filter. For illustrative purposes, we compute these measures for filter spans generally
applied to monthly time series, that is, 9-, 13-, and 23-term, even if filter of any
length can be considered. Table 9.2 shows that the third order standard Laplace ker-
nel presents large discrepancies for each span indicating that asymptotically it does
not approach to the CSS. On the other hand, the equivalent kernel representation
derived in the Sobolev space R2 provides the worst approximation for the shortest
filter length, but it has a better performance as the span increases.
Given the poor performance of the standard Laplace estimators, we looked
for another density function. To be coherent with the results of Eq. (9.21), we
considered the log F2m1 ;2m2 distribution family introduced by Prentice [28] which
includes the standard Normal (m1 ; m2 ! 0) and Laplace (m1 ; m2 ! 1) as limiting
cases. Many densities belong to this class of distributions, among others, the logistic
(m1 D m2 D 1) and the exponential (m1 ¤ 0; m2 ! 0). We concentrate on
the former given its strong connection with the standard Laplace and other widely
applied density functions, as recently shown by Lin and Hu [23]. These authors
modified and extended previous work by Mudholkar and George [24], providing
a characterization of the logistic density in terms of sample median and Laplace
distribution, as well as in terms of the smallest order statistics and exponential
density [20]. Furthermore, George and Ojo [16], and George et al. [15] studied the
logistic distribution as approximation of Student’s t functions. The logistic density
is defined as
1
1 2 1 t˛
f0L .t/ D sech ; t 2 .1; 1/; (9.23)
4ˇ 2 ˇ
where ˛ is the mean, set equal to 0, ˇ > 0 is the dispersion parameter, and sech is
the hyperbolic secant function.
Table 9.2 shows the results for the third order kernel within the hierarchy derived
by f0L , where the dispersion parameter has been set equal to 0.2. It is given by
5 5 21 2085 2
K3L .t/ D sech2 t t : (9.24)
4 2 16 878
This kernel representation closely approximates the CSS as shown in Table 9.2.
234 9 A Unified View of Trend-Cycle Predictors in Reproducing Kernel Hilbert. . .
Fig. 9.1 13-Term (a) second and (b) third order kernels
The kernels derived by means of the RKHS methodology provide a unified view of
presently used nonparametric trend-cycle estimators. The third order kernels in the
tricube, biweight, and logistic hierarchies are equivalent kernels for the classical
LOESS of degree 2 (LOESS 2), Henderson filter, and cubic smoothing spline,
respectively. No comparisons can be made for the third order Gaussian estimator
which is already a kernel function, and for which no counterpart exists in the
literature.
The reproducing kernel representation has important consequences in the deriva-
tion of the corresponding asymmetric smoothers, which are of crucial importance
in current analysis where the aim is to obtain the estimate of the trend-cycle for the
most recent seasonally adjusted values. We derived the asymmetric smoothers by
adapting the kernel functions to the length of the last m asymmetric filters, such that
j
K3 b
wq;j D P j D m; : : : ; qI q D 0; : : : ; m 1;
q j
jDm K3 b
Fig. 9.3 (a) Gain and (b) phase shift functions of the asymmetric (end point) weights of the third
order tricube kernel and the classical LOESS 2
Fig. 9.4 (a) Gain and (b) phase shift functions of the asymmetric (end point) weights of the third
order biweight kernel and the classical Henderson filter
of the series with respect to the middle, and represents a drawback, particularly for
economic time series where turning points are important to be identified. As shown
in Fig. 9.3a, the last point LOESS asymmetric kernel exhibits a gain function with
better properties of signal passing and noise suppression relative to the classical one.
This implies smaller filter revisions as new data are added to the series. The phase
shifts for both filters, shown in Fig. 9.3b, are smaller than one month in the signal
frequency band 0 < ! 0:06, usually assumed to correspond to the trend-cycle
component.
Similar conclusions can be derived for the last point asymmetric Henderson filter
which were developed by Musgrave [25], as shown in Fig. 9.4. They are based on the
minimization of the mean squared revisions between the final estimates (obtained
by the application of the symmetric filter) and the preliminary ones (obtained by
the application of an asymmetric filter) subject to the constraint that the sum of the
weights is equal to one [13]. The assumption made is that at the end of the series, the
seasonally adjusted values follow a linear trend-cycle plus a purely random irregular
"t , such that "t NID.0; 2 /.
The asymmetric filters for the natural cubic smoothing splines are obtained by
adding additional constraints, ensuring that the function is of degree one beyond
the boundary knots. In this study, the asymmetric classical splines are obtained by
fixing the parameter to have a symmetric filter of 2m C 1 terms, and then selecting
the last m rows of the influential matrix A./. We illustrate the results for the last
point asymmetric weights corresponding to a 23-term symmetric filter because the
9.2 Boundary Behavior 237
Fig. 9.5 (a) Gain and (b) phase shift functions of the asymmetric (end point) weights of the third
order spline kernel and the CSS smoother
spline gives very poor results for short lengths such as 13 or less. Figure 9.5a shows
that the asymmetric kernel exhibits a gain function with better properties of signal
passing and noise suppression, without implying larger phase shifts. These latter are
smaller than 1 month for both filters (see Fig. 9.5b).
This section examines how the kernel estimators perform on real seasonally adjusted
data in comparison with the LOESS, Henderson filter, and cubic smoothing spline.
The last point filters are applied to a set of fifty time series taken from the Hyd-
man’s time series library (http://www-personal.buseco.monash.edu.au/~hyndman/
TSDL/). These series are seasonally adjusted and related to different fields (finance,
labor market, production, sales, transports, meteorology, hydrology, physics, and
health). The periods selected vary to sufficiently cover the various lengths published
for these series. For each series, the length of the Henderson and LOESS filters is
selected according to the I=C (noise to signal) ratio that ranges from 0.16 to 7.35,
hence filters of length 9-, 13-, and 23-term are applied.
On the other hand, for the cubic smoothing spline the generalized cross validation
(GCV) criteria [6] is used to estimate the parameter, and consequently the length
of the filters to be applied by considering the number of non-null elements in a
central row of the matrix A./. The kernel estimator of the same length is then
calculated. The smoothing parameter is known as hyperparameter in the Bayesian
terminology, and it has the interpretation of a noise to signal ratio: the larger the ,
the smoother the output. In our sample, ranges from a minimum of 0.013, at which
corresponds a filter length equal to 7 terms, to a maximum of 15, which corresponds
to a 43-term smoother. This enables us to analyze the pattern of the two estimators
on series characterized by different degrees of variability.
238 9 A Unified View of Trend-Cycle Predictors in Reproducing Kernel Hilbert. . .
The comparisons are based on the relative filter revisions between the final
symmetric filter S and the last point asymmetric filter A, that is,
S t At
Rt D ; t D 1; 2; : : : ; n (9.25)
St
For each series and for each estimator, it is calculated the ratio between the mean
square error (MSE) of the revisions corresponding to the third order kernel RK and
K/
to the corresponding classical filter RC , that is MSE.R
MSE.RC /
.
For all the estimators, the results illustrated in Table 9.3 indicate that the ratio
is always smaller than one, showing that the kernel last point predictors introduce
smaller revisions than the classical ones. This implies that the kernel estimates will
be more reliable and efficient than the ones obtained by the application of the
classical LOESS, Henderson filter, and cubic spline. In particular, for the LOESS
estimator, 96 % of the sample has a ratio less than 0.7 and, in general, it is never
greater than 0.806. For the Henderson filter, the ratio is less than 0.7 in 20 % of the
series and for the cubic spline in 34 % of the sample. In both cases, the highest value
of the ratio is close to 0.895.
For current economic analysis, it is also of great importance to assess the
performance of the several kernel estimators for turning point detection. A turning
point is generally defined to occur at time t if (downturn):
or (upturn)
for k D 3 and m D 1 given the smoothness of the trend-cycle data [8]. For each
estimator, the time lag to detect the true turning point is obtained by calculating the
number of months it takes for the revised trend series to signal a turning point in the
same position as in the final trend series.
True turning points are those recognized in the economy after occurring. For
instance, the Business Cycle Dating Committee of the National Bureau of Economic
Research (NBER) had identified the last trough in the US economy in June 2009.
To determine this date, the behavior of various monthly indicators of the economic
activity has been analyzed. Among others, the committee considered a measure of
monthly Gross Domestic Product (GDP) developed by the private forecasting firm
Macroeconomic Advisers (MAGDP), monthly GDP and Gross Domestic Income
(GDI) developed by James H. Stock and Mark W. Watson (SWGDP, SWGDI),
real personal income excluding transfers (INC), the payroll (PAY) and household
measures of total employment (HOU). Monthly data series for industrial production
(IIP) and manufacturing trade sales (MAN) have also been taken into account. The
committee designated June as the month of the trough for several of these indicators,
that is, MAGDP, SWGDP, SWGDI, MAN, and IIP. On the other hand, the trough
was identified in October 2009 for INC, and in December 2009 for PAY and HOU.
9.2 Boundary Behavior 239
Table 9.3 MSE revision ratio between kernel and classical last point predictors
Series Ratio
Health Number of cases of measles, New York city 0.806
Number of cases of measles, Baltimore 0.512
Body weight of rats 0.544
Number of chickenpox, New York city 0.597
Hydrology Temperature, copper mine 0.539
Flows, Colorado river Lees ferry 0.580
Lake Frie levels 0.568
Flows, Chang Jiang 0.696
Labour market Wisconsin employment series, fabricated metals 0.502
US male (20 years and over) unemployment figures 0.540
Number of employed persons in Australia 0.318
Macroeconomics Consumer price index 0.346
Meteorology Degree days per heating in Chicago 0.541
Microeconomics Gambling expenditure in Victoria, Australia 0.584
Logged flour price indices over the 9-years 0.558
Miscellaneous Average daily calls to directory assistance 0.607
Physics Zurich sunspot numbers 0.546
Critical radio frequencies in Washington DC 0.621
Mean thickness (Dobson units) ozone column Switzerland 0.685
Production Basic iron production in Australia 0.696
Production of chocolate confectionery in Australia 0.628
Production of Portland cement 0.687
Electricity production in Australia 0.783
Production of blooms and slabs in Australia 0.666
Production of blooms and slabs 0.677
Sales Sales of Tasty Cola 0.459
Unit sales, Winnebago Industries 0.594
Sales of new one-family houses sold in US 0.603
Sales of souvenir shop in Queensland, Australia 0.570
Demand for carpet 0.502
Transport and Tourism Portland Oregon average monthly bus ridership 0.511
US air passenger miles 0.546
International airline passengers 0.473
Weekday bus ridership, Iowa city (monthly averages) 0.567
Passenger miles flow domestic UK 0.703
Utilities Average residential gas usage Iowa 0.633
Total number of consumer 0.615
240 9 A Unified View of Trend-Cycle Predictors in Reproducing Kernel Hilbert. . .
Table 9.4 Time lag in detecting true turning points (TP) for classical and kernel estimators
LOESS Henderson CSS
Series True TP Length Classic Kernel Length Classic Kernel Length Classic Kernel
MAGDP 06/09 13 6 4 13 4 3 27 6 4
SWGDP 06/09 13 6 4 13 4 3 29 6 6
SWGDI 06/09 13 6 4 13 5 3 21 4 4
MAN 06/09 9 4 3 9 4 2 33 8 8
IIP 06/09 9 4 3 9 2 2 25 6 6
INC 10/09 13 6 4 13 4 3 17 6 6
PAY 12/09 9 4 3 9 4 4 13 6 5
HOU 12/09 9 4 3 9 2 1 21 6 6
The results obtained by applying the classical and kernel estimators to all of these
series are shown in Table 9.4.
It can be noticed that kernel estimators perform better than classical filters.
In particular, the third order biweight kernel always detects the turning points only
very few months after it has occurred, and faster than the classical Henderson filter,
which is very well-known for fast turning point detection. Tricube LOESS kernel
also presents good time lags, not greater than 4 months, that are smaller than those
observed for the classical LOESS estimator. On the other hand, the length selected
for the classical and kernel CSS estimators is generally greater than 20 terms and,
as expected, they are generally poor predictors of the true turning points, with a
slightly better performance for the logistic kernel than for the classical estimator.
References
1. Aronszajn, N. (1950). Theory of reproducing kernels. Transaction of the AMS, 68, 337–404.
2. Berlinet, A. (1993). Hierarchies of higher order kernels. Probability Theory and Related Fields,
94, 489–504.
3. Berlinet, A., & Thomas-Agnan, C. (2003). Reproducing kernel Hilbert spaces in probability
and statistics. Boston, MA: Kluwer Academic Publishers.
4. Bianconcini, S. (2006). Trend-Cycle Estimation in Reproducing Kernel Hilbert Spaces (Ph.D.
thesis, Department of Statistics, University of Bologna).
5. Cleveland, W. S. (1979). Robust locally regression and smoothing scatterplots. Journal of the
American Statistical Association, 74, 829–836.
6. Craven, P., & Wahba, G. (1979). Smoothing noisy data with spline functions. Numerical
Mathematics, 31, 377–403.
7. Cristianini, N., & Shawe-Taylor, J. (2000). An introduction to support vector machines and
other kernel-based learning methods. New York: Cambridge University Press.
8. Dagum, E. B. (1996). A new method to reduce unwanted ripples and revisions in trend-cycle
estimates from X11ARIMA. Survey Methodology, 22(1), 77–83.
9. Dagum, E. B., & Bianconcini, S. (2006). Local polynomial trend-cycle predictors in reproduc-
ing kernel Hilbert spaces for current economic analysis. Anales de Economia Aplicada, 1–22.
10. Dagum, E. B., & Bianconcini, S. (2008). The Henderson smoother in reproducing kernel
Hilbert space. Journal of Business and Economic Statistics, 26(4), 536–545.
References 241
11. Dagum, E. B., & Bianconcini, S. (2013). A unified probabilistic view of nonparametric
predictors via reproducing kernel Hilbert spaces. Econometric Reviews, 32(7), 848–867.
12. De Boor, C., & Lynch, R. (1966). On splines and their minimum properties. Journal of
Mathematics and Mechanics, 15, 953–969.
13. Doherty, M. (1992). The Surrogate Henderson Filters in X-11. Working Paper. Wellington,
New Zealand: Statistics New Zealand.
14. Evgeniou, T., Pontil, M., & Poggio, T. (2000). Regularization networks and support vector
machines. Advanced in Computational Mathematics, 13, 1–50.
15. George, E. O., El-Saidi, M., & Singh, K. (1986). A generalized logistic approximation of the
Student t distribution. Communications in Statistics B - Simulation Computing, 15(1), 261–277.
16. George, E. O., & Ojo, M. O. (1980). On a generalization of the logistic distribution. Annals of
the Institute of Statistical Mathematics, 32, 161–169.
17. Gray, A., & Thomson, P. (1996). On a family of moving-average trend filters for the ends of
series. In Proceedings of the Business and Economic Statistics Section. American Statistical
Association Annual Meeting, Chicago.
18. Green, P. J., & Silverman, B. W. (1994). Nonparametric regression and generalized linear
models. London: Chapman and Hall.
19. Henderson, R. (1916). Note on graduation by adjusted average. Transactions of the Actuarial
Society of America, 17, 43–48.
20. Jones, M. C. (2006). The logistic and the log F distributions. In N. Balakrishan (Ed.), Handbook
of logistic distribution. New York: Dekker.
21. Kimeldorf, G. S., & Wahba, G. (1970). A correspondence between Bayesian estimation
on stochastic processes and smoothing by splines. Annals of Mathematical Statistics, 41,
495–502.
22. Kimeldorf, G. S., & Wahba, G. (1970). Splines functions and stochastic processes. Sankhya,
Series A, 32, 173–180.
23. Lin, G. D., & Hu, C. Y. (2008). On the characterization of the logistic distribution. Journal of
Statistical Planning and Inference, 138(4), 1147–1156.
24. Mudholkar, G. S., & George, E. O. (1978). A remark on the shape of the logistic distribution.
Biometrika, 65, 667–668.
25. Musgrave, J. (1964). A Set of End Weights to End All End Weights. Working paper. Washington:
US Bureau of the Census.
26. Parzen, E. (1959). Statistical Inference on Time Series by Hilbert Space Methods. Technical
Report No. 53. , Stanford, CA: Statistics Department, Stanford University.
27. Pearce, N. D., & Wand, M. P. (2006). Penalized splines and reproducing kernel methods. The
American Statistician, 60(3), 223–240.
28. Prentice, R. L. (1976). A generalization of probit and logit methods for dose response curves.
Biometrics, 32, 761–768.
29. Priestley, M. B. (1981). Spectral analysis and time series. Probability and mathematical
statistics. New York: Academic.
30. Schoenberg, I. (1964). Monosplines and quadrature formulae. In T. Greville (Ed.), Theory and
applications of spline functions. Madison, WI: University of Wisconsin Press.
31. Silverman, B. (1994). Spline smoothing: The equivalent kernel method. Annals of Statistics,
12, 898–916.
32. Thomas-Agnan, C. (1991). Splines functions and stochastic filtering. Annals of Statistics, 9(3),
1512–1527.
33. Wahba, G. (1990). Spline models for observational data. Philadelphia: SIAM.
34. Wahba, G. (1999). Support vector machine, reproducing kernel Hilbert spaces, and randomized
GACV. In B. Scholkopf, C. Burges, & A. Smola (Eds.), Advanced in kernel methods: Support
vector learning (pp. 69–88). Cambridge, MA: MIT Press.
35. Whittaker, E. T. (1923). On a new method of graduation. In Proceedings of the Edinburgh
Mathematical Association (Vol. 78, pp. 81–89).
36. Whittaker, E. T., & Robinson, G. (1924). Calculus of observations: A treasure on numerical
calculations. London: Blackie and Son.
Chapter 10
Real Time Trend-Cycle Prediction
Abstract For real time trend-cycle analysis, official statistical agencies generally
produce estimates derived from asymmetric moving average techniques. However,
the use of the asymmetric filters introduces revisions as new observations are incor-
porated to the series and also delays in detecting true turning points. This chapter
deals with a reproducing kernel approach to obtain time-varying asymmetric trend-
cycle filters that converge fast and monotonically to the corresponding symmetric
ones. It shows that the bandwidth parameter that minimizes the distance between the
gain functions of the asymmetric and symmetric filters is to be preferred. Respect
to the asymmetric filters incorporated in the X12ARIMA software currently applied
by the majority of official statistical agencies, the new set of weights produces both
smaller revisions and time delay to detect true turning points. The theoretical results
are empirically corroborated with a set of leading, coincident, and lagging indicators
of the US economy.
yt D TCt C It : (10.1)
The noise It is assumed to be white noise, WN.0; I2 /, or, more generally, a
stationary and invertible autoregressive moving average (ARMA) process. The
signal TCt ; t D 1; : : : ; n; is assumed to be a smooth function of time, such that it can
be represented locally by a polynomial of degree p in a variable j; which measures
the distance between yt and its neighboring observations ytCj ; j D m; : : : ; m.
This is equivalent to estimate the trend-cycle TC c t as a weighted moving average
as follows:
X
m
ct D
TC wj ytCj D w0 y t D m C 1; : : : ; n m; (10.2)
jDm
246 10 Real Time Trend-Cycle Prediction
where w0 D wm w0 wm contains the weights to be applied to the input
c t for each point in time.
data y0 D ytm yt ytCm to get the estimate TC
As discussed in Sect. 9.2, several nonparametric estimators, based on different
sets of weights w, have been developed in the literature. It is shown in Sect. 8.3.2
that the Henderson filter can be derived using the Reproducing Kernel Hilbert Space
methodology. In this context, its equivalent kernel representation is given by
3
X
K4 .t/ D Pi .t/Pi .0/f0B .t/ t 2 Œ1; 1
; (10.3)
iD0
q
K4 .j=bq / det.H04 Œ1; j=bq
/.1=bq /f0B .j=bq /
wq;j D Pq q
D ; (10.6)
K4 .j=bq / det.Ha /
jDm
j D m; : : : ; qI q D 0; : : : ; m 1;
where H04 Œ1; j=bq
is the Hankel matrix whose elements are the moments of
f0B , and where the first column has been substituted by the vector j=bq 0 D
1 .j=bq / .j=bq /2 .j=bq /3 . On the other hand, Ha D H04 Œ1; Sq
with Sq D
q q q q 0 Pq r
1 j j
S0 S1 S2 S3 , being Srq D jDm bq bq f0B bq the discrete approximation
of r . Finally, bq ; q D 0; : : : ; m 1, is the local bandwidth, specific for each
q
asymmetric filter. It allows to relate the discrete domain of the filter, that is,
fm; : : : ; qg, for each q D 0; : : : ; m 1, to the continuous domain of the kernel
function, that is Œ1; q
.
Proposition 1 Each asymmetric filter wq D Œwq;m wq;q
0 of length .m C q C
1/; q D 0; : : : ; m 1, admits the following matrix representation:
j D m; : : : ; qI q D 0; : : : ; m 1:
det.H04 Œ1; j=bq
/ b1q f0B bjq det.H04 Œ1; j=bq
/ b1q f0B bjq
D D : (10.9)
det.H04 Œ1; Sq
/ det.Ha /
The expression (10.9) is exactly the same we would obtain by solving for ˇO0 D TC
ct
the system of linear equations
Ha ˇ D X0q Fq y:
Q 0 Œ1; c
/
det.H det.H04 Œ1; c
/
ˇO0 D 4
D :
det.Ha / det.Ha /
Pq
1 j
Given that c D jDm
j
bq bq f0B bq ytCj , it follows that
X
q
j 1 j
det.H04 Œ1; bq
/ D det.H04 Œ1;
/ f0B ytCj
jDm
bq bq bq
and therefore
X
q det.H04 Œ1; j=bq
/ b1q f0B j
bq
ct D
TC ytCj :
jDm
det.Ha /
10.1 Asymmetric Filters and RKHS 249
Hence,
w0 D e01 Ha 1 X0q Fq :
Since the trend-cycle estimates for the last m data points do not use 2m C 1
observations as for any interior point, but 2m; 2m 1; : : : ; m C 1 data, they are
subject to revisions due to new observations entering in the estimation and filters
change. As said before, we will concentrate on the reduction of revisions due to
filters change. The reduction of these revisions is an important property that the
asymmetric filters should possess together with a fast detection of true turning
points. In the specific case of the RKHS filters, Eq. (10.8) shows how the asymmetric
filter weights are related to the symmetric ones given in Eq. (10.4). It is clear that
the convergence depends on the relationship between the two discretized biweight
density functions, truncated and non-truncated, jointly with the relationship between
their respective truncated Srq and untruncated Sr discrete moments. The latter provide
an approximation of the continuous moments r , which improves as the asymmetric
filter length increases. Similarly, the convergence of Srq ; q D 0; : : : ; m, to the
corresponding non-truncated moment Sr depends on the length of the asymmetric
filter given by q, and on the local bandwidth bq . It should be noticed that bq plays
a very important role in the convergence property. For the last trend-cycle point
weight, q D 0, Eq. (10.8) reduces to
4 15
w0;0 D :
S00 4 0
S2 2 16b0
It is apparent that the largest b0 , the smaller is the weight given to the last
trend-cycle point. Since the sum of all the weights of the last point asymmetric
filter, w0;m ; : : : ; w0;0 , must be equal to one, this implies that the weights for the
remaining points are very close to one another. This can be seen in Fig. 10.1 (right
side) that shows, for m D 6, the truncated continuous biweight density function and
its discretized version when b0 is equal to 12. The opposite is observed when b0
is smaller, as shown in the same figure (left side) for b0 equal to 7. Since a larger
weight is given to the last point, much smaller weights have to be assigned to the
remaining ones for all of them to add to one. Next, we introduce time-varying local
bandwidths to improve the properties of the asymmetric filters in terms of size of
revisions and time delay to signal the upcoming of true turning points.
250 10 Real Time Trend-Cycle Prediction
0∗ 0∗
0 0
0.8
0.8
b0 = 7 b0 = 12
0.4
0.4
0.0
0.0
−1.0 −0.5 0.0 0.5 1.0 −1.0 −0.5 0.0 0.5 1.0
t t
The main effects introduced by a linear filter on a given input can be fully described
in the frequency domain by its transfer function
X
m
.!/ D wj exp.i2!j/ ! 2 Œ1=2; 1=2
;
jDm
where, for better interpretation, the frequencies ! are given in cycles for unit of
time instead of radians. .!/ represents the Fourier transform of the filter weights,
wj ; j D m; : : : ; m, and it relates the spectral density hy .!/ and hg .!/ of the input
and of the output, respectively, by
Thus, the transfer function .!/ measures the effect of the filter on the total variance
of the input at different frequencies. It is generally expressed in polar coordinates
c t D .!/ exp.i2!t/
TC
D G.!/ exp.i2.!// exp.i2!t/
D G.!/ expŒi2.!t .!//
:
10.2 Optimal Bandwidth Selection 251
G.!/ D j .!/j is called the gain of the filter and measures the amplitude of the
output for a sinusoidal input of unit amplitude, whereas .!/ is called the phase
function and shows the shift in phase of the output compared with the input. Hence,
the transfer function plays a fundamental role to measure that part of the total
revisions due to filters change.
A measure of total revisions introduced by Musgrave [19] is given by
2 32
X
q
X
m
E4 wq;j ytj wj ytj 5 ; q D 0; : : : ; m 1; (10.11)
jDm jDm
where, in our specific case, wq;j and wj are given by Eqs. (10.8) and (10.4),
respectively. This criterion can be expressed in the frequency domain as follows:
2 32
X
q
X
m
2
E4 wq;j ei2!.tj/ wj ei2!.tj/ 5 D E .q .!/ .!//ei2!t
jDm jDm
Z 1=2
D jq .!/ .!/j2 ei4!t hy .!/d!; (10.12)
1=2
2
cos.q .!// D 2 sin q !2 . Based on Eq. (10.13), the mean square filter revision
252 10 Real Time Trend-Cycle Prediction
The first component reflects the part of the total mean square filter error which is
attributed to the amplitude function of the asymmetric filter. On the other hand,
the second term measures the distinctive contribution of the phase shift. The term
Gq .!/G.!/ is a scaling factor which accounts for the fact that the phase function is
dimensionless, i.e., it does not convey level information [21].
As previously discussed, once the length of the filter is fixed, the properties of
the asymmetric filters derived in RKHS are strongly affected by the choice of the
time-varying local bandwidths bq ; q D 0; : : : ; m 1. A filter is said to be optimal if
it minimizes both revisions and time delay to detect a true turning point.
The LHS of Eq. (10.14) is a measure of total filter revision that provides the best
compromise between the amplitude function of the asymmetric filter (gain) and its
phase function (time displacement) [3, 4, 9]. Optimal asymmetric filters in this sense
can be derived using local bandwidth parameters selected according to the following
criterion:
s
Z 1=2
bq; D min 2 jq .!/ .!/j2 d!: (10.15)
bq 0
Based on the decomposition of the total filter revision error provided in Eq. (10.14),
further bandwidth selection criteria can be defined by emphasizing more the gain or
phase shift effects, and/or by attaching varying importance to the different frequency
components, depending on whether they appear in the spectrum of the input time
series or not. In the context of smoothing a monthly input, the frequency domain
˝ D f0 ! 0:50g can be partitioned into two main intervals: (1) ˝S D f0
! 0:06g associated with cycles of 16 months or longer attributed to the signal
(trend-cycle) of the series and (2) ˝N S D f0:06 < ! 0:50g corresponding to short
cyclical fluctuations attributed to the noise.
A class of optimal asymmetric filters based on bandwidth parameters bq ; q D
0; : : : ; m 1, is selected as follows:
s
Z 1=2
bq;G D min 2 jGq .!/ G.!/j2 d!; (10.16)
bq 0
and
s Z
bq; D min 2 Gq .!/G.!/ 1 cos.q .!// : (10.17)
bq ˝S
10.2 Optimal Bandwidth Selection 253
Table 10.1 Optimal bandwidth values selected for each of the biweight asymmetric filters
corresponding to the 9-, 13-, and 23-term Henderson symmetric filters
q 0 1 2 3
bq; 6.47 5.21 4.90 4.92
bq;G 8.00 5.67 4.87 4.90
bq; 4.01 4.45 5.97 6.93
q 0 1 2 3 4 5
bq; 9.54 7.88 7.07 6.88 6.87 6.94
bq;G 11.78 9.24 7.34 6.85 6.84 6.95
bq; 6.01 6.01 7.12 8.44 9.46 10.39
q 0 1 2 3 4 5 6 7 8 9 10
bq; 17.32 15.35 13.53 12.47 12.05 11.86 11.77 11.77 11.82 11.91 11.98
bq;G 21.18 18.40 16.07 13.89 12.44 11.90 11.72 11.73 11.83 11.92 11.98
bq; 11.01 11.01 11.01 11.01 11.41 13.85 15.13 16.21 17.21 18.15 19.05
It has to be noticed that the minimization of the phase error in Eq. (10.17) is very
close to minimizing the average phase shift in month for the signal, that is
Z
1 .!/
bq; D min d! : (10.18)
bq 0:06 ˝S 2!
0.8
S q1 q3 q5 S q1 q3 q5
q0 q2 q4 q0 q2 q4
0.6
0.6
0.4
0.4
0.2
0.2
0.0
0.0
−6 −4 −2 0 2 4 6 −6 −4 −2 0 2 4 6
j j
Fig. 10.2 Time path of the asymmetric filters based on bq; (left), bq;G (right) corresponding to
the 13-term symmetric filter
0.8
0.8
S q1 q3 q5 S q1 q3 q5
q0 q2 q4 q0 q2 q4
0.6
0.6
0.4
0.4
0.2
0.2
0.0
0.0
−6 −4 −2 0 2 4 6 −6 −4 −2 0 2 4 6
j j
Fig. 10.3 Time path of the asymmetric filters based on bq; (left) and of the Musgrave asymmetric
filters (right) corresponding to the 13-term symmetric filter
in months over the signal domain. However, as shown in Fig. 10.3, their time path
is only very close to that of the filters derived by Musgrave [19] up to q D 2 but
there is no monotonic convergence of these asymmetric filters to their final ones.
As already said, the Musgrave filters are based on the minimization of the mean
squared revision between the final estimates, obtained by the application of the
symmetric filter, and the preliminary estimates, obtained by the application of an
asymmetric filter, subject to the constraint that the sum of the weights is equal to
one [12, 18]. These filters have the good property of fast detection of turning points.
This property is reflected in their phase shift function that, for the last point filter,
is illustrated in Fig. 10.4. As can be seen, both the last point Musgrave filter and
10.3 Empirical Application 255
1.2
1.5
Sym b0,G b0, b0, Mus
b0,G
0.5
0.8
b0,
( )
G( )
b0,
−0.5
0.4
Mus
−2.0
0.0
Fig. 10.4 Gain (left) and phase shift (right) functions for the last point asymmetric filters based
on b0; , b0;G , and b0; compared with the last point Musgrave filter
the one based on b0; produce almost one half of the phase shift introduced by the
filter based on b0; and a quarter of the one introduced by the filter based on b0;G at
the signal frequency band. However, the reduced phase shift produced by these two
filters is compensated by larger revisions introduced in the final estimates. Indeed,
as shown by the corresponding gain functions, the last point Musgrave filter and
the one based on b0; suppress much less noise than the filters obtained through
minimization of Eqs. (10.15) and (10.16). Furthermore, the Musgrave filter has the
worst performance since it introduces a large amplification of the power attributed
to the trend and suppresses less noise.
The asymmetric filters previously derived can be applied in many fields, such
as, macroeconomic, finance, health, hydrology, meteorology, criminology, physics,
labor markets, and utilities. In fact, in any time series where the impact of the trend-
cycle is of relevance. A set of leading, coincident, and lagging indicators of the US
economy is chosen to corroborate the theoretical conclusions discussed before. The
leading indicators are time series that have a turning point before the economy as
a whole changes, whereas the coincident indicators change direction approximately
at the same time as the whole economy, thereby providing information about the
current state of the economy. On the other hand, the lagging indicators are those that
usually change direction after the economy as a whole does. The composite indexes
are typically reported in the financial and trade press, and the data analyzed in this
study are from the St. Louis Federal Reserve Bank database, the Bureau of Labor
256 10 Real Time Trend-Cycle Prediction
Table 10.2 Ratio of the mean square percentage revision errors of the last point asymmetric filters
based on b0;G , b0; , and b0; , and the last point Musgrave filter
b0;G b0; b0;
Macro-area Series Mus Mus Mus
Leading Average weekly overtime hours: manufacturing 0.492 0.630 0.922
New orders for durable goods 0.493 0.633 0.931
New orders for nondefense capital goods 0.493 0.633 0.931
New private housing units authorized by building permits 0.475 0.651 0.927
S&P 500 stock price index 0.454 0.591 0.856
M2 money stock 0.508 0.655 0.932
10-year treasury constant maturity rate 0.446 0.582 0.849
University of Michigan: consumer sentiment 0.480 0.621 0.912
Coincident All employees: total nonfarm 0.517 0.666 0.951
Real personal income excluding current transfer receipts 0.484 0.627 0.903
Industrial production index 0.477 0.616 0.884
Manufacturing and trade sales 0.471 0.606 0.869
Lagging Average (mean) duration of unemployment 0.509 0.649 0.937
Inventory to sales ratio 0.483 0.618 0.894
Index of total labor cost per unit of output 0.515 0.663 0.983
Commercial and industrial loans at all commercial banks 0.473 0.610 0.871
the X11/X12ARIMA procedure [17]. In the sample, the ratio ranges from 0.20 to
1.98, hence filters of length 9 and 13 terms are applied.
The comparisons are based on the relative filter revisions between the final
symmetric filter S and the last point asymmetric filter A, that is,
S t At
Rt D ; t D 1; 2; : : : ; n: (10.19)
St
For each series and for each estimator, we calculate the ratio between the mean
square percentage error (MSPE) of the revisions corresponding to the filters
derived following the RKHS methodology and those corresponding to the last
point Musgrave filter. For all the estimators, the results illustrated in Table 10.2
indicate that the ratio is always smaller than one, indicating that the kernel last
point predictors, based on time-varying bandwidth parameters, introduce smaller
revisions than the Musgrave filter. This implies that the estimates obtained by the
former will be more accurate than those derived by the application of the latter. In
particular, as expected, the best performance is shown by the filter based on the
optimal bandwidth b0;G obtained by minimizing the criterion (10.16). In almost all
the series its ratio with the last point Musgrave filter is less than one half and, on
average, around 0.480. This implies that when applied to real data, the filter based
on b0;G produces a reduction of almost 50 % of the revisions introduced in the real
time trend-cycle estimates given by the Musgrave filter. The filter based on b0; ,
derived to minimize the size of total filter revisions as defined by Eq. (10.15), also
258 10 Real Time Trend-Cycle Prediction
performs very well with more than the 30 % of revision reduction with respect to
the Musgrave filter. In this case, the ratio is greater than the one corresponding to
the filter based on b0;G , but always less than 0.7 for all the series, being, on average,
around 0.636. The filter whose bandwidth parameter is selected to minimize the
average phase shift over the signal domain performs more similar to the last point
Musgrave filter but still shows revisions reduction, on average, around 10 %.
It is important that the reduction of revisions in real time trend-cycle estimates is not
achieved at the expense of increasing the time lag to detect the upcoming of a true
turning point. A turning point is generally defined to occur at time t if (downturn):
or (upturn)
Table 10.3 Time lag in detecting true turning points for the asymmetric filters based on bq;G , bq; ,
and bq; , and the Musgrave filters
Macro-area Series bq;G bq; bq; Musgrave
Leading Average weekly overtime hours: manufacturing 1 1 1 1
New orders for durable goods 1 2 3 2
New orders for nondefense capital goods 1 2 2 3
New private housing units authorized by building 2 2 3 3
permits
S&P 500 stock price index 1 2 2 2
10-Year treasury constant maturity rate 1 1 1 2
University of Michigan: consumer sentiment 1 1 1 1
Coincident All employees: total nonfarm 1 1 1 2
Real personal income excluding current transfer 1 1 1 1
receipts
Industrial production index 1 1 1 1
Manufacturing and trade sales 1 2 3 3
Lagging Average (mean) duration of unemployment 3 3 4 3
Inventory to sales ratio 1 1 1 2
Index of total labor cost per unit of output 2 2 3 2
Commercial and industrial loans at all commercial 1 1 1 1
banks
Average time lag in months 1.27 1.67 1.93 2.00
The fastest the upcoming of a turning point is detected, the fastest new policies
can be applied to counteract the impact of the business-cycle stage. Failure to
recognize the downturn in the cycle or taking a long time delay to detect it may lead
to the adoption of policies to curb expansion when in fact, a recession is already
underway.
The filters based on local bandwidth parameters selected to minimize crite-
rion (10.16) are optimal, since they drastically reduce the total revisions by one half
with respect to the Musgrave filters and, similarly, almost by one third the number
of months needed to pick up the June 2009 turning point. The real time trend-cycle
filter calculated with the bandwidth parameter that minimizes the distance between
the asymmetric and symmetric filters gain functions is to be preferred. This last point
trend-cycle filter reduced by one half the size of the revisions and by one third the
time delay to detect the June 2009 turning point relative to the Musgrave filters. For
illustrative purposes, Tables 10.4 and 10.5 give the weight systems of these filters
for 9- and 13-term symmetric filters.
260
0:22362 0:21564 0:19266 0:157478 0:11444 0:06902 0:02714 0:00000 0:00000 0:00000 0:00000 0:00000 0:00000
0:21065 0:22352 0:21065 0:174523 0:12230 0:06460 0:01357 0:01982 0:00000 0:00000 0:00000 0:00000 0:00000
0:15391 0:21013 0:23100 0:21013 0:15391 0:08000 0:01250 0:02600 0:025570 0:00000 0:00000 0:00000 0:00000
0:06338 0:14212 0:20452 0:22808 0:20452 0:14212 0:06338 0:00217 0:02978 0:01617 0:00000 0:00000 0:00000
0:00258 0:06319 0:14245 0:20533 0:22909 0:20533 0:14245 0:06319 0:00258 0:02996 0:01593 0:00000 0:00000
0:02983 0:00060 0:06762 0:14651 0:20848 0:23179 0:20848 0:14651 0:06762 0:00060 0:02983 0:01855 0:00000
0:01986 0:02982 0:00217 0:07010 0:14921 0:21106 0:23429 0:21106 0:149208 0:07010 0:00217 0:02982 0:01986
0:00000 0:01855 0:02983 0:00066 0:06762 0:14651 0:20848 0:23179 0:20848 0:14651 0:06762 0:00060 0:02983
0:00000 0:00000 0:01593 0:02996 0:00258 0:06319 0:14245 0:20533 0:22909 0:20533 0:14245 0:06319 0:00258
0:00000 0:00000 0:00000 0:01617 0:02978 0:00217 0:06338 0:14212 0:20452 0:22808 0:20452 0:14212 0:06338
0:00000 0:00000 0:00000 0:00000 0:02557 0:02600 0:01250 0:08000 0:15391 0:21013 0:23100 0:21013 0:15391
0:00000 0:00000 0:00000 0:00000 0:00000 0:01982 0:01357 0:06460 0:12230 0:17452 0:21065 0:22352 0:21065
0:00000 0:00000 0:00000 0:00000 0:00000 0:00000 0:02714 0:06902 0:11444 0:15748 0:19267 0:21564 0:22362
261
262 10 Real Time Trend-Cycle Prediction
References
1. Berlinet, A. (1993). Hierarchies of higher order kernels. Probability Theory and Related Fields,
94, 489–504.
2. Bianconcini, S., & Quenneville, B. (2010). Real time analysis based on reproducing kernel
Henderson filters. Estudios de Economia Aplicada, 28(3), 553–574.
3. Dagum, E. B. (1982). Revisions of time varying seasonal filters. Journal of Forecasting, 1,
173–187.
4. Dagum, E. B. (1982). The effects of asymmetric filters on seasonal factor revisions. Journal of
American Statistical Association, 77, 732–738.
5. Dagum, E. B. (1996). A new method to reduce unwanted ripples and revisions in trend-cycle
estimates from X11ARIMA. Survey Methodology, 22, 77–83.
6. Dagum, E. B., & Bianconcini, S. (2008). The Henderson smoother in reproducing kernel
Hilbert space. Journal of Business and Economic Statistics, 26(4), 536–545.
7. Dagum, E. B., & Bianconcini, S. (2013). A unified probabilistic view of nonparametric
predictors via reproducing kernel Hilbert spaces. Econometric Reviews, 32(7), 848–867.
8. Dagum, E. B., & Bianconcini, S. (2015). A new set of asymmetric filters for tracking the short-
term trend in real time. The Annals of Applied Statistics, 9, 1433–1458.
9. Dagum, E. B., & Laniel, N. (1987). Revisions of trend-cycle estimators of moving average
seasonal adjustment methods. Journal of Business and Economic Statistics, 5, 177–189.
10. Dagum, E. B., & Luati, A. (2009). A cascade linear filter to reduce revisions and turning points
for real time trend-cycle estimation. Econometric Reviews, 28(1–3), 40–59.
11. Dagum, E. B., & Luati, A. (2012). Asymmetric filters for trend-cycle estimation. In W. R.
Bell, S. H. Holan, & T. S. McElroy (Eds.), Economic time series: Modeling and seasonality
(pp. 213–230). Boca Raton, FL: Chapman&Hall.
12. Doherty, M. (2001). The surrogate Henderson filters in X-11. Australian and New Zealand
Journal of Statistics, 43(4), 385–392.
13. Gasser, T., & Muller, H. G. (1979). Kernel estimation of regression functions. In Lecture notes
in mathematics (Vol. 757, pp. 23–68). New York: Springer.
14. Gray, A. G., & Thomson, P. J. (2002). On a family of finite moving-average trend filters for the
ends of series. Journal of Forecasting, 21, 125–149.
15. Henderson, R. (1916). Note on graduation by adjusted average. Transaction of Actuarial
Society of America, 17, 43–48.
16. Kyung-Joon, C., & Schucany, W. R. (1998). Nonparametric Kernel regression estimation near
endpoints. Journal of Statistical Planning and Inference, 66(2), 289–304.
17. Ladiray, D., & Quenneville, B. (2001). Seasonal adjustment with the X-11 method. New York:
Springer-Verlag.
18. Laniel, N. (1985). Design Criteria for the 13-term Henderson End Weights. Working paper.
Ottawa: Methodology Branch. Statistics Canada.
19. Musgrave, J. (1964). A Set of End Weights to End All End Weights. Working paper.
Washington, DC: U.S. Bureau of Census.
20. Quenneville, B., Ladiray, D., & Lefrancois, B. (2003). A note on Musgrave asymmetrical trend-
cycle filters. International Journal of Forecasting, 19(4), 727–734.
21. Wildi, M. (2008). Real-time signal extraction: Beyond maximum likelihood principles. Berlin:
Springer.
22. Zellner, A., Hong, C., & Min, C. (1991). Forecasting turning points in international output
growth rates using Bayesian exponentially weighted autoregression, time-varying parameter,
and pooling techniques. Journal of Econometrics, 49(1–2), 275–304.
Chapter 11
The Effect of Seasonal Adjustment
on Real-Time Trend-Cycle Prediction
seasonality evolves through time. Deterministic methods assume that the seasonal
pattern can be predicted with no error or with variance of the prediction error null.
On the contrary, stochastic methods assume that seasonality can be represented
by a stochastic process, that is, a process governed by a probability law and,
consequently, the variance of the prediction error is not null. The best known
seasonal adjustment methods belong to the following types:
(i) regression methods which assume global or local simple functions of time,
(ii) stochastic model-based methods which assume simple ARIMA models, and
(iii) moving average methods which are based on linear filtering and hence do not
assume explicit parametric models.
Only methods (i), which assume global simple functions for each component, are
deterministic; the others are considered stochastic. Moving averages and ARIMA
model-based (AMB) methods are those mainly applied by statistical agencies to
produce officially seasonally adjusted series. A brief description is given in the
following and we refer the readers to Chaps. 4 and 5, where each method is
discussed in detail.
11.1.1 X12ARIMA
The X12ARIMA is today the most often applied seasonal adjustment method by
statistical agencies. It was introduced by Findley et al. [10] and is an enhanced
version of the X11ARIMA method developed at Statistics Canada by Estela Bee
Dagum [5].
The major modifications concern: (1) extending the automatic identification and
estimation of ARIMA models for the extrapolation option to many more than the
three models available in the X11ARIMA, and (2) estimating trading day variations,
moving holidays, and outliers in what is called regARIMA model. The latter
consists of regression models with ARIMA errors. More precisely, they are models
in which the mean function of the time series (or its logs) is described by a linear
combination of regressors, and the covariance structure of the series is that of an
ARIMA process. If no regressors are used, indicating that the mean is assumed to
be zero, the regARIMA model reduces to an ARIMA model.
Whether or not special problems requiring the use of regressors are present
in the series to be adjusted, an important use of regARIMA models is to extend
the series with forecasts (and backcasts) in order to improve the seasonal adjust-
ments of the most recent (and the earliest) data. Doing this reduces problems
inherent in the trend estimation and asymmetric seasonal averaging processes of
the type used by the X11 method near the ends of the series. The provision of
this extension was the most important improvement offered by the X11ARIMA
11.1 Seasonal Adjustment Methods 265
program. Its theoretical and empirical benefits have been documented in many
publications, such as Dagum [6], Bobbit and Otto [3], and the references therein.
The X12ARIMA method has all the seasonal adjustment capabilities of the
X11ARIMA variant. The same seasonal and trend moving averages are available,
and the program still offers the X11 calendar and holiday adjustment routines
incorporated in X11ARIMA. But several important new options have been included.
The modeling module is designed for regARIMA model building with sea-
sonal socioeconomic time series. To this end, several categories of predetermined
regression variables are available, including trend constants or overall means, fixed
seasonal effects, trading day effects, holiday effects, pulse effects (additive outliers),
level shifts, temporary change outliers, and ramp effects. User-specified regression
variables can also be included in the models. The specification of a regARIMA
model requires specification of both the regression variables to be included in the
model and the type of ARIMA model for the regression errors (i.e., the order
.p; d; q/.P; D; Q/s).
Specification of the regression variables depends on user knowledge about
the series being modeled. Identification of the ARIMA model for the regression
errors follows well-established procedures based on examination of various sample
autocorrelation and partial autocorrelation functions produced by the X12ARIMA
program.
Once a regARIMA model has been specified, X12ARIMA estimates its param-
eters by maximum likelihood using an Iterative Generalized Least Squares (IGLS)
algorithm. Diagnostic checking involves examination of residuals from the fitted
model for signs of model inadequacy. X12ARIMA produces several standard
residual diagnostics for model checking, as well as provides sophisticated methods
for detecting additive outliers and level shifts. Finally, X12ARIMA can produce
point forecasts, forecast standard errors, and prediction intervals from the fitted
regARIMA model.
Trading day effects occur when a series is affected by the different day-of-the-
week compositions of the same calendar month in different years. Trading day
effects can be modeled with seven variables that represent (no. of Mondays), . . . ,
(no. of Sundays) in month t. Bell and Hillmer [1] proposed a better parameterization
of the same effects using six variables defined as (no. of Mondays/ .no.
of Sundays), . . . , (no. of Saturdays/ .no. of Sundays), along with a seventh
variable for Length of Month (LOM) or its deseasonalized version, the leap-
year regressor (lpyear). In X12ARIMA the six variables are called the tdnolpyear
variables. Instead of using a seventh regressor, a simpler and often better way to
handle multiplicative leap-year effects is to re-scale the February values of the
original time series before transformation to m N Feb yt =mt , where yt is the original
time series before transformation, mt is the length of month t (28 or 29), and
mN Feb D 28:25 is the average length of February. If the regARIMA model includes
seasonal effects, these can account for the length of month effect except in
Februaries, so the trading day model only has to deal with the leap-year effect.
When this is done, only the tdnolpyear variables need be included in the model.
X12ARIMA allows explicit choice of either approach, as well as an option (td)
266 11 The Effect of Seasonal Adjustment on Real-Time Trend-Cycle Prediction
that makes a default choice of how to handle length of month effects. When
the time series being modeled represents the aggregation of some daily series
(typically unobserved) over calendar months they are called monthly flow series.
If the series instead represents the value of some daily series at the end of
the month, called a monthly stock series, then different regression variables are
appropriate.
Holiday effects in a monthly flow series arise from holidays whose dates vary
over time if: (1) the activity measured by the series regularly increases or decreases
around the date of the holiday, and (2) this affects two (or more) months depending
on the date the holiday occurs each year. Effects of holidays with a fixed date, such
as Christmas, are indistinguishable from fixed seasonal effects. Easter effects are
the most frequently found holiday effects in American and European economic
time series, since the date of Easter Sunday varies between March 22 and April
25.
X12ARIMA provides four other types of regression variables to deal with abrupt
changes in the level of a series of a temporary or permanent nature: additive outliers
(AO), level shifts (LS), temporary changes (TC), and ramps. Identifying the location
and nature of potential outliers is the object of the outlier detection methodology
implemented. This methodology can be used to detect AOs, TCs, and LSs (not
ramps); any outlier that is detected is automatically added to the model as regression
variable. Prespecified AOs, LSs, TCs, and ramps are actually simple forms of
interventions as discussed by Box and Tiao [4].
The regARIMA modeling routine fulfills the role of prior adjustment and forecast
extension of the time series. That is, it first estimates some deterministic effects,
such as calendar effects and outliers using predefined built-in options or user-
defined regressors, and it removes them from the observed series. An appropriate
ARIMA model is identified for the pre-adjusted series in order to extend it
with forecasts. Then, this extended pre-adjusted series is decomposed into the
unobserved components of the series using moving averages, also accounting for
the presence of extreme values, as follows:
1. a preliminary estimate of the trend-cycle is obtained using a centered thirteen
(five) term weighted moving average of the monthly (quarterly) original series;
2. this trend-cycle series is removed from the original one to give an initial estimate
of the seasonal and irregular components, often called SI ratios;
3. initial preliminary seasonal factors are estimated from these initial SI ratios by
applying weighted seasonal moving averages for each month over several years,
separately;
4. an initial estimate of the irregular component is obtained by removing the initial
preliminary seasonal factors from the initial seasonal irregular (SI) ratios;
5. extreme values are identified and (temporarily) adjusted to replace the initial SI
ratios with the modified ones;
6. steps 3. and 4. are repeated on the modified SI ratios, initial seasonal factors are
derived by normalizing the initial preliminary seasonal factors;
11.1 Seasonal Adjustment Methods 267
11.1.2 TRAMO-SEATS
Gomez and Maravall [12] developed at the Bank of Spain, a seasonal adjustment
software called TRAMO-SEATS which is currently applied mainly by European
statistical agencies. TRAMO stands for “Time series Regression with ARIMA
noise, Missing observations and Outliers,” and SEATS for “Signal Extraction in
ARIMA Time Series.” First, TRAMO estimates via regression the deterministic
components, which are after removed from the input data. In a second round, SEATS
estimates the seasonal and trend-cycle components from the ARIMA model fitted
to the data where the deterministic components are removed. SEATS uses the filters
derived from the linearized ARIMA model that describes the stochastic behavior of
the time series.
It should be mentioned that Eurostat, in collaboration with the National Bank
of Belgium, the US Bureau of the Census, the Bank of Spain, and the European
Central Bank, has developed an interface of TRAMO-SEATS and X12ARIMA
called DemetraC. In the Bank of Spain and Eurostat websites it is also possible
to find a considerable number of papers relevant to TRAMO-SEATS as well as in
the European Statistical System (ESS) Guideline.
TRAMO is a regression method that performs the estimation, forecasting, and
interpolation of missing observations with ARIMA errors, in the presence of possi-
bly several types of outliers. The ARIMA model can be identified automatically or
by the user. It consists of the following steps:
1. automatic detection of several types of outliers and, if necessary, estimation of
other regression variables, such as calendar effects;
2. automatic identification of an ARIMA model;
3. then, TRAMO removes all the estimated components (trading days, moving
holidays, and outliers) from the original series and passes the linearized series
to SEATS, where the actual decomposition is performed.
SEATS belongs to the class of procedures based on ARIMA models for the
decomposition of time series into unobserved components and consists of the
following steps:
1. ARIMA model estimation. SEATS starts by fitting an ARIMA model to a series
not affected by deterministic components, such as trading day variations, moving
holidays, and outliers. Let yt denote this linearized series, and consider an
additive decomposition model (multiplicative if applied to the log transformation
of yt ), such that
zt D .1 B/d .1 Bs /D yt (11.1)
represent the “differenced” series. The corresponding model for the differenced
linearized series zt can be written as
where c is equal to .B/Ny, being yN the mean of the linearized series yt . In words,
the model that SEATS assumes is a linear time series with Gaussian innovations.
When used with TRAMO, estimation of the ARIMA model is made by the exact
maximum likelihood method described in Gomez and Maravall [11].
2. Derivation of the ARIMA models for each component. The program proceeds
by decomposing the series that follows the ARIMA model (11.4) into several
components. The decomposition can be multiplicative or additive. Next we shall
discuss the additive model, since the multiplicative relation can be taken care
with the log transformation of the data. That is,
yt D Tt C Ct C St C It ; (11.5)
where Tt denotes the trend component, Ct the cycle, St represents the seasonal
component, and It the irregulars. The decomposition is done in the frequency
domain. The spectrum (or pseudospectrum) of yt is partitioned into additive
spectra, associated with the different components which are determined, mostly,
from the AR roots of the model. The trend component represents the long-term
evolution of the series and displays a spectral peak at frequency 0, whereas the
seasonal component captures the spectral peaks at seasonal frequencies (e.g.,
for monthly data these are 0.524, 1.047, 1.571, 2.094, 2.618, and 3.142). The
cyclical component captures periodic fluctuations with period longer than a year,
associated with a spectral peak for a frequency between 0 and (2=s), and short-
term variation associated with low order MA components and AR roots with
small moduli. Finally, the irregular component captures white noise behavior, and
hence has a flat spectrum. The components are determined and fully derived from
the structure of the (aggregate) ARIMA model (11.4) for the linearized series
directly identified from the data. The program is aimed at monthly or quarterly
frequency data and the maximum number of observations that can be processed is
600. One important assumption is that of orthogonality among the components,
and each one will have in turn an ARIMA model. In order to identify the
components, the canonical decomposition is used which implies that the variance
of the irregulars is maximized, whereas the trend, seasonal, and cycle are as
stable as possible (compatible with the stochastic nature of model (11.2)). The
270 11 The Effect of Seasonal Adjustment on Real-Time Trend-Cycle Prediction
Due to major economic and financial changes of global nature, seasonally adjusted
data are not smooth enough to be able to provide a clear signal of the short-term
trend. Hence, further smoothing is required, but the main objections for trend-cycle
estimation are
1. the size of the revisions of the most recent values (generally much larger than for
the corresponding seasonally adjusted estimates), and
2. the time lag in detecting true turning points.
With the aim to overcome such limitations, Dagum and Bianconcini [7–9]
have recently developed nonparametric trend-cycle filters that reduce significantly
revisions and turning point detection with respect to the Musgrave [13] filters which
are applied by X12ARIMA. These authors rely on the assumption that the input
series fyt ; t D 1; 2; : : : ; ng is seasonally adjusted (where trading day variations and
outliers, if present, have been also removed), such that it can be decomposed into
the sum of a systematic component (signal) TCt , that represents the trend and cycle
usually estimated jointly, plus an irregular component It , called the noise, as follows:
yt D TCt C It ; t D 1; : : : ; n: (11.6)
The noise It is assumed to be either white noise, WN.0; I2 / or, more generally, a
stationary and invertible ARMA process.
On the other hand, the signal TCt ; t D 1; : : : ; n; is assumed to be a smooth
function of time that can be estimated as a weighted moving average as follows:
X
m
ct D
TC wj ytCj ; t D m C 1; : : : ; n m; (11.7)
jDm
11.2 Trend-Cycle Prediction in Reproducing Kernel Hilbert Space (RKHS) 271
where fwj ; j D m; : : : ; mg are the weights applied to the (2m C 1/ observations
surrounding the target observation yt . Using the reproducing kernel Hilbert space
methodology [2], the weights can be derived from the following kernel function:
3
X
K4 .s/ D Pi .s/Pi .0/f0 .s/; s 2 Œ1; 1
; (11.8)
iD0
where f0 .s/ D 15 16
.1 s2 /2 ; s 2 Œ1; 1
, is the biweight density function, and Pi ; i D
0; 1; 2; 3; are the corresponding orthonormal polynomials. In particular, the generic
weight wj ; j D m; : : : ; m, is given by
2 2 3
4 2 j
6 71
b j
wj D 4 5 f0 ; (11.9)
S0 4 S2 2 b b
R1 P r
1 j
where r D 1 sr f0 .s/ds are the moments of f0 , and Sr D m jDm b b f0 bj
their discrete approximation that depends on the bandwidth parameter b whose
choice is of fundamental importance to guarantee specific inferential properties
to the trend-cycle estimator. It has to be selected to ensure that only (2m C 1)
observations surrounding the target point will receive nonzero weight and to
approximate, as close as possible, the continuous density function f0 , as well as
the moments r with Sr . Dagum and Bianconcini ([7] and [8]) have suggested a
time invariant global bandwidth b equal to m C 1, which gave excellent results.
At the end of the sample period, that is, t D n .m C 1/; : : : ; n, only
2m; : : : ; m C 1 observations are available, and time-varying asymmetric filters have
to be considered. At the boundary, the effective domain of the kernel function K4 is
Œ1; q
, where q << 1, instead of Œ1; 1
as for any interior point. This implies
that the symmetry of the kernel is lost, and it does not integrate to unity on the
R q
asymmetric support ( 1 K4 .s/ds ¤ 1). Based on these considerations, Dagum
and Bianconcini [8] have suggested to derive the asymmetric weights by “cutting
and normalizing” the symmetric kernel K4 , that means by omitting that part of the
function lying between q and 1, and by normalizing it on [1,q]. Hence, the
corresponding asymmetric weights result
2 2 3
4 2 j
6 bq 7 1 j
wq;j D4 q q 5 f0 ; (11.10)
S0 4 S2 2 bq bq
j D m; : : : ; qI q D 0; : : : ; m 1:
272 11 The Effect of Seasonal Adjustment on Real-Time Trend-Cycle Prediction
Pq r
1
where Srq D jDm bq bq
j
f0 bjq is the discrete approximation of q r D
R q r
1 s f0 .s/ds that are proportional to the moments of the truncated biweight density
f0 on the support Œ1; q
.
Dagum and Bianconcini [9] have analyzed in detail the statistical properties
of these asymmetric filters as function of the bandwidth parameters bq ; q D
0; : : : ; m 1. They have shown, both theoretically and empirically, that filters
based on local time-varying bandwidth parameters bq ; q D 0; : : : ; m 1, selected
to minimize the distance between the gain functions of each asymmetric filter
fwq;j ; j D 0; : : : ; q; q D 0; : : : ; mg and the symmetric one fwj ; j D m; : : : ; mg have
excellent properties. In particular, the last point trend-cycle filter reduces around
a half the size of the total revisions as well as the time delay to detect a true
turning point with respect to the Musgrave [13] filter. For a detailed discussion of
the properties of these filters and of the corresponding weight systems, we refer the
reader to Chap. 10.
Since the RKHS trend-cycle filters discussed in the previous section are applied to
seasonally adjusted data, it is important to evaluate how their statistical properties
are affected by the seasonal adjustment method used to produce the input data.
In this regard, in this section, it is analyzed the behavior of the Dagum and
Bianconcini [9] asymmetric filters in terms of size of revisions in the estimates as
new observations are added to the series, and time lag to detect true turning points.
We looked at seasonally adjusted data with outliers replaced if present.
The filters are applied to a sample of series that cover various sectors, such as
labor, imports, exports, housing, industrial production, and inventories. The series
have been observed on different time periods. Specifically, we chose a set of leading,
coincident, and lagging indicators of the US economy. The leading indicators are
time series that have a turning point before the economy as a whole changes,
whereas the coincident indicators change direction approximately at the same time
as the whole economy, thereby providing information about the current state of the
economy. On the other hand, the lagging indicators are those that usually change
direction after the economy as a whole does. The composite indexes are typically
reported in the financial and trade press, and the data analyzed in this study are from
the St. Louis Federal Reserve Bank database, the Bureau of Labor Statistics, and the
National Bureau of Economic Research (NBER). In particular, the analyzed leading
indicators are
– Average weekly overtime hours, manufacturing
– New orders for durable goods
– New orders for nondefense capital goods
– New private housing units authorized by building permits
11.3 Empirical Application 273
Original series
12
X12ARIMA SA series
10
8
6
4
Fig. 11.2 Original and X12ARIMA (default option) seasonally adjusted US Unemployment Rate
for Males (16 years and over) from January 1992 to December 2013
274 11 The Effect of Seasonal Adjustment on Real-Time Trend-Cycle Prediction
12
Original series
TRAMO-SEATS SA series
10
8
6
4
Fig. 11.3 Original and TRAMO-SEATS (default option) seasonally adjusted US Unemployment
Rate for Males (16 years and over) from January 1992 to December 2013
S t At
Rt D ; t D 1; 2; : : : ; n: (11.11)
St
For
q each series, it is calculated the root mean square percentage error (RMSPE),
100 PN 2
N tD1 Rt , and the mean absolute percentage error (MAPE) of the revisions,
100 P N
N tD1 jRt j, corresponding to the filters derived following the RKHS methodol-
ogy. Table 11.1 gives the RMSPE and MAPE of the revisions introduced by the
RKHS filters when applied to the SA data produced by X12ARIMA and TRAMO-
SEATS. The size of the revisions is small and very close to each other. An indication
is that the seasonal adjustment method chosen has no impact on the final results from
the revisions viewpoint.
or (upturn)
Following Zellner et al. [14], k D 3 and m D 1 are chosen given the smoothness
of the trend-cycle data. For the values given by the RKHS method, the time lag to
detect the true turning point is affected by the convergence path of its asymmetric
filters fwq;j ; j D m; : : : ; qI q D 0; : : : ; m 1g to the symmetric one fwj ; j D
m; : : : ; mg.
276
Table 11.1 Values of RMSPE and MAPE of the revisions induced by RKHS filters applied on different SA series
RMSPE MAPE
Macro-area Series X12ARIMA TRAMO-SEATS X12ARIMA TRAMO-SEATS
Leading Average weekly overtime hours: manufacturing 0.489 0.501 3.135 3.160
New orders for durable goods 0.438 0.450 3.075 3.210
New orders for nondefense capital goods 0.497 0.508 3.844 3.979
New private housing units authorized by building permits 0.577 0.586 4.160 4.294
S&P 500 stock price index 0.936 0.945 6.169 6.674
10-Year treasury constant maturity rate 0.986 1.046 7.685 8.055
University of Michigan: consumer sentiment 0.544 0.583 4.297 4.431
Coincident All employees: total nonfarm 0.049 0.049 0.419 0.420
Unemployment rate for males (16 years and over) 0.604 0.635 2.241 2.420
Real personal income excluding current transfer receipts 0.340 0.356 2.866 3.000
Industrial production index 0.128 0.125 0.962 0.959
Manufacturing and trade sales 0.319 0.281 2.497 2.354
Lagging Average (mean) duration of unemployment 0.364 0.390 2.778 3.021
Index of total labor cost per unit of output 0.150 0.183 1.202 1.336
Commercial and industrial loans at all commercial banks 0.242 0.240 2.127 2.102
Mean 0.395 0.406 2.858 3.019
11 The Effect of Seasonal Adjustment on Real-Time Trend-Cycle Prediction
References 277
Table 11.2 Ratio in the time delay in identifying true turning points for RKHS filters applied on
seasonally adjusted series
Time lag
Macro-area Series X12ARIMA TRAMO-SEATS
Leading Average weekly overtime hours: manufacturing 2.60 3.00
New orders for durable goods 2.33 2.50
New orders for nondefense capital goods 2.33 2.43
New private housing units authorized 2.67 3.00
by building permits
S&P 500 stock price index 2.50 2.75
10-Year treasury constant maturity rate 1.00 2.00
University of Michigan: consumer sentiment 1.00 1.00
Coincident All employees: total nonfarm 1.00 1.00
Unemployment rate for males (16 years and over) 1.61 1.94
Industrial production index 1.80 2.33
Manufacturing and trade sales 1.75 2.33
Lagging Average (mean) duration of unemployment 1.00 1.00
Index of total labor cost per unit of output 2.67 3.29
Commercial and industrial loans 1.20 1.50
at all commercial banks
Mean 1.759 2.177
To determine the time lag to detect a true turning point, we calculate the number
of months it takes for the real time trend-cycle estimate to signal a turning point in
the same position as in the final trend-cycle series. We have used a very large sample
of series and, but we illustrate here only the results for the leading, coincident, and
lagging indicators shown in Table 11.2. For the three sets in total 40 turning points
were detected and the average delay in month is equal to 1.759 when the input is
seasonally adjusted with X12ARIMA and 2.177 when done with TRAMO-SEATS.
This latter takes longer to identify a true turning point. In fact, two series of the
leading indicator set and one of the lagging indicator set show time delays of 3 and
3.29 months.
We can conclude that the seasonally adjusted series with X12ARIMA is to be
preferred relative to that from TRAMO-SEATS concerning the time delay to detect
a true turning point.
References
1. Bell, W. H., & Hillmer, S. C. (1984). Issues involved with the seasonal adjustment of economic
time series. Journal of Business and Economic Statistics, 2, 291–320.
2. Berlinet, A. (1993). Hierarchies of higher order kernels. Probability Theory and Related Fields,
94, 489–504.
278 11 The Effect of Seasonal Adjustment on Real-Time Trend-Cycle Prediction
3. Bobbit, L., & Otto, M. C. (1990). Effects of forecasts on the revisions of seasonally
adjusted values using the X11 seasonal adjustment procedure. In Proceedings of the American
Statistical Association Business and Economic Statistics Session (pp. 449–453).
4. Box, G. E. P., & Tiao, G. C. (1975). Intervention analysis with applications to economics and
environmental problems. Journal of the American Statistical Association, 70, 70–79.
5. Dagum, E. B. (1978). Modelling, forecasting and seasonally adjusting economic time series
with the X-11 ARIMA method. The Statistician, 27(3), 203–216.
6. Dagum, E. B. (1988). The X-11 ARIMA/88 seasonal adjustment method-foundations and user’s
manual. Ottawa, Canada: Time Series Research and Analysis Centre, Statistics Canada.
7. Dagum, E. B., & Bianconcini, S. (2008). The Henderson smoother in reproducing kernel
Hilbert space. Journal of Business and Economic Statistics, 26(4), 536–545.
8. Dagum, E. B., & Bianconcini, S. (2013). A unified probabilistic view of nonparametric
predictors via reproducing kernel Hilbert spaces. Econometric Reviews, 32(7), 848–867.
9. Dagum, E. B., & Bianconcini, S. (2015). A new set of asymmetric filters for tracking the short-
term trend in real time. The Annals of Applied Statistics, 9, 1433–1458.
10. Findley, D. F., Monsell, B. C., Bell, W. R., Otto, M. C., & Chen, B. C. (1998). New capabilities
and methods of the X12ARIMA seasonal adjustment program. Journal of Business and
Economic Statistics, 16(2), 127–152.
11. Gomez, V., & Maravall, A. (1994). Estimation, prediction and interpolation for nonstationary
series with the Kalman filter. Journal of the American Statistical Association, 89, 611–624.
12. Gomez, V., & Maravall, A. (1996). Program TRAMO and SEATS: Instructions for users.
Working Paper 9628. Service de Estudios, Banco de Espana.
13. Musgrave, J. (1964). A set of end weights to end all end weights. Working Paper. U.S. Bureau
of Census, Washington, DC.
14. Zellner, A., Hong, C., & Min, C. (1991). Forecasting turning points in international output
growth rates using Bayesian exponentially weighted autoregression, time-varying parameter,
and pooling techniques. Journal of Econometrics, 49(1–2), 275–304.
Glossary
Structural Time Series Model Model that consists of a number of linear stochas-
tic processes that stand for the trend, cycle, seasonality, and remaining stationary
dynamic features in an observed times series.
Time Series Set of observations on a given phenomenon that are ordered in time.
Formally, it is defined as a sample realization of a stochastic process, that is, a
process that evolves according to a probability law.
Trading Day Effects Variations associated with the composition of the calendar
which are due to the fact that the activities of some days of the week are
more important than others. Trading day variations imply the existence of a
daily pattern analogous to the seasonal pattern. However, these daily factors are
usually referred to as daily coefficients. This occurs because only non-leap year
Februaries have four of each day—four Mondays, four Tuesdays, etc. All other
months have an excess of some types of days. If an activity is higher on some
days compared to others, then the series can have a trading day effect.
TRAMO-SEATS It is a seasonal adjustment software composed of two main
parts: (1) TRAMO that automatically selects a regARIMA model to estimate
deterministic components such as trading day variations, moving holidays, and
outliers that are removed from the time series and (2) SEATS that uses the
time series where the deterministic components have been removed in order to
estimate the trend-cycle and seasonality and produces a seasonally adjust series.
Transfer Function It measures the effect of a linear filter on the total variance
of the input at different frequencies, and it is generally expressed in polar
coordinates.
Trend The concept of trend is used in economics and other sciences to represent
long-term smooth variations. The causes of these variations are often associated
with structural phenomena such as population growth, technological progress,
capital accumulation, new practices of business, and economic organization. For
most economic time series, the trends evolve smoothly and gradually, whether in
a deterministic or stochastic manner.
Trend-Cycle The trend-cycle is a short-term trend that includes the impact of the
long-term trend plus that of the business cycle. This is due to the fact that most
series are relatively short and then it is impossible to estimate the long-term
trend by itself. It is important to assess the current stage of the business cycle,
particularly to forecast the coming of a turning point.
Unwanted Ripple A 10-month cycle identified by the presence of high power
at the frequency 0.10 which, due to its periodicity, often leads to the wrong
identification of a true turning point.
White Noise Process Stochastic process characterized by random variables that
have a constant expected value (usually zero), a finite variance, and are mutually
uncorrelated. If the variance is assumed constant (homoscedastic condition), the
process is referred as white noise in the strict sense. If the variance is finite but
not constant (heteroscedastic condition), the process is called white noise in the
weak sense.
X11ARIMA The X11ARIMA is an enhanced version of the Census Method II–
X11 variant that was developed mainly to produce a better current seasonal
Glossary 283