Train, K. (2003) - Discrete Choice Methods With Simulation
Train, K. (2003) - Discrete Choice Methods With Simulation
Kenneth E. Train
University of California, Berkeley
and
National Economic Research Associates, Inc.
iii
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-FMA CB495/Train September 18, 2002 10:54 Char Count= 0
C Kenneth E. Train 2003
A catalog record for this book is available from the British Library.
iv
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-FMA CB495/Train September 18, 2002 10:54 Char Count= 0
Contents
1 Introduction page 1
1.1 Motivation 1
1.2 Choice Probabilities and Integration 3
1.3 Outline of Book 7
1.4 Topics Not Covered 8
1.5 A Couple of Notes 11
Part I Behavioral Models
2 Properties of Discrete Choice Models 15
2.1 Overview 15
2.2 The Choice Set 15
2.3 Derivation of Choice Probabilities 18
2.4 Specific Models 21
2.5 Identification of Choice Models 23
2.6 Aggregation 33
2.7 Forecasting 36
2.8 Recalibration of Constants 37
3 Logit 38
3.1 Choice Probabilities 38
3.2 The Scale Parameter 44
3.3 Power and Limitations of Logit 46
3.4 Nonlinear Representative Utility 56
3.5 Consumer Surplus 59
3.6 Derivatives and Elasticities 61
3.7 Estimation 64
3.8 Goodness of Fit and Hypothesis Testing 71
3.9 Case Study: Forecasting for a New
Transit System 75
3.10 Derivation of Logit Probabilities 78
v
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-FMA CB495/Train September 18, 2002 10:54 Char Count= 0
vi Contents
4 GEV 80
4.1 Introduction 80
4.2 Nested Logit 81
4.3 Three-Level Nested Logit 90
4.4 Overlapping Nests 93
4.5 Heteroskedastic Logit 96
4.6 The GEV Family 97
5 Probit 101
5.1 Choice Probabilities 101
5.2 Identification 104
5.3 Taste Variation 110
5.4 Substitution Patterns and Failure of IIA 112
5.5 Panel Data 114
5.6 Simulation of the Choice Probabilities 118
6 Mixed Logit 138
6.1 Choice Probabilities 138
6.2 Random Coefficients 141
6.3 Error Components 143
6.4 Substitution Patterns 145
6.5 Approximation to Any Random Utility Model 145
6.6 Simulation 148
6.7 Panel Data 149
6.8 Case Study 151
7 Variations on a Theme 155
7.1 Introduction 155
7.2 Stated-Preference and Revealed-Preference Data 156
7.3 Ranked Data 160
7.4 Ordered Responses 163
7.5 Contingent Valuation 168
7.6 Mixed Models 170
7.7 Dynamic Optimization 173
Part II Estimation
8 Numerical Maximization 189
8.1 Motivation 189
8.2 Notation 189
8.3 Algorithms 191
8.4 Convergence Criterion 202
8.5 Local versus Global Maximum 203
8.6 Variance of the Estimates 204
8.7 Information Identity 205
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-FMA CB495/Train September 18, 2002 10:54 Char Count= 0
Contents vii
Bibliography 319
Index 331
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-01Drv CB495/Train KEY BOARDED September 18, 2002 11:2 Char Count= 0
1 Introduction
1.1 Motivation
When I wrote my first book, Qualitative Choice Analysis, in the mid-
1980s, the field had reached a critical juncture. The breakthrough con-
cepts that defined the field had been made. The basic models – mainly
logit and nested logit – had been introduced, and the statistical and eco-
nomic properties of these models had been derived. Applications had
proven successful in many different areas, including transportation, en-
ergy, housing, and marketing – to name only a few.
The field is at a similar juncture today for a new generation of proce-
dures. The first-generation models contained important limitations that
inhibited their applicability and realism. These limitations were well
recognized at the time, but ways to overcome them had not yet been
discovered. Over the past twenty years, tremendous progress has been
made, leading to what can only be called a sea change in the approach
and methods of choice analysis. The early models have now been sup-
plemented by a variety of more powerful and more flexible methods.
The new concepts have arisen gradually, with researchers building on
the work of others. However, in a sense, the change has been more like
a quantum leap than a gradual progression. The way that researchers
think about, specify, and estimate their models has changed. Importantly,
a kind of consensus, or understanding, seems to have emerged about the
new methodology. Among researchers working in the field, a definite
sense of purpose and progress prevails.
My purpose in writing this new book is to bring these ideas together,
in a form that exemplifies the unity of approach that I feel has emerged,
and in a format that makes the methods accessible to a wide audience.
The advances have mostly centered on simulation. Essentially, simu-
lation is the researcher’s response to the inability of computers to per-
form integration. Stated more precisely, simulation provides a numerical
1
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-01Drv CB495/Train KEY BOARDED September 18, 2002 11:2 Char Count= 0
2 Introduction
Introduction 3
4 Introduction
Introduction 5
6 Introduction
Now suppose that a closedform exists for the integral in large brackets.
Label this formula g(ε1 ) ≡ ε2 I [h(x, ε1 , ε2 ) = y] f (ε2 | ε1 ) dε2 , which
on the value of ε1 . The probability then becomes
is conditional
P(y | x) = ε1 g(ε1 ) f (ε1 ) dε1 . If a closed-form solution does not ex-
ist for this integral, then it is approximated through simulation. Note
that it is simply the average of g over the marginal density of ε1 . The
probability is simulated by taking draws from f (ε1 ), calculating g(ε1 )
for each draw, and averaging the results.
This procedure is called convenient error partitioning (Train, 1995).
The integral over ε2 given ε1 is calculated exactly, while the integral over
ε1 is simulated. There are clear advantages to this approach over com-
plete simulation. Analytic integrals are both more accurate and easier to
calculate than simulated integrals. It is useful, therefore, when possible,
to decompose the random terms so that some of them can be integrated
analytically, even if the rest must be simulated. Mixed logit (in Chap-
ter 6) is a prominent example of a model that uses this decomposition
P1: GEM/IKJ P2: GEM/IKJ QC: GEM/ABE T1: GEM
CB495-01Drv CB495/Train KEY BOARDED September 18, 2002 11:2 Char Count= 0
Introduction 7
8 Introduction
Introduction 9
10 Introduction
Introduction 11
12