ps4
ps4
1. Let X̂ = X + V . Suppose X and X̂ both take values in {0, 1}. Is it possible for E[V ] = 0 and
Cov[X, V ] = 0? Conclude that one cannot have classical measurement error in a binary variable.
Y = X ′β + U .
Suppose Y takes values in {0, 1} and that E[Y |X] = X ′ β. Is it reasonable to assume that Var[U |X]
does not depend on X? Explain briefly.
Y = β0 + β1 X + U ,
4. Let (Y, X) ∼ P , where Y takes values in R and X takes values in Rk+1 . Suppose E[XX ′ ] and E[XY ]
both exist and that there is no perfect colinearity in X. Let (Y1 , X1 ), . . . , (Yn , Xn ) be an i.i.d. sequence
of random vectors with distribution P . Let A be an k + 1 × k + 1 invertible matrix and define W = AX.
5. (Generalized Least Squares) Let (Y1 , X1 ), . . . , (Yn , Xn ) be an i.i.d. sequence of random vectors where Yi
take values in R and Xi takes values in Rk+1 . Suppose that E[Yi |Xi ] = Xi′ β, and Var[Yi |Xi ] = σ 2 (Xi )
and σ 2 (·) is known and σ 2 (Xi ) > 0 for all 1 ≤ i ≤ n. Define Y = (Y1 , . . . , Yn )′ , X = (X1 , . . . , Xn )′ , and
D = diag(σ 2 (X1 ), . . . , σ 2 (Xn )). Consider an estimator of β of the form
β̃n = A′ Y ,
where A = A(X1 , . . . , Xn ).
1
6. (Ridge Regression) Let (Y1 , X1 ), . . . , (Yn , Xn ) be an i.i.d. sequence of random vectors. Suppose E[Xi Xi′ ]
and E[Xi Yi ] exists. Suppose further that there is no perfect collinearity in Xi . Hence, E[Xi Xi′ ] is
invertible.
is invertible.
(c) Suppose λn → 0 as n → ∞. Find the limit in probability of
−1
1 X 1 X
β̃n = (Xi Xi′ + λn I) Xi Yi .
n n
1≤i≤n 1≤i≤n
√ d
7. Suppose n(β̂n − β) → N (0, Ω) as n → ∞. Let f : Rk+1 → R be continuously differentiable at β with
nonzero derivative f ′ (β). Let Ω̂n be a consistent estimate of Ω. Suppose that Ω is non-singular.
√
(a) Derive the limiting distribution of n(f (β̂n ) − f (β)).
(b) Construct a test of
H0 : f (β) ≤ 0 versus H1 : f (β) > 0
P {f (β) ∈ Cn } → 1 − α
as n → ∞.
Ui = Yi − BLP(Yi |Wi ) ,
where
Wi = (1, Xi , Zi )′ .
Suppose Yi takes values in {0, 1} and that E[Yi |Wi ] = Wi′ β, where β = (β0 , β1 , β2 )′ .
2
(f) Can you propose a “more efficient” estimator of β? (Hint: Remember Gauss-Markov!)
9. You are interested in the effect of a binary treatment on an outcome of interest. To this end, you
collect an i.i.d. sample of n individuals and assign them to the treatment or control group with equal
probability. Let Yi denote the outcome of the ith individual. It is assumed that E[Yi2 ] < ∞. Let Di
denote the treatment status of the ith individual, where Di = 1 if the ith individual is treated and
Di = 0 if the ith individual is not treated. You assume the following model for Yi :
Yi = αi + βi Di ,
where (αi , βi ) are independent of Di . Note that αi and βi are random variables, and βi is the effect of
the treatment on the outcome. Furthermore, the effect is “heterogeneous” in the sense that βi differs
across individuals.
which you interpret as the best linear predictor of Yi given Di . Express α and β in terms of the
distribution of (αi , βi ). In particular, show that β = E[βi ].
(c) For α ∈ (0, 1), construct Cn such that
P {β ∈ Cn } → 1 − α .
10. Download the dataset ps4.csv from Canvas. Please code up your solutions in Matlab, R or Python
and include a pdf of your code with your submission. The point of this problem is to get you to work
through linear regression by example, so please do not use any regression packages to implement your
solutions. Restrict yourself to just matrix operations. (Of course, you may want to use packages to
check your answers!) Please include your answers in the write-up that you are handing in.
Consider the regression:
Y = β0 + β1 X1 + β2 X2 + ϵ
11. (Constrained Least Squares and a Lagrange Multiplier Test) Let (Y, X, U ) satisfy
Y = X ′β + U ,
where Y and U take values in R, X = (1, X1 , . . . , Xk )′ takes values in Rk+1 , β = (β0 , . . . , βk )′ . Suppose
E[XU ] = 0, E[XX ′ ] < ∞, Var[XU ] is non-singular, and there is no perfect colinearity in X. Suppose
further that Rβ = c, where R is a p × (k + 1) matrix such that the rows of R are linearly independent.
3
Let (Y1 , X1 ), . . . , (Yn , Xn ) be an i.i.d. sample from the distribution of (Y, X). Define the constrained
least squares (CLS) estimator of β, β̃n , as the solution to
1 X
min (Yi − Xi′ b)2 .
b∈Rk+1 :Rb=c n
1≤i≤n
1 ∂
(The 2 out front just makes the algebra work out a bit more nicely.) Compute ∂β L(β, λ) and
∂
∂λ L(β, λ). Let β̃n and λ̃n be such that these two derivatives are equal to zero.
(b) Show that
−1 −1
1 X
λ̃n = R Xi Xi′ R′ Rβ̂n − c ,
n
1≤i≤n
Y = β0 + β1 X + U ,
where Y , X and U all take values in R. Suppose E[XU ] = E[U ] = 0. Suppose X is unobserved, but
X̂ = X + V is observed, where E[V ] = E[XV ] = E[U V ] = 0.