R07 Correlation and Regression IFT Notes
R07 Correlation and Regression IFT Notes
1. Introduction ............................................................................................................................. 2
2. Correlation Analysis ................................................................................................................ 2
2.1. Scatter Plots ......................................................................................................................... 2
2.2. Correlation Analysis ............................................................................................................ 3
2.3. Calculating and Interpreting the Correlation Coefficient .................................................... 4
2.4. Limitations of Correlation Analysis..................................................................................... 5
2.5. Uses of Correlation Analysis ............................................................................................... 5
2.6. Testing the Significance of the Correlation Coefficient ...................................................... 6
3. Linear Regression .................................................................................................................... 9
3.1. Linear Regression with One Independent Variable ............................................................. 9
3.2. Assumptions of the Linear Regression Model ................................................................... 10
3.3. The Standard Error of Estimate ......................................................................................... 12
3.4. The Coefficient of Determination ...................................................................................... 14
3.5. Hypothesis Testing............................................................................................................. 15
3.6. Analysis of Variance in a Regression with One Independent Variable ............................. 20
3.7. Prediction Intervals ............................................................................................................ 22
3.8. Limitations of Regression Analysis ................................................................................... 24
4. Summary ................................................................................................................................ 24
This document should be read in conjunction with the corresponding reading in the 2019 Level II
CFA® Program curriculum. Some of the graphs, charts, tables, examples, and figures are
copyright 2018, CFA Institute. Reproduced and republished with permission from CFA Institute.
All rights reserved.
Required disclaimer: CFA Institute does not endorse, promote, or warrant the accuracy or
quality of the products or services offered by IFT. CFA Institute, CFA®, and Chartered
Financial Analyst® are trademarks owned by CFA Institute.
1. Introduction
In this reading, we look at two important concepts to examine the relationship between two or
more financial variables: correlation analysis and regression analysis. For example, how to
determine if there is a relationship between the returns of the U.S. stock market and the Japanese
stock market over the past five years, or between unemployment and inflation?
LO.a: Calculate and interpret a sample covariance and a sample correlation coefficient;
and interpret a scatter plot.
2. Correlation Analysis
In this section, we look at two methods to examine how two sets of data are related to each other:
scatter plots and correlation analysis.
A scatter plot is a graph that shows the relationship between the observations for two data series
in two dimensions (x-axis and y-axis). The scatter plot below is reproduced from the curriculum:
Figure 1.
Scatter Plot of
Annual Money
Supply Growth Rate
and Inflation Rate by
Country, 1970–2001
Interpretation of Figure 1:
• The two data series here are the average growth in money supply (on the x-axis) plotted
against the average annual inflation rate (on the y-axis) for six countries.
• Each point on the graph represents (money growth, inflation rate) pair for one country.
From the six points, it is evident that there is an increase in inflation as money supply
grows.
Correlation analysis is used to measure the strength of the relationship between two variables. It
is represented as a number. The correlation coefficient is a measure of how closely related two
data series are. In particular, the correlation coefficient measures the direction and extent of
linear association between two variables.
The three scatter plots below show a positive linear, negative linear, and no linear relation
between two variables A and B. They have correlation coefficients of +1, -1 and 0 respectively.
Figure 2.
Variables with a
Correlation of 1.
Figure 3.
Variables with a
Correlation of -1.
Figure 4:
Variables with a
Correlation of 0.
In order to calculate the correlation coefficient between two variables, X and Y, we need the
following:
∑𝐍𝐍 � �
𝐢𝐢 = 𝟏𝟏(𝐗𝐗 𝐢𝐢 − 𝐗𝐗)(𝐘𝐘𝐢𝐢 − 𝐘𝐘)
Cov (X, Y) =
𝐧𝐧 − 𝟏𝟏
The table below illustrates how to apply the covariance formula. Our data is the money supply
growth rate (X i ) and the inflation rate (Y i ) for six different countries. � represents the average
X
� represents the average inflation rate.
money supply growth rate and Y
Squared Squared
Country Xi Yi Cross-Product
Deviations Deviations
Notes:
1. Divide the cross-product sum by n − 1 (with n = 6) to obtain the covariance of X and Y.
2. Divide the squared deviations sums by n − 1 (with n = 6) to obtain the variances of X and Y.
Source: International Monetary Fund.
Given the covariance between X and Y and the two standard deviations, the sample correlation
can be easily calculated.
The following equation shows the formula for computing the sample correlation of X and Y:
𝐂𝐂𝐂𝐂𝐂𝐂 (𝐗𝐗, 𝐘𝐘)
𝐫𝐫 =
𝐬𝐬𝐱𝐱 ∗ 𝐬𝐬𝐲𝐲
Cov (X,Y) 0.000437
r= = = 0.870236
σx ∗ σy 0.028071 ∗ 0.017889
LO.c: Formulate a test of the hypothesis that the population correlation coefficient equals
zero, and determine whether the hypothesis is rejected at a level of significance.
The uses of correlation analysis are highlighted through six examples in the curriculum. Instead
of reproducing the examples, the specific scenarios where they are used are listed below:
• Evaluating economic forecasts: Inflation is often predicted using the change in the
consumer price index (CPI). By plotting actual vs predicted inflation, analysts can
determine the accuracy of their inflation forecasts.
• Style analysis correlation: Correlation analysis is used in determining the appropriate
benchmark to evaluate a portfolio manager’s performance. For example, assume the
portfolio managed consists of 200 small value stocks. The Russell 2000 Value Index and
the Russell 2000 Growth Index are commonly used as benchmarks to measure the small-
cap value and small-cap growth equity segments, respectively. If there is a high
correlation between the returns to the two indexes, then it may be difficult to distinguish
between small-cap growth and small-cap value as different styles.
• Exchange rate correlations: Correlation analysis is also used to understand the
correlations among many asset returns. This helps in asset allocation, hedging strategy
and diversification of the portfolio to reduce risk. Historical correlations are used to set
expectations of future correlation. For example, suppose an investor who has an exposure
to foreign currencies. He needs to ascertain whether to increase his exposure to the
Canadian dollar or to Japanese Yen. By analyzing the historical correlations between
USD returns to holding the Canadian dollar and USD returns to holding the Japanese yen,
he will be able to come to a conclusion. If they are not correlated, then holding both the
assets helps in reducing risk.
• Correlations among stock return series: Analyzing the correlations among the stock
market indexes such as large-cap, small-cap and mid-cap helps in asset allocation and
diversifying risk. For instance, if there is a high correlation between the returns to the
large-cap index and the small-cap index, then their combined allocation may be reduced
to diversify risk.
• Correlations of debt and equity returns: Similarly, the correlation among different
asset classes, such as equity and debt, is used in portfolio diversification and asset
allocation. For example, high-yield corporate bonds may have a high correlation to equity
returns, whereas long-term government bonds may have a low correlation to equity
returns.
• Correlations among net income, cash flow from operations, and free cash flow to the
firm: Correlation analysis shows if an analyst’s decision to value a firm based only on NI
and ignore CFO and FCFF is correct. FCFF is the cash flow available to debt holders and
shareholders after all operating expenses have been paid and investments in working and
fixed capital have been made. If there is a low correlation between NI and FCFF, then the
analyst’s decision to use NI instead of FCFF/CFO to value a company is questionable.
The objective of a significance test is to assess whether there is really a correlation between
random variables, or if it is a coincidence. If it can be ascertained that the relationship is not a
result of chance, then one variable can be used to predict another variable using the correlation
coefficient.
A t-test is used to determine whether the correlation between two variables is significant. The
population correlation coefficient is denoted by ρ (rho). As long as the two variables are
distributed normally, we can use hypothesis testing to determine whether the null hypothesis
should be rejected using the sample correlation, r. The formula for the t-test is:
r √n − 2
t=
√1 − r2
Important points from Examples 7 through 10 in the curriculum are summarized below:
Example 8: Testing the correlation between money supply growth rate and inflation
Data given: The sample correlation between long-term supply growth and long-term inflation in
six countries during the 1970 - 2001 period is 0.8702. The sample has six observations.
Test the null hypothesis that the true correlation in the population is zero (ρ = 0).
Solution:
0.8702 ∗ √6 − 2
Compute the test statistic: t = = 3.532
√1 − 0.87022
Conclusion: Since the test statistic 3.532 is greater than 2.776, we can conclude that there is a
strong relationship between long-term money supply growth and long-term inflation in six
countries.
Data given: Sample correlation between the USD monthly returns to Swedish kronor and the
Japanese yen for the period from January 1990 to December 1999 is 0.2860.
Test the null hypothesis that the true correlation in the population is zero (ρ = 0).
Solution:
Conclusion: Since the test statistic 3.242 is greater than 1.98, we can reject the null hypothesis
and conclude that there is a correlation between the USD monthly return to Swedish kronor and
the Japanese yen. The correlation coefficient is smaller at 0.2860 but is still significant because
the sample is larger. As n increases, the critical value decreases and t increases.
Example 11: Testing the correlation between net income and free cash flow to the firm
Data given: The sample correlation between NI and FCF for six women’s clothing stores was
0.4045 in 2001. The sample has six observations. Test the null hypothesis that the true
correlation in the population is 0 against the alternative hypothesis that the correlation in the
population is different from 0.
Solution:
0.4045 ∗ �6 − 2
The test statistic t = = 0.8846
�1 − 2
0.4045
Critical test statistic = 2.776 (at 0.05 significance level with 4 degrees of freedom)
Decision rule: If the test statistic is greater than 2.776 or less than -2.776, then reject the null
hypothesis.
Conclusion: Since t = 0.8846 is less than 2.776, we cannot reject the null hypothesis. For this
sample of women’s clothing stores, there is no statistically significant correlation between NI
and FCFF.
LO.d: Distinguish between the dependent and independent variables in a linear regression.
3. Linear Regression
3.1. Linear Regression with One Independent Variable
Linear regression allows us to use one variable to make predictions about another, test
hypotheses about the relation between two variables, and quantify the strength of the relationship
between the two variables. Linear regression assumes a linear relationship between the
dependent and independent variables.
In simple terms, regression analysis uses the historical relationship between the independent
variable and the dependent variable to predict the values of the dependent variable. The
regression equation is expressed as follows:
Yi = b0 + bi Xi + εi
where
i = 1,….n
Y = dependent variable
b 0 = intercept
b 1 = slope coefficient
X = independent variable
ε = error term
b 0 and b 1 are called the regression coefficients.
Dependent variable is the variable being predicted. It is denoted by Y in the equation. The
variable used to explain changes in the dependent variable is the independent variable. It is
denoted by X. The equation shows how much Y changes when X changes by one unit.
A linear regression model or linear least squares method, computes the best-fit line through
the scatter plot, or the line with the smallest distance between itself and each point on the scatter
plot. The regression line may pass through some points, but not through all of them. The vertical
distances between the observations and the regression line are called error terms, denoted by ε i.
Linear regression chooses the estimated values for intercept � 𝑏𝑏0 and slope �
𝑏𝑏1 such that the sum of
the squared vertical distances between the observations and the regression line is minimized.
This is represented by the following equation. The error terms are squared, so that they don’t
cancel out each other. The objective of the model is that the sum of the squared error terms
should be minimized.
N
�1 Xi �2
�0 − b
� �Yi − b
i=1
Where b� �
0 and b1 are estimated parameters.
�0
Note: The predictions in a regression model are based on the population parameter values of 𝑏𝑏
and 𝑏𝑏�1 and not on b 0 or b 1. Figure 5 below shows the regression line.
Interpretation of the graph:
• The graph plots money growth rate, the independent variable, on the x-axis against
inflation rate, the dependent variable, on the y-axis.
• The equation estimates the value of the long-term inflation rate as b 0 + b 1 (long-term rate
of money growth) + ε.
• The error term or regression residual is the distance between the line and each of the six
points. This is also equal to the difference between the actual value of the dependent
variable and the predicted value of the dependent variable.
LO.e: Explain the assumptions underlying linear regression, and interpret regression
coefficients.
The classic linear regression model (Yi = b0 + b1 Xi + εi where i = 1,…,n) is based on the
following six assumptions:
1. The relationship between Y and X is linear in the parameters b 0 and b 1. This means that
b 0, b 1, and X can only be raised to the power of 1. b 0 and b 1 cannot be multiplied or
divided by another regression parameter. This assumption is critical for a valid linear
regression.
2. X is not random.
3. The expected value of the error term is zero. This ensures the model produces correct
estimates for b 0 and b 1.
4. The variance of the error term is constant for all observations. This is called
homoskedasticity. If the variance of the error term is not constant, then it is called
heteroskedasticity.
5. The error term, ε, is uncorrelated across observations.
6. The error term, ε, is normally distributed.
This example discusses the importance of making accurate and unbiased forecasts. It is based on
the following premise: if forecasts are accurate, every prediction of change in an economic
variable will be equal to the actual change. For an unbiased forecast, the expected value of the
error term is zero and E (actual change – predicted change) = 0
Figure 6 below shows a scatter plot of the mean forecast of current-quarter percentage change in
the CPI from the previous quarter and the actual percentage change in the CPI. It shows the fitted
regression line for the following equation:
Figure 6.
What we get:
The fitted regression line is drawn based on the equation:
Actual change = - 0.0140 + 0.9637 (predicted change)
Since b 0 and b 1 are close to 0 and 1 respectively, we can conclude that the forecast is unbiased.
LO.f: Calculate and interpret the standard error of the estimate, the coefficient of
determination, and a confidence interval for a regression coefficient.
Figure 7 below shows a fitted regression line to the scatter plot on monthly returns to the S&P
500 Index and the monthly inflation rate in the United States during the 1990s. The monthly
inflation rate is the independent variable. The stock returns is the dependent variable as stock
returns vary with the inflation rate.
Figure 7:
In contrast to the fitted regression line in Figure 6, many actual observations in Figure 7 are
farther from the regression line. This may imply the predictions were not strong, resulting in an
inaccurate forecast. The standard error of estimate (SEE) measures how well a given linear
regression model captures the relationship between the dependent and independent variables. In
simple terms, it is the standard deviation of the prediction errors.
The formula for the standard error of estimate is given below:
1 1
2 2
�Yi − b� � ∑N �2 2
0 − b1 Xi � i=1 (εı )
SEE = �∑N
i=1 �= � �
n−2 n−2
where
Numerator: Regression residual, 𝜀𝜀𝑖𝑖 = dependent variable’s actual value for each observation –
predicted value for each observation
Denominator: degrees of freedom = n – 2 = n observations – two estimated parameters 𝑏𝑏� �
0 and 𝑏𝑏1 .
The reason 2 is subtracted is because SEE describes characteristics of two variables.
Example 13:
This example illustrates how to compute the standard error of estimate based on money supply
and inflation data, and regression equation: Yi = 0.0084 + 0.5545Xi
Table 7. Computing the Standard Error of Estimate
Data given:
Column 2, 3: Values are given for money supply growth rate (the independent variable) and an
inflation rate (the dependent variable).
Calculated values: Let us see how to calculate the values for each column:
Column 4: Predicted value calculated for each observation from the regression equation
Column 5: Residual = actual value – predicted value = Yi − (b �0 + b�1 Xi )
Column 6: Squared residual and finally, the sum of squared residuals, is calculated as 0.000386
1
0.000386 2
Using Equation 5, SEE is calculated as � � = 0.009823. SEE is about 0.98 percent.
6−2
Note: From an exam point of view, the testability of a question like this is low that asks you to
calculate SEE given x and y data.
The SEE gives some indication of how certain we can be about a particular prediction of Y using
the regression equation; it still does not tell us how well the independent variable explains
variation in the dependent variable. The coefficient of determination does exactly this: it
measures the fraction of the total variation in the dependent variable that is explained by the
independent variable. It is denoted by R2.
Method 1: Square the correlation coefficient between the dependent and the independent
variable. The drawback of this method is that it cannot be used when there is more than one
independent variable.
Method 2: The percent of variation that can be explained by the regression equation.
If there was no regression equation, then the predicted value of any observation would be mean
of y i , y�ı .
Solution:
Since there is only one independent variable, both the methods can be used and they should yield
the same answer.
Method 1: in section 2.3 we had calculated the correlation coefficient for the same data to be
0.870236. Squaring it, we get coefficient of determination as 0.7573.
Method 2: using Equation 5,
LO.g: Formulate a null and alternative hypothesis about a population value of a regression
coefficient, and determine the appropriate test statistic and whether the null hypothesis is
rejected at a level of significance.
Hypothesis testing is used to assess whether the evidence supports the claim about a population.
The three things required to perform a hypothesis test using the confidence interval approach are
listed below:
• �0 or b
The estimated parameter value, b �1
• The hypothesized value of the parameter, b 0 or b 1
• A confidence interval around the estimated parameter
There are two ways to perform hypothesis testing using either the confidence interval or testing
for a significance level, that we will see in the following example.
Suppose we regress a stock’s returns on a stock market index’s returns and find that the slope
coefficient is 1.5 with a standard error estimate of 0.200. Assume we need 60 monthly
observations in our regression analysis. The hypothesized value of the parameter is 1.0, which is
the market average slope coefficient. Define the null and alternative hypothesis. Compute the test
statistic. At a 0.05 level of significance, should we reject H 0?
Solution:
Step 1: Select the significance level for the test. For instance, if we are testing at a significance
level of 0.05, then we will construct a 95 percent confidence interval.
Step 2: Define the null hypothesis. For this example, the null hypothesis is thatthere is a 95%
level of confidence that the confidence interval includes b 1. H 0 : b 1 = 1.0; H a: b 1 ≠ 1.0
Step 3: Determine the critical t value t c for a given level of significance and degrees of freedom.
In this regression with one independent variable, there are two estimated parameters, the
intercept and the coefficient on the independent variable.
Step 4: Construct the confidence interval which will span fromb �1 − t c sb� to b
�1 + t c sb� where t c
1 1
is the critical t value.
Confidence interval = 1.5 – 2(0.2) to 1.5 + 2(0.2) = 1.10 to 1.90
Step 5: Conclusion: Make a decision to reject or fail to reject the null hypothesis.
Since the hypothesized value of 1.0 for the slope coefficient falls outside the confidence interval,
we can reject the null hypothesis. This means we are 95 percent confident that the interval for the
slope coefficient does not contain 1.0.
Step 3: Compare the absolute value of the t-statistic to t c. and make a decision to reject or fail to
reject the null hypothesis. If the absolute value of t is greater than t c , then reject the null
hypothesis. Since 2.5 > 2, we can reject the null hypothesis that b 1 = 1.0. Notice that both the
approaches give the same result.
p-value: At times financial analysts report the p-value or probability value for a particular
hypothesis. The p-value is the smallest level of significance at which the null hypothesis can be
rejected. It allows the reader to interpret the results rather than be told that a certain hypothesis
has been rejected or accepted. In most regression software packages, the p-values printed for
regression coefficients apply to a test of the null hypothesis that the true parameter is equal to 0
against the alternative that the parameter is not equal to 0, given the estimated coefficient and the
standard error for that coefficient. Here are a few important points connecting t-statistic and p-
value:
Given the table below that shows the results of the regression analysis, test the null hypotheses
that β for GM stock equals 1 against the alternative hypotheses that β does not equal 1.
The regression equation is Y = 0.0036 + 1.1958X + ε
Regression Statistics
Multiple R 0.5549
R-squared 0.3079
Standard error of estimate 0.0985
Observations 60
Coefficients Standard Error t-Statistic
Alpha 0.0036 0.0127 0.284
Solution:
We will test the null hypotheses using the confidence interval and t-test approaches.
Method 1: See if the hypothesized value falls within the confidence interval. We follow the
same steps as the previous example.
•The critical value of the test statistic at the 0.05 significance level with 58 degrees of
freedomist c ≈ 2.
• Construct the 95 percent confidence interval for the data for any hypothesized value of β:
β� ± t c ∗ sβ� = 1.1958 ± 2 (0.2354) = 0.7250 to 1.6666
• Conclusion: Since the hypothesized value of β = 1 falls in this confidence interval, we
cannot reject the hypothesis at the 0.05 significance level. This also means that we
cannot reject the hypotheses that GM stock has the systematic risk as the market.
Method 2: Using the t-statistic to test the significance level.
β� − β 1.1958 − 1.0
• Compute the t-statistic for GM using equation 7: t = = = 0.8318
sβ
� 0.2354
• Since t-statistic is less than the critical value of 2, we fail to reject the null hypothesis.
Example 16: Explaining the company value based on returns to invested capital
This example shows a regression hypothesis test with a one-sided alternative. The regression
EV
equation IC i = b0 + b1 (ROICi − WACCi ) + εi where the subscript i is an index to identify the
i
EV
company, tests the relationship between IC
and ROIC − WACC. Our null hypothesis is H 0 : b 1 <=
EV
0; the significance level is 0.05. Use hypothesis testing to test the relationship between IC
and
(ROIC - WACC). The results of the regression are given:
Regression Statistics
Multiple R 0.9469
R-squared 0.8966
Standard error of estimate 0.7422
Observations 9
Coefficients Standard Error t-Statistic
Solution:
The t-statistic for a coefficient reported by software programs assumes the hypothesized value to
be 0. If you are testing for a different null hypothesis, as in the previous example, then the t-
statistic must be computed. In this example, however, we are testing for b 1 = 0. Since the t-
statistic of 7.7928 is greater than t c , we can reject the null hypothesis and conclude there is a
statistically significant relationship between EV/IC and (ROIC - WACC). R2 of 0.8966 implies
ROIC - WACC explains about 90 percent of the variation in EV/IC.
In this example, we test the null hypothesis for two parameters: a slope of 0 and a slope
coefficient of 1. In Example 11, we saw that for an unbiased forecast, the expected value of the
error term is zero and E (actual change – predicted change) = 0. For the average forecast error to
be 0, the value of b 0 (the intercept) should be 0 and the value of b 1 (slope) should be 1 in this
regression equation:
Regression Statistics
Multiple R 0.7138
R-squared 0.5095
Standard error of estimate 1.0322
Observations 80
Coefficients Standard Error t-Statistic
Intercept – 0.0140 0.3657 – 0.0384
Forecast (slope) 0.9637 0.1071 9.0008
Sources: Federal Reserve Banks of Philadelphia and St. Louis.
Solution:
1. Define the null hypotheses. First null hypothesis H 0 : b 0 = 0; alternative hypothesis H a : b 0
≠ 0; second null hypothesis H 0 : b 1 = 1; b 1 ≠ 1.
2. Select the significance level. We choose a significance level of 0.05.
3. Determine the critical t value. The t-statistic at 0.05 significance level and 78 degrees of
freedom is 1.99.
4. For the first null hypothesis, see if the hypothesized value of b 0 = 0 falls within a 95
LO.h: Describe the use of analysis of variance (ANOVA) in regression analysis; interpret
ANOVA results, calculate and interpret the F-statistic.
For a meaningful regression model the slope coefficients should be non-zero. This is determined
through the F-test which is based on the F-statistic. The F-statistic tests whether all the slope
coefficients in a linear regression are equal to 0. In a regression with one independent variable,
this is a test of the null hypothesis H 0: b 1 = 0 against the alternative hypothesis H a : b 1 ≠ 0. The F-
statistic also measures how well the regression equation explains the variation in the dependent
variable. The four values required to construct the F-statistic for null hypothesis testing are:
The F-statistic is the ratio of the average regression sum of squares to the average sum of the
squared errors. Average regression sum of squares (RSS) is the amount of variation in Y
explained by the regression equation.
RSS
1 Mean regression sum of squares
F= SSE =
Mean squared error
n−2
where
regression sum of squares
Mean regression sum of squares =
number of slope parameters estimated
sum of squared errors
Mean squared error =
number of observations n – total number of parameters estimated
• The higher the F-statistic, the better it is as the regression model does a good job of
explaining the variation in the dependent variable.
• The F-statistic is 0 if the independent variable explains no variation in the dependent
variable.
• The F-statistic is used to evaluate whether a fund has generated positive alpha. The null
hypothesis for this test will be: H 0: α = 0 versus H a : α ≠ 0.
• F-statistic is usually not used in a regression with one independent variable because the
F-statistic is the square of the t-statistic for the slope coefficient.
This example illustrates how the F-test reveals nothing more than what the t-test already does.
The ANOVA table and results of regression to evaluate the performance of the Dreyfus
appreciation fund is given below:
Regression Statistics
Multiple R 0.928
R-squared 0.8611
Standard error of estimate 0.0174
Observations 60
Degrees of Freedom Mean Sum of
ANOVA F
(df) Squares (MSS)
Regression 1 0.1093 359.64
Residual 58 0.0003
Total 59
Test the null hypothesis that α = 0, i.e. the fund did not have a significant excess return beyond
the return associated with the market risk of the fund.
Solution:
Recall that the software tests the null hypothesis for slope coefficient = 0, so we can use the
given t-statistic. Since the t-statistic of 18.9655 is high, the probability that the coefficient is 0 is
very small.
We get the same result using the F-statistic as well.
0.1093
F= 1
0.0176 = 360.19. The p-value for F-statistic is less than 0.0001. This is much lower than the
60−2
significance level.
LO.i: Calculate the predicted value for the dependent variable, given an estimated
regression model and a value for the independent variable.
LO.j: Calculate and interpret a confidence interval for the predicted value of the
dependent variable.
We use regression intervals to make predictions about a dependent variable. Earlier we discussed
the construction of a confidence interval, which is a range of values that is likely to contain the
value of an unknown population parameter. With a 95 percent confidence interval, are we 95
percent confident that, say β, will have a value of 1?
Prediction intervals focus on the accuracy. It represents an interval of values associated with a
parameter that we believe includes the true parameter value, b 1 , with a specified probability.
Let us consider the regression equation: Y = b 0 + b 1 X. The predicted value of �Y = b �0 + b�1 X.
The two sources of uncertainty in regression analysis using the estimated parameters to make a
prediction are:
Note: You need not memorize this formula, but understand the factors that affect s f 2, like higher
the n, lower the variance, and the better it is.
Once the variance of the prediction error is known, it is easy to determine the confidence interval
around the prediction. The steps are:
This part is not covered in the curriculum. But, it may help in understanding how a prediction
interval is different from a confidence interval, as both are essentially a range of values:
Given the results of a regression analysis below, predict the ratio of enterprise value to invested
capital for the company at a 95 percent confidence interval. The return spread between ROIC and
WACC is given as 10 percentage points.
Regression Statistics
Multiple R 0.9469
R-squared 0.8966
Standard error of estimate 0.7422
Observations 9
Standard
Coefficients t-Statistic
Error
Intercept 1.3478 0.3511 3.8391
Spread 30.0169 3.8519 7.7928
Source: Nelson (2003).
Solution:
1 (X − � X) 2 1 (0.1 − 0.0647)2
sf2 2
= s ∗ �1 + + 2
� = 0.7422 ∗ �1 + + �
n (n − 1)sx2 9 (9 − 1)0.004641
= 0.630556
S f = 0.7941
3. t c at the 95 percent confidence level and 7 degrees of freedom is 2.365.
4. 95 percent prediction interval using equation 8 is:
4.3495 – 2.365 (0.7941) to 4.3495 + 2.365 (0.7941) = 2.4715 to 6.2275.
• Regression relations can change over time as do correlations. This is called parameter
instability. This characteristic is observed in both cross-series and time-series regression
relationships.
• Regression analysis is often difficult to apply because of specification issues and
assumption violations.
o Specification issues: Identifying the independent variables and the dependent
variable, and formulating the regression equation may be challenging.
o Assumptions violations: Often there is an uncertainty in determining if an
assumption has been violated.
• Public knowledge of regression relationships may limit their usefulness in the future. For
example, is low P/E stocks in the sugar industry have had historically high returns during
Oct-Jan every year, then this knowledge may cause other analysts to buy the stock which
will push their prices up.
4. Summary
Below is a summary of the important points discussed in this reading:
LO.a: Calculate and interpret a sample covariance and a sample correlation coefficient;
and interpret a scatter plot.
• Correlation analysis is used to measure the strength of the relationship between two
variables.
• The sample correlation is denoted by r, and the population correlation by ρ. The
correlation coefficient has no units. It can take a value between -1 and 1.
∑N � )(Yi − Y
(Xi − X �)
• The sample covariance is Cov (X,Y) = i = 1
n−1
Cov(X,Y)
• The sample correlation coefficient of two variables X and Y is
sx ∗sy
LO.b: Describe the limitations to correlation analysis.
LO.c: Formulate a test of the hypothesis that the population correlation coefficient equals
zero, and determine whether the hypothesis is rejected at a level of significance.
LO.d: Distinguish between the dependent and independent variables in a linear regression.
LO.e: Explain the assumptions underlying linear regression, and interpret regression
coefficients.
• Linear regression allows us to use one variable to make predictions about another, test
hypotheses about the relation between two variables, and quantify the strength of the
relationship between the two variables.
• Linear regression assumes a linear relationship between the dependent and the
independent variables.
• A simple linear regression using one independent variable can be expressed as:
Yi = b0 + bi Xi + εi where i = 1,2,…n
• The regression model chooses the estimated values for b �0 and b
�1 such that this term
2
∑N � �
i = 1�Yi − b0 − b1 X i � is minimized.
• The variable used to explain the dependent variable is the independent variable, X. The
variable being explained is the dependent variable, Y.
LO.f: Calculate and interpret the standard error of the estimate, the coefficient of
determination, and a confidence interval for a regression coefficient.
1 1
�1 Xi�2 2
�0 − b ∑N �2 2
�Yi − b i = 1(εı )
• Standard error of estimate: SEE = �∑N
i=1 � =� �
n−2 n−2
• Coefficient of determination
• The higher the R2, the more useful the model. R2 has a value between 0 and 1.
• The higher the R2, the more useful the model. R2 has a value between 0 and 1.
• Correlation coefficient, r, is also called multiple-R.
• A confidence interval is an interval of values that we believe includes the true parameter
value, b 1 , with a given degree of confidence.
LO.g: Formulate a null and alternative hypothesis about a population value of a regression
coefficient, and determine the appropriate test statistic and whether the null hypothesis is
rejected at a level of significance.
LO.h: Describe the use of analysis of variance (ANOVA) in regression analysis; interpret
ANOVA results, calculate and interpret the F-statistic.
LO.i: Calculate the predicted value for the dependent variable, given an estimated
regression model and a value for the independent variable.
LO.j: Calculate and interpret a confidence interval for the predicted value of the
dependent variable.
• Analysis of variance is a statistical procedure for dividing the variability of a variable into
components that can be attributed to different sources. We use ANOVA to determine the
usefulness of the independent variable or variables in explaining variation in the
dependent variable.
• The F-statistic measures how well the regression equation explains the variation in the
dependent variable.
RSS
Mean regression sum of squares
• F= 1
SSE =
Mean squared error
n−2
where
regression sum of squares
Mean regression sum of squares =
number of slope parameters estimated
sum of squared errors
Mean squared error =
number of observations n – total number of parameters estimated
• � ± t c ∗ sf
The prediction interval is given by: Y
• Regression relations can change over time as do correlations. This is called parameter
instability.
• Regression analysis is often difficult to apply because of specification issues and
assumption violations.
• Public knowledge of regression relationships may limit their usefulness in the future.