0% found this document useful (0 votes)
8 views

Regression DPP 01 Discussion Notes664745df1b2c900018f5ac7e

Uploaded by

Ayush
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views

Regression DPP 01 Discussion Notes664745df1b2c900018f5ac7e

Uploaded by

Ayush
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 32

Data Science &

Artificial Intelligence
Machine Learning

Regression

DPP-01 Discussion Notes By- Astha Ma’am


[MCQ]
#Q. The parameters acquired through linear regression:

A can take any value in the real space

B are strictly integers

C always lie in the range [0,1]

D can take only non-zero values


[MCQ]
#Q. Which of the statements is/are True ?

A Ridge has sparsity constraint, and it will drive coefficients with low values to 0.

B Lasso has a closed form solution for the optimization problem, but this is not the
case for Ridge.

C Ridge regression does not reduce the number of variables since it never leads a
coefficient to zero but only minimizes it.

D If there are two or more highly collinear variables, Lasso will select one of them
randomly
[MCQ]
#Q. The relation between studying time (in hours) and grade on the final
examination (0-100) in a random sample of students in the Introduction to
Machine Learning Class was found to be : Grade = 30.5 + 15.2 (h) How will a
student's grade be affected if she studies for four hours ?

A It will go down by 30.4 points.

B It will go down by 30.4 points.

C It will go up by 60.8 points.

D The grade will remain unchanged.

E It cannot be determined from the information given


[MCQ]
#Q. Which of the following statements about principal components in Principal
Component Regression (PCR) is true ?

A Principal components are calculated based on the correlation matrix of the


original predictors.
The first principal component explains the largest proportion of the variation in
B
the dependent variable.
Principal components are linear combinations of the original predictors that are
C
uncorrelated with each other.
PCR selects the principal components with the highest p-values for inclusion in
D
the regression model.

E PCR always results in a lower model complexity compared to ordinary least


squares regression.
[MCQ]
#Q. Which statement is true about outliers in Linear regression ?

A Linear regression model is not sensitive to outliers

B Linear regression model is sensitive to outliers

C Can't say

D None of these
[MCQ]
#Q. What does the slope coefficient in a linear regression model indicate ?

A The point where the regression line intersects the y-axis

B The dependent variable changes for every one-unit change in the


independent variable
C The average value of the dependent variable

D The dispersion of the dependent variable


[MCQ]
#Q. Find the mean of squared error for the given predications :
Y F(X)
1 2
2 3
4 5
8 9
16 15
32 31
Hind : Find the squared error for each predication and take the mean of that.

A 1 B 2

C 1.5 D 0
[MCQ]
#Q. What is the primary assumption of linear regression regarding the
relationship between the independent and dependent variables?

A Non-linearity

B Independence of errors

C Homoscedasticity

D Linearity
[MCQ]
#Q. Which of the following statements is true regarding Partial Least Squares
(PLS) regression ?

A PLS is a dimensionality reduction technique that maximizes the covariance


between the predictors and the dependent variable.
PLS is only applicable when there is no multicollinearity among the
B
independent variables.
PLS can handle situations where the number of predictors is larger than the
C
number of observations.
PLS estimates the regression coefficients by minimizing the residual sum of
D
squares.

E PLS is based on the assumption of normally distributed residuals.


F All of the above.

G None of the above.


[MCQ]
#Q. The confidence interval is an interval which is an estimate of -

A The mean value of the dependent variable

B The standard deviation value of the dependent variable

C The mean value of the independent variable

D The standard deviation value of the independent variable


[MCQ]
#Q. In the regression model (y = a + bx) where x = 2.50, y = 5.50 and a = 1.50 (x
and y denote mean of variables x and y and a is a constant), which one of the
following values of parameter 'b' of the model is correct ?

A 1.75

B 1.60

C 2.00

D 2.50
[NAT]
#Q. There is no value of x that can simultaneously satisfy both the given
equations. Therefore, find the ‘least squares error’ solution to the two
equations, i.e ., find the value of x that minimize the sum of squares of the
errors in the two equations. ____________.
2x = 3
4x = 1
[MCQ]
#Q. For a bivariate data set on (x, y), if the means, standard deviations and
correlation coefficient are x = 1.0, y = 2.0, sx = 3.0, sy = 9.0, r = 0.8.
Then the regression line of y on x is:

A y = 1 + 2.4 (x − 1)

B y = 2 + 0.27 (x − 1)

C y = 2 + 2.4 (x − 1)

D y = 1 + 0.27 (x − 1)
[MCQ]
#Q. What is the purpose of regularization in linear regression ?

A To make the model more complex

B To avoid underfitting

C To encourage overfitting

D To reduce the complexity of the model


[NAT]
#Q. A set of observations of independent variable (x) and the corresponding
dependent variable (y) is given below :

X 5 2 4 3
Y 16 10 13 12

Based on the data, the coefficient a of the linear regression model.


y = a + bx is estimated as 6.1
The coefficient b is ____________ .(round off to one decimal place )
[MCQ]
#Q. The purpose of using a dummy variable in the regression model is ?

A Some of the independent variables are categorical data

B The dependent variable is categorical data

C Both independent and dependent variable may have a categorical data value

D The dependent and independent variable must be a numerical data value


[MCQ]
#Q. The random error  in multiple linear regression model y = X +  are assumed
to be identically and independently distributed following the normal
distribution with zero mean and constant variance. Here y is a n  1 vectors of
observations on response variable, X is a n  K matrix of n observations on each
of the K explanatory variables,  is a K  1 vectors of regression coefficients and
 is a n  1 vectors of random errors. The residuals ˆ = y − yˆ based on the
ordinary least squares estimator of  have , in general.
A Zero mean, constant variance and are independent

B Zero mean, constant variance and are not independent

C Zero mean, non constant variance and are not independent

D non Zero mean, non constant variance and are not independent
[MCQ]
#Q. A residual is defined as ?

A The difference between the actual Y values and the mean of Y.

B The difference between the actual Y values and the predicted Y values.

C The predicted value of Y for the average X value.

D The square root of the slope.


[MCQ]
#Q. The linear regression model y=a0+a1x1+a2x2+...+apxp is to be fitted to a set
of N training data points having p attributes each. Let X be Nx(p+1) vectors
of input values (augmented by 1's), Y be Nx1 vector of target values, and
theta (0) be (p+1)×1 vector of parameter values (a0, a1, a2,...,ap). If the sum
squared error is minimized for obtaining the optimal regression model,
which of the following equation holds ?
A XTX=XY

B XO=XTY

C XTXO=Y

D XTX0=XTY
[NAT]
#Q. Use the regression equation to predict the glucose level given the age.
Consider the following is the data set for understanding the concept of
Linear Regression Numerical Example with One Independent Variable.

SUBJECT AGE X GLUSCOSE LEVEL Y


1 43 99
2 21 65
3 25 79
4 42 75
5 57 87
6 59 81
7 55 ?
[MCQ]
#Q. In linear regression, what is the primary difference between Lasso (L1
regularization) and Ridge (L2 regularization) ?

A Lasso tends to produce sparse coefficient vectors, while Ridge does not.

B Ridge tends to produce sparse coefficient vectors, while Lasso does not.

C Both Lasso and Ridge produce sparse coefficient vectors.

D Both Lasso and Ridge tend to produce non-sparse coefficient vectors.


[NAT]
#Q. Consider the following set of points: {(-2 , -1) , (1 , 1) , (3 , 2)}
(a) Find the least square regression line for the given data points.
(b) Plot the given points and the regression line in the same rectangular
system of axes.
[NAT]
#Q. In the table below, the xi column shows scores on the aptitude test. Similarly,
the yi column shows statistics grades. The last two columns show deviations
scores - the difference between the student's score and the average score on
each measurement. The last two rows show sums and mean scores.
Find the regression equation
( xi − x ) (yi − y )
2 2
Student xi yi
1 95 85 289 64
2 85 95 49 324
3 80 70 4 49
4 70 65 64 144
5 60 70 324 49
Sum 390 385 730 630
Mean 78 77
[MCQ]
#Q. What does (x(5), y (5)) represent or imply ?

A There are 5 training examples

B The values of x and y are 5

C The fourth training examples

D The fifth training example.


[NAT]
#Q. The values of y and their corresponding values of y are shown in the table
below.

X 0 1 2 3 4
y 2 3 5 4 6

(a) Find the least square regression line y = a x + b.


(b) Estimate the value of y when x = 10.
[MCQ]
#Q. In the context of linear regression, what is the purpose of the F - test ?

A To determine the significance of individual coefficients.

B To test the overall significance of the regression model.

C To assess the presence of multicollinearity among independent variables.

D To evaluate the normality of residuals.


[MCQ]
#Q. When performing linear regression, multicollinearity can be problematic.
Which of the following statements about multicollinearity is true ?

A Multicollinearity occurs when there is no correlation between independent


variables.
B Multicollinearity makes it easier to interpret the individual coefficients in
the regression model.
C Multicollinearity inflates the standard errors of the regression coefficients.

D Multicollinearity always improves the predictive performance of the model.


[MCQ]
#Q. What is heteroscedasticity, and how does it affect the assumptions of linear
regression?

A Heteroscedasticity refers to the presence of outliers in the dataset, violating


the assumption of linearity
B Heteroscedasticity refers to the non-constant variance of residuals,
violating the assumption of homoscedasticity
C Heteroscedasticity occurs when there is perfect multicollinearity among
independent variables, violating the assumption of independence.
D Heteroscedasticity refers to the presence of correlated errors, violating
the assumption of normality.
[MCQ]
#Q. When should one prefer ridge regression over lasso regression ?

A When the goal is to select a subset of important predictors.

B When the coefficients of irrelevant predictors should be exactly zero.

C When there is multicollinearity among the independent variables.

D When the dataset has a large number of observations.


THANK - YOU

You might also like