IBA Important Questions Sessional 2
IBA Important Questions Sessional 2
Prescriptive analytics is a type of data analytics that attempts to answer the question “What
do we need to do to achieve this?” It involves the use of technology to help businesses make
better decisions through the analysis of raw data.
Prescriptive analytics tries to answer the question “How do we get to this point?” It relies
on artificial intelligence (AI) techniques, such as machine learning (the ability of a computer
program without additional human input), to understand and advance from the data it
acquires, adapting all the while.
Machine learning makes it possible to process a tremendous amount of data available today.
As new or additional data becomes available, computer programs adjust automatically to
make use of it, in a process that is much faster and more comprehensive than human
capabilities could manage.
Prescriptive analytics works with another type of data analytics: predictive analytics, which
involves the use of statistics and modelling to determine future performance, based on
current and historical data. However, it goes further: Using predictive analytics’ estimation
of what is likely to happen, it recommends what future course to take.
Advantages
Prescriptive analytics can cut through the clutter of immediate uncertainty and changing
conditions. It can help prevent fraud, limit risk, increase efficiency, meet business goals, and
create more loyal customers. When used effectively, it can help organizations make
decisions based on highly analyzed facts rather than jump to under informed conclusions
based on instinct.
Prescriptive analytics can simulate the probability of various outcomes and show the
probability of each, helping organizations to better understand the level of risk and
uncertainty they face than they could be by relying on averages. Organizations that use it can
gain a better understanding of the likelihood of worst-case scenarios and plan accordingly.
Disadvantages
Prescriptive analytics is not foolproof, though. It is only effective if organizations know
what questions to ask and how to react to the answers. As such, it’s only effective if its
inputs are valid. If the input assumptions are invalid, then the output results will not be
accurate.
This form of data analytics is only suitable for short-term solutions. This means businesses
shouldn’t use prescriptive analytics to make any long-term ones. That’s because it becomes
more unreliable if more time is needed.
Not all prescriptive analytics providers are made the same. So it’s important for businesses
to carefully consider the technology and who provides it. Some may provide real, concrete
results, while others make the promise of big data and fail to deliver.
Pros
Prevents fraud, reduces risk, and increases efficiency, among other things
Cons
Example- You can take help of Decision Tree Analysis and take any examples from the
business world and build modelling around it.
• Decision trees are organized as follows: An individual makes a big decision, such as
undertaking a capital project or choosing between two competing ventures. These
decisions, which are often depicted with decision nodes, are based on the expected
outcomes of undertaking particular courses of action.
2. Differentiate between Prescriptive and Predictive Analytics.
Both predictive and prescriptive analytics leverage statistical methods and modeling to assess
future performance based on current and historical data. Predictive analytics focuses on
forecasting outcomes to help you make decisions. Prescriptive analytics takes the extra step
to recommend the specific, optimal course of action or strategy for you to take. Both types of
advanced analytics simplify complex situations and help you make better, data-driven
decisions.
Regression is a statistical method used in finance, investing, and other disciplines that
attempts to determine the strength and character of the relationship between a dependent
variable and one or more independent variables.
Linear regression is the most common form of this technique. Also called simple regression
or ordinary least squares (OLS), linear regression establishes the linear relationship between
two variables.
Understanding Regression
Regression captures the correlation between variables observed in a data set and quantifies
whether those correlations are statistically significant or not.
The two basic types of regression are simple linear regression and multiple linear regression,
although there are nonlinear regression methods for more complicated data and analysis .
Simple linear regression uses one independent variable to explain or predict the outcome of
the dependent variable Y, while multiple linear regression uses two or more independent
variables to predict the outcome. Analysts can use stepwise regression to examine each
independent variable contained in the linear regression model.
Regression can help finance and investment professionals. For instance, a company might
use it to predict sales based on weather, previous sales, gross domestic product (GDP)
growth, or other types of conditions. The capital asset pricing model (CAPM) is an often-
used regression model in finance for pricing assets and discovering the costs of capital.
Calculating Regression
Linear regression models often use a least-squares approach to determine the line of best fit.
The least-squares technique is determined by minimizing the sum of squares created by a
mathematical function. A square is, in turn, determined by squaring the distance between a
data point and the regression line or mean value of the data set.
Once this process has been completed (usually done today with software), a regression model
is constructed. The general form of each type of regression model is:
Y=a+bX+u
Additional variables such as the market capitalization of a stock, valuation ratios, and recent
returns can be added to the CAPM to get better estimates for returns. These additional
factors are known as the Fama-French factors, named after the professors who developed the
multiple linear regression model to better explain asset returns.3
Although there is some debate about the origins of the name, the statistical technique
described above most likely was termed “regression” by Sir Francis Galton in the 19th
century to describe the statistical feature of biological data (such as heights of people in a
population) to regress to some mean level. In other words, while there are shorter and taller
people, only outliers are very tall or short, and most people cluster somewhere around (or
“regress” to) the average.4
A regression model output may be in the form of Y = 1.0 + (3.2) X1 - 2.0(X2) + 0.21.
Here we have a multiple linear regression that relates some variable Y with two explanatory
variables X1 and X2. We would interpret the model as the value of Y changes by 3.2× for
every one-unit change in X 1 (if X1 goes up by 2, Y goes up by 6.4, etc.) holding all else
constant. That means controlling for X 2, X1 has this observed relationship. Likewise, holding
X1 constant, every one unit increase in X 2 is associated with a 2× decrease in Y. We can
also note the y-intercept of 1.0, meaning that Y = 1 when X 1 and X2 are both zero. The error
term (residual) is 0.21.2
Assumptions for Regression Models
To properly interpret the output of a regression model, the following main assumptions
about the underlying data process of what you are analyzing must hold:
• Residuals,
• Co-efficient of determination
• Regression Weights
• Intercepts
Residuals: Residuals are the differences between the observed values and the
predicted values of the response variable (also known as dependent variable) in a
regression analysis. They represent the error or the deviation of the predicted values
from the actual values. Mathematically, the residuals are calculated as the difference
between the observed value and the predicted value for each data point. They are used
to evaluate the performance of the regression model and to detect outliers or patterns
in the data that are not explained by the model.
Intercept: The intercept is the value of the dependent variable when all independent
variables are equal to zero. It represents the baseline value of the dependent variable
and it is one of the parameters of the regression model that is estimated from the data.
It can be interpreted as the expected value of the dependent variable when the
independent variables have no effect on it.
5. What do you understand by Time Series analysis? Explain its advantages in Business
Analytics.
“Time series analysis is a specific way of analyzing a sequence of data points collected over
an interval of time.”
In time series analysis, analysts record data points at consistent intervals over a set period of
time rather than just recording the data points intermittently or randomly. However, this type
of analysis is not merely the act of collecting data over time.
What sets time series data apart from other data is that the analysis can show how variables
change over time? In other words, time is a crucial variable because it shows how the data
adjusts over the course of the data points as well as the final results. It provides an additional
source of information and a set order of dependencies between the data.
Time series analysis typically requires a large number of data points to ensure consistency
and reliability. An extensive data set ensures you have a representative sample size and that
analysis can cut through noisy data. It also ensures that any trends or patterns discovered are
not outliers and can account for seasonal variance. Additionally, time series data can be used
for forecasting—predicting future data based on historical data.
Time series analysis helps organizations understand the underlying causes of trends or
systemic patterns over time. Using data visualizations, business users can see seasonal trends
and dig deeper into why these trends occur. With modern analytics platforms, these
visualizations can go far beyond line graphs.
When organizations analyze data over consistent intervals, they can also use time series
forecasting to predict the likelihood of future events. Time series forecasting is part
of predictive analytics. It can show likely changes in the data, like seasonality or cyclic
behavior, which provides a better understanding of data variables and helps forecast better.
For example, DPS analyzed five years of student achievement data to identify at-risk students
and track progress over time. Today’s technology allows us to collect massive amounts of
data every day and it’s easier than ever to gather enough consistent data for comprehensive
analysis.
Time series analysis is used for non-stationary data—things that are constantly fluctuating
over time or are affected by time. Industries like finance, retail, and economics frequently use
time series analysis because currency and sales are always changing. Stock market analysis is
an excellent example of time series analysis in action, especially with automated trading
algorithms. Likewise, time series analysis is ideal for forecasting weather changes, helping
meteorologists predict everything from tomorrow’s weather report to future years of climate
change. Examples of time series analysis in action include:
Weather data
Rainfall measurements
Temperature readings
Heart rate monitoring (EKG)
Quarterly sales
Stock prices
Industry forecasts
Interest rates
Because time series analysis includes many categories or variations of data, analysts
sometimes must make complex models. However, analysts can’t account for all variances,
and they can’t generalize a specific model to every sample. Models that are too complex or
that try to do too many things can lead to a lack of fit. Lack of fit or overfitting models lead to
those models not distinguishing between random error and true relationships, leaving analysis
skewed and forecasts incorrect.
Curve fitting: Plots the data along a curve to study the relationships of variables
within the data.
Descriptive analysis: Identifies patterns in time series data, like trends, cycles, or
seasonal variation.
Explanative analysis: Attempts to understand the data and the relationships within it,
as well as cause and effect.
Exploratory analysis: Highlights the main characteristics of the time series data,
usually in a visual format.
Forecasting: Predicts future data. This type is based on historical trends. It uses the
historical data as a model for future data, predicting scenarios that could happen along
future plot points.
Segmentation: Splits the data into segments to show the underlying properties of the
source information.
Factor analysis, or correlational analysis, is a statistical technique that reduces a large number
of variables into a few data sets that are more manageable and understandable. This makes it
easier to work with research data. It can be an efficient way to simplify complex data sets
with many variables. The variables can form groups of categories, called factors, based on
observations. You may determine the similarities and differences between the factors to
establish a correlation between them.
There are different methods that we use in factor analysis from the data set:
It is the most common method which the researchers use. Also, it extracts the maximum
variance and put them into the first factor. Subsequently, it removes the variance explained by
the first factor and extracts the second factor. Moreover, it goes on until the last factor.
It’s the second most favoured technique by researchers. Also, it extracts common variance and
put them into factors. Furthermore, this technique doesn’t include the variance of all variables
and is used in SEM.
3. Image Factoring
It is on the basis of the correlation matrix and makes use of OLS regression technique in order to
predict the factor in image factoring.
It also works on the correlation matrix but uses a maximum likelihood method to factor.
Alfa factoring outweighs least squares. Weight square is another regression-based method that
we use for factoring.
Factor loading- Basically it the correlation coefficient for the factors and variables. Also, it
explains the variable on a particular factor shown by variance.
Eigenvalues- Characteristics roots are its other name. Moreover, it explains the variance shown
by that particular factor out of the total variance. Furthermore, commonality column helps to
know how much variance the first factor explained out of total variance.
Factor Score- It’s another name is the component score. Besides, it’s the score of all rows and
columns that we can use as an index for all variables and for further analysis. Moreover, we can
standardize it by multiplying it with a common term.
Rotation method- This method makes it more reliable to understand the output. Also, it affects
the eigenvalues method but the eigenvalues method doesn’t affect it. Besides, there are 5
rotation methods: (1) No Rotation Method, (2) Varimax Rotation Method, (3) Quartimax
Rotation Method, (4) Direct Oblimin Rotation Method, and (5) Promax Rotation Method.
5. It is also based on the linearity assumption. So, we can also use non-linear variables.
However, after a transfer, they change into a linear variable.
Exploratory factor analysis- It assumes that any variable or indicator can be associated with any
factor. Moreover, it is the most common method used by researchers. Furthermore, it isn’t based
on any prior theory.
Confirmatory Factor Analysis- It is used to determine the factors loading and factors of
measured variables, and to confirm what it expects on the basis of pre-established assumption.
Besides, it uses two approaches:
A correlational analysis is essential because it can describe a vast data set using fewer
dimensions than the original variables. It attempts to discover the unexplained factors that
influence how similar variables form co-variations among multiple observations. It provides
more details about concepts that a single variable may not accurately measure. This can be
helpful, especially in survey research, where the responses to each question represent an
outcome. Another importance of dimension reduction is that it helps uncover clusters of
reactions, making analyzing results more accessible and faster. For example, it can help
improve customer satisfaction by getting their feedback on products.
Business analysts use this process to derive conclusions from the data of their company or
business. They may use different techniques to determine the most accurate factors for a
particular data set. This can provide them with conclusions that help a company make
effective business decisions about their products, services, and daily operations. It can also
provide them with ways to increase customer satisfaction.
7. Classify the following into predictive and prescriptive Analysis. Write of about roles of
Predictive and prescriptive analysis in Business Analytics
• Regression Analysis
• Cluster Analytics
• Econometrics
• Time Series Analysis
• Text Analytics
• Web Analytics
• Factor Analytics
• Decision Tree Analytics
Predictive Analytics techniques-
Regression Analysis
• Cluster Analytics
• Econometrics
• Time Series Analysis
• Factor Analytics
Prescriptive Analytics Techniques
• Text Analytics
• Web Analytics
• Decision Tree Analytics