0% found this document useful (0 votes)
130 views

Seminar PPT On Pca

PCA is used to reduce the dimensionality of high-dimensional data while retaining variation. It works by transforming variables into a new set of uncorrelated principal components. The first principal component accounts for as much variation in the data as possible, and each succeeding component accounts for as much remaining variation as possible. The steps are to standardize data, calculate the covariance matrix, compute eigenvectors and eigenvalues to identify principal components, and reduce the dimensions by keeping only the most informative components.

Uploaded by

prince Singh
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
130 views

Seminar PPT On Pca

PCA is used to reduce the dimensionality of high-dimensional data while retaining variation. It works by transforming variables into a new set of uncorrelated principal components. The first principal component accounts for as much variation in the data as possible, and each succeeding component accounts for as much remaining variation as possible. The steps are to standardize data, calculate the covariance matrix, compute eigenvectors and eigenvalues to identify principal components, and reduce the dimensions by keeping only the most informative components.

Uploaded by

prince Singh
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

PCA IN MACHINE

LEARNING

Presented by: Vaibhav Saran


NEED FOR PCA
➢ High Dimension Data is
extremely complex to process
due to inconsistency in the
features.
➢ Increase the computation
time and make data
processing hard.
BENEFITS OF PCA
➢ Better perspective
and less complexity.
➢ Better visualization.
➢ Size reduction.
WHAT IS PCA????
The main idea of principal component
analysis (PCA) is to reduce the
dimensionality of a data set consisting
of many variables correlated with each
other, either heavily or lightly, while
retaining the variation present in the
dataset, up to the maximum extent.
STEPS TO PERFORM PCA
STANDARDIZATION OF CALCULATE THE REDUCING THE
DATA EIGENVECTORS AND DIMENSIONS OF THE
EIGENVALUES DATA

STEP 1 STEP 3 STEP 5

STEP 2 STEP 4

COMPUTE THE COMPUTING THE


COVARIANCE MATRIX PRINCIPAL
COMPONENTS
STANDARDIZATION OF DATA
COVARIANCE MATRIX
EIGENVECTORS AND EIGENVALUES

➢ Principal Components are the new set of variables that are


obtained from the initial set of variables. They compress and
possess most of the useful information that was scattered
among the initial variables.
➢ Eigenvectors are those vectors when a linear transformation
is performed on them, then their direction does not change.
➢ Eigenvalues simple denote the scalars of the respective
eigenvectors
COMPUTING THE PRINCIPAL COMPONENTS

➢ PC1 is the most significant


and stores the maximum
possible information.
➢ PC2 is the second most
significant PC and stores
remaining maximum
information and so on.
THANK YOU

A PRESENTATION BY VAIBHAV SARAN

You might also like