SlideShare a Scribd company logo
1
Pattern Recognition
Pattern recognition is:
1. The name of the journal of the Pattern Recognition
Society.
2. A research area in which patterns in data are
found, recognized, discovered, …whatever.
3. A catchall phrase that includes
• classification
• clustering
• data mining
• ….
2
Two Schools of Thought
1. Statistical Pattern Recognition
The data is reduced to vectors of numbers
and statistical techniques are used for
the tasks to be performed.
2. Structural Pattern Recognition
The data is converted to a discrete structure
(such as a grammar or a graph) and the
techniques are related to computer science
subjects (such as parsing and graph matching).
3
In this course
1. How should objects to be classified be
represented?
2. What algorithms can be used for recognition
(or matching)?
3. How should learning (training) be done?
4
Classification in Statistical PR
• A class is a set of objects having some important
properties in common
• A feature extractor is a program that inputs the
data (image) and extracts features that can be
used in classification.
• A classifier is a program that inputs the feature
vector and assigns it to one of a set of designated
classes or to the “reject” class.
With what kinds of classes do you work?
5
Feature Vector Representation
 X=[x1, x2, … , xn],
each xj a real number
 xj may be an object
measurement
 xj may be count of
object parts
 Example: object rep.
[#holes, #strokes,
moments, …]
6
Possible features for char rec.
7
Some Terminology
 Classes: set of m known categories of objects
(a) might have a known description for each
(b) might have a set of samples for each
 Reject Class:
a generic class for objects not in any of
the designated known classes
 Classifier:
Assigns object to a class based on features
8
Discriminant functions
 Functions f(x, K)
perform some
computation on
feature vector x
 Knowledge K
from training or
programming is
used
 Final stage
determines class
9
Classification using nearest class
mean
 Compute the
Euclidean distance
between feature vector
X and the mean of
each class.
 Choose closest class,
if close enough (reject
otherwise)
10
Nearest mean might yield poor
results with complex structure
 Class 2 has two
modes; where is
its mean?
 But if modes are
detected, two
subclass mean
vectors can be
used
11
Scaling coordinates by std dev
12
Nearest Neighbor Classification
• Keep all the training samples in some efficient
look-up structure.
• Find the nearest neighbor of the feature vector
to be classified and assign the class of the neighbor.
• Can be extended to K nearest neighbors.
13
Receiver Operating Curve ROC
 Plots correct
detection rate
versus false
alarm rate
 Generally, false
alarms go up
with attempts to
detect higher
percentages of
known objects
14
Confusion matrix shows
empirical performance
15
Bayesian decision-making
16
Classifiers often used in CV
• Decision Tree Classifiers
• Artificial Neural Net Classifiers
• Bayesian Classifiers and Bayesian Networks
(Graphical Models)
• Support Vector Machines
17
Decision Trees
#holes
moment of
inertia
#strokes #strokes
best axis
direction
#strokes
- / 1 x w 0 A 8 B
0
1
2
< t  t
2 4
0 1
0
60
90
0 1
18
Decision Tree Characteristics
1. Training
How do you construct one from training data?
Entropy-based Methods
2. Strengths
Easy to Understand
3. Weaknesses
Overtraining
19
Entropy-Based Automatic
Decision Tree Construction
Node 1
What feature
should be used?
What values?
Training Set S
x1=(f11,f12,…f1m)
x2=(f21,f22, f2m)
.
.
xn=(fn1,f22, f2m)
Quinlan suggested information gain in his ID3 system
and later the gain ratio, both based on entropy.
20
Entropy
Given a set of training vectors S, if there are c classes,
Entropy(S) =  -pi log (pi)
Where pi is the proportion of category i examples in S.
i=1
c
2
If all examples belong to the same category, the entropy
is 0.
If the examples are equally mixed (1/c examples of each
class), the entropy is a maximum at 1.0.
e.g. for c=2, -.5 log .5 - .5 log .5 = -.5(-1) -.5(-1) = 1
2 2
21
Information Gain
The information gain of an attribute A is the expected
reduction in entropy caused by partitioning on this attribute.
Gain(S,A) = Entropy(S) -  ----- Entropy(Sv)
v  Values(A)
|Sv|
|S|
where Sv is the subset of S for which attribute A has
value v.
Choose the attribute A that gives the maximum
information gain.
22
Information Gain (cont)
Attribute A
v1 vk
v2
Set S
Set S 
repeat
recursively
Information gain has the disadvantage that it prefers
attributes with large number of values that split the
data into small, pure subsets.
S={sS | value(A)=v1}
23
Gain Ratio
Gain ratio is an alternative metric from Quinlan’s 1986
paper and used in the popular C4.5 package (free!).
GainRatio(S,A) = ------------------
Gain(S,a)
SplitInfo(S,A)
SplitInfo(S,A) =  - ----- log ------
|Si|
|S|
|Si|
|S|
where Si is the subset of S in which attribute A has its ith value.
2
i=1
ni
SplitInfo measures the amount of information provided
by an attribute that is not specific to the category.
24
Information Content
Note:
A related method of decision tree construction using
a measure called Information Content is given in the
text, with full numeric example of its use.
25
Artificial Neural Nets
Artificial Neural Nets (ANNs) are networks of
artificial neuron nodes, each of which computes
a simple function.
An ANN has an input layer, an output layer, and
“hidden” layers of nodes.
.
.
.
.
.
.
Inputs
Outputs
26
Node Functions
a1
a2
aj
an
output
output = g ( aj * w(j,i) )
Function g is commonly a step function, sign function,
or sigmoid function (see text).
neuron i
w(1,i)
w(j,i)
27
Neural Net Learning
That’s beyond the scope of this text; only
simple feed-forward learning is covered.
The most common method is called back propagation.
We’ve been using a free package called NevProp.
What do you use?
28
Support Vector Machines (SVM)
Support vector machines are learning algorithms
that try to find a hyperplane that separates
the differently classified data the most.
They are based on two key ideas:
• Maximum margin hyperplanes
• A kernel ‘trick’.
29
Maximal Margin
0
0
0
0
1
1
1
1
Margin
Hyperplane
Find the hyperplane with maximal margin for all
the points. This originates an optimization problem
Which has a unique solution (convex problem).
30
Non-separable data
0
0
0 0
0
0
0
0
1
1
1
1
1
0
0
1
1
1
1
0
0
What can be done if data cannot be separated with a
hyperplane?
31
The kernel trick
The SVM algorithm implicitly maps the original
data to a feature space of possibly infinite dimension
in which data (which is not separable in the
original space) becomes separable in the feature space.
0 0
0 0
0
1
1 1
Original space Rk
0
0
0
0
0
1
1
1
Feature space Rn
1
1
Kernel
trick
32
Our Current Application
• Sal Ruiz is using support vector machines in his
work on 3D object recognition.
• He is training classifiers on data representing deformations
of a 3D model of a class of objects.
• The classifiers are starting to learn what kinds of
surface patches are related to key parts of the model
(ie. A snowman’s face)
33
Snowman with Patches
Ad

More Related Content

Similar to Pattern Recognition- Basic Lecture Notes (20)

Machine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.pptMachine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.ppt
Anshika865276
 
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkkOBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkkOBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
Algorithm chapter 1
Algorithm chapter 1Algorithm chapter 1
Algorithm chapter 1
chidabdu
 
Classifiers
ClassifiersClassifiers
Classifiers
Ayurdata
 
Lect4
Lect4Lect4
Lect4
sumit621
 
机器学习Adaboost
机器学习Adaboost机器学习Adaboost
机器学习Adaboost
Shocky1
 
Anomaly detection using deep one class classifier
Anomaly detection using deep one class classifierAnomaly detection using deep one class classifier
Anomaly detection using deep one class classifier
홍배 김
 
20070702 Text Categorization
20070702 Text Categorization20070702 Text Categorization
20070702 Text Categorization
midi
 
MS CS - Selecting Machine Learning Algorithm
MS CS - Selecting Machine Learning AlgorithmMS CS - Selecting Machine Learning Algorithm
MS CS - Selecting Machine Learning Algorithm
Kaniska Mandal
 
Unit-1 Introduction and Mathematical Preliminaries.pptx
Unit-1 Introduction and Mathematical Preliminaries.pptxUnit-1 Introduction and Mathematical Preliminaries.pptx
Unit-1 Introduction and Mathematical Preliminaries.pptx
avinashBajpayee1
 
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
csandit
 
Analytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion miningAnalytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion mining
csandit
 
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
cscpconf
 
Lecture 2
Lecture 2Lecture 2
Lecture 2
butest
 
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Universitat Politècnica de Catalunya
 
Machine learning and Neural Networks
Machine learning and Neural NetworksMachine learning and Neural Networks
Machine learning and Neural Networks
butest
 
Islamic University Pattern Recognition & Neural Network 2019
Islamic University Pattern Recognition & Neural Network 2019 Islamic University Pattern Recognition & Neural Network 2019
Islamic University Pattern Recognition & Neural Network 2019
Rakibul Hasan Pranto
 
Yulia Honcharenko "Application of metric learning for logo recognition"
Yulia Honcharenko "Application of metric learning for logo recognition"Yulia Honcharenko "Application of metric learning for logo recognition"
Yulia Honcharenko "Application of metric learning for logo recognition"
Fwdays
 
8.clustering algorithm.k means.em algorithm
8.clustering algorithm.k means.em algorithm8.clustering algorithm.k means.em algorithm
8.clustering algorithm.k means.em algorithm
Laura Petrosanu
 
Machine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.pptMachine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.ppt
Anshika865276
 
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkkOBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkkOBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
OBJECTRECOGNITION1.pptxjjjkkkkjjjjkkkkkkk
shesnasuneer
 
Algorithm chapter 1
Algorithm chapter 1Algorithm chapter 1
Algorithm chapter 1
chidabdu
 
Classifiers
ClassifiersClassifiers
Classifiers
Ayurdata
 
机器学习Adaboost
机器学习Adaboost机器学习Adaboost
机器学习Adaboost
Shocky1
 
Anomaly detection using deep one class classifier
Anomaly detection using deep one class classifierAnomaly detection using deep one class classifier
Anomaly detection using deep one class classifier
홍배 김
 
20070702 Text Categorization
20070702 Text Categorization20070702 Text Categorization
20070702 Text Categorization
midi
 
MS CS - Selecting Machine Learning Algorithm
MS CS - Selecting Machine Learning AlgorithmMS CS - Selecting Machine Learning Algorithm
MS CS - Selecting Machine Learning Algorithm
Kaniska Mandal
 
Unit-1 Introduction and Mathematical Preliminaries.pptx
Unit-1 Introduction and Mathematical Preliminaries.pptxUnit-1 Introduction and Mathematical Preliminaries.pptx
Unit-1 Introduction and Mathematical Preliminaries.pptx
avinashBajpayee1
 
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
csandit
 
Analytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion miningAnalytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion mining
csandit
 
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
cscpconf
 
Lecture 2
Lecture 2Lecture 2
Lecture 2
butest
 
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Universitat Politècnica de Catalunya
 
Machine learning and Neural Networks
Machine learning and Neural NetworksMachine learning and Neural Networks
Machine learning and Neural Networks
butest
 
Islamic University Pattern Recognition & Neural Network 2019
Islamic University Pattern Recognition & Neural Network 2019 Islamic University Pattern Recognition & Neural Network 2019
Islamic University Pattern Recognition & Neural Network 2019
Rakibul Hasan Pranto
 
Yulia Honcharenko "Application of metric learning for logo recognition"
Yulia Honcharenko "Application of metric learning for logo recognition"Yulia Honcharenko "Application of metric learning for logo recognition"
Yulia Honcharenko "Application of metric learning for logo recognition"
Fwdays
 
8.clustering algorithm.k means.em algorithm
8.clustering algorithm.k means.em algorithm8.clustering algorithm.k means.em algorithm
8.clustering algorithm.k means.em algorithm
Laura Petrosanu
 

Recently uploaded (20)

Surveying through global positioning system
Surveying through global positioning systemSurveying through global positioning system
Surveying through global positioning system
opneptune5
 
Artificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptxArtificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptx
rakshanatarajan005
 
Water Industry Process Automation & Control Monthly May 2025
Water Industry Process Automation & Control Monthly May 2025Water Industry Process Automation & Control Monthly May 2025
Water Industry Process Automation & Control Monthly May 2025
Water Industry Process Automation & Control
 
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
ajayrm685
 
Redirects Unraveled: From Lost Links to Rickrolls
Redirects Unraveled: From Lost Links to RickrollsRedirects Unraveled: From Lost Links to Rickrolls
Redirects Unraveled: From Lost Links to Rickrolls
Kritika Garg
 
C_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdf
C_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdfC_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdf
C_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdf
amanpathak160605
 
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...
IJCNCJournal
 
PRIZ Academy - Functional Modeling In Action with PRIZ.pdf
PRIZ Academy - Functional Modeling In Action with PRIZ.pdfPRIZ Academy - Functional Modeling In Action with PRIZ.pdf
PRIZ Academy - Functional Modeling In Action with PRIZ.pdf
PRIZ Guru
 
Machine Learning basics POWERPOINT PRESENETATION
Machine Learning basics POWERPOINT PRESENETATIONMachine Learning basics POWERPOINT PRESENETATION
Machine Learning basics POWERPOINT PRESENETATION
DarrinBright1
 
Novel Plug Flow Reactor with Recycle For Growth Control
Novel Plug Flow Reactor with Recycle For Growth ControlNovel Plug Flow Reactor with Recycle For Growth Control
Novel Plug Flow Reactor with Recycle For Growth Control
Chris Harding
 
Jacob Murphy Australia - Excels In Optimizing Software Applications
Jacob Murphy Australia - Excels In Optimizing Software ApplicationsJacob Murphy Australia - Excels In Optimizing Software Applications
Jacob Murphy Australia - Excels In Optimizing Software Applications
Jacob Murphy Australia
 
Applications of Centroid in Structural Engineering
Applications of Centroid in Structural EngineeringApplications of Centroid in Structural Engineering
Applications of Centroid in Structural Engineering
suvrojyotihalder2006
 
ML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdf
ML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdfML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdf
ML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdf
rameshwarchintamani
 
Autodesk Fusion 2025 Tutorial: User Interface
Autodesk Fusion 2025 Tutorial: User InterfaceAutodesk Fusion 2025 Tutorial: User Interface
Autodesk Fusion 2025 Tutorial: User Interface
Atif Razi
 
NOMA analysis in 5G communication systems
NOMA analysis in 5G communication systemsNOMA analysis in 5G communication systems
NOMA analysis in 5G communication systems
waleedali330654
 
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
ijflsjournal087
 
Analog electronic circuits with some imp
Analog electronic circuits with some impAnalog electronic circuits with some imp
Analog electronic circuits with some imp
KarthikTG7
 
Interfacing PMW3901 Optical Flow Sensor with ESP32
Interfacing PMW3901 Optical Flow Sensor with ESP32Interfacing PMW3901 Optical Flow Sensor with ESP32
Interfacing PMW3901 Optical Flow Sensor with ESP32
CircuitDigest
 
ZJIT: Building a Next Generation Ruby JIT
ZJIT: Building a Next Generation Ruby JITZJIT: Building a Next Generation Ruby JIT
ZJIT: Building a Next Generation Ruby JIT
maximechevalierboisv1
 
Evonik Overview Visiomer Specialty Methacrylates.pdf
Evonik Overview Visiomer Specialty Methacrylates.pdfEvonik Overview Visiomer Specialty Methacrylates.pdf
Evonik Overview Visiomer Specialty Methacrylates.pdf
szhang13
 
Surveying through global positioning system
Surveying through global positioning systemSurveying through global positioning system
Surveying through global positioning system
opneptune5
 
Artificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptxArtificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptx
rakshanatarajan005
 
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
ajayrm685
 
Redirects Unraveled: From Lost Links to Rickrolls
Redirects Unraveled: From Lost Links to RickrollsRedirects Unraveled: From Lost Links to Rickrolls
Redirects Unraveled: From Lost Links to Rickrolls
Kritika Garg
 
C_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdf
C_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdfC_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdf
C_Dayyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyy 3.pdf
amanpathak160605
 
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...
IJCNCJournal
 
PRIZ Academy - Functional Modeling In Action with PRIZ.pdf
PRIZ Academy - Functional Modeling In Action with PRIZ.pdfPRIZ Academy - Functional Modeling In Action with PRIZ.pdf
PRIZ Academy - Functional Modeling In Action with PRIZ.pdf
PRIZ Guru
 
Machine Learning basics POWERPOINT PRESENETATION
Machine Learning basics POWERPOINT PRESENETATIONMachine Learning basics POWERPOINT PRESENETATION
Machine Learning basics POWERPOINT PRESENETATION
DarrinBright1
 
Novel Plug Flow Reactor with Recycle For Growth Control
Novel Plug Flow Reactor with Recycle For Growth ControlNovel Plug Flow Reactor with Recycle For Growth Control
Novel Plug Flow Reactor with Recycle For Growth Control
Chris Harding
 
Jacob Murphy Australia - Excels In Optimizing Software Applications
Jacob Murphy Australia - Excels In Optimizing Software ApplicationsJacob Murphy Australia - Excels In Optimizing Software Applications
Jacob Murphy Australia - Excels In Optimizing Software Applications
Jacob Murphy Australia
 
Applications of Centroid in Structural Engineering
Applications of Centroid in Structural EngineeringApplications of Centroid in Structural Engineering
Applications of Centroid in Structural Engineering
suvrojyotihalder2006
 
ML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdf
ML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdfML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdf
ML_Unit_V_RDC_ASSOCIATION AND DIMENSIONALITY REDUCTION.pdf
rameshwarchintamani
 
Autodesk Fusion 2025 Tutorial: User Interface
Autodesk Fusion 2025 Tutorial: User InterfaceAutodesk Fusion 2025 Tutorial: User Interface
Autodesk Fusion 2025 Tutorial: User Interface
Atif Razi
 
NOMA analysis in 5G communication systems
NOMA analysis in 5G communication systemsNOMA analysis in 5G communication systems
NOMA analysis in 5G communication systems
waleedali330654
 
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
ijflsjournal087
 
Analog electronic circuits with some imp
Analog electronic circuits with some impAnalog electronic circuits with some imp
Analog electronic circuits with some imp
KarthikTG7
 
Interfacing PMW3901 Optical Flow Sensor with ESP32
Interfacing PMW3901 Optical Flow Sensor with ESP32Interfacing PMW3901 Optical Flow Sensor with ESP32
Interfacing PMW3901 Optical Flow Sensor with ESP32
CircuitDigest
 
ZJIT: Building a Next Generation Ruby JIT
ZJIT: Building a Next Generation Ruby JITZJIT: Building a Next Generation Ruby JIT
ZJIT: Building a Next Generation Ruby JIT
maximechevalierboisv1
 
Evonik Overview Visiomer Specialty Methacrylates.pdf
Evonik Overview Visiomer Specialty Methacrylates.pdfEvonik Overview Visiomer Specialty Methacrylates.pdf
Evonik Overview Visiomer Specialty Methacrylates.pdf
szhang13
 
Ad

Pattern Recognition- Basic Lecture Notes

  • 1. 1 Pattern Recognition Pattern recognition is: 1. The name of the journal of the Pattern Recognition Society. 2. A research area in which patterns in data are found, recognized, discovered, …whatever. 3. A catchall phrase that includes • classification • clustering • data mining • ….
  • 2. 2 Two Schools of Thought 1. Statistical Pattern Recognition The data is reduced to vectors of numbers and statistical techniques are used for the tasks to be performed. 2. Structural Pattern Recognition The data is converted to a discrete structure (such as a grammar or a graph) and the techniques are related to computer science subjects (such as parsing and graph matching).
  • 3. 3 In this course 1. How should objects to be classified be represented? 2. What algorithms can be used for recognition (or matching)? 3. How should learning (training) be done?
  • 4. 4 Classification in Statistical PR • A class is a set of objects having some important properties in common • A feature extractor is a program that inputs the data (image) and extracts features that can be used in classification. • A classifier is a program that inputs the feature vector and assigns it to one of a set of designated classes or to the “reject” class. With what kinds of classes do you work?
  • 5. 5 Feature Vector Representation  X=[x1, x2, … , xn], each xj a real number  xj may be an object measurement  xj may be count of object parts  Example: object rep. [#holes, #strokes, moments, …]
  • 7. 7 Some Terminology  Classes: set of m known categories of objects (a) might have a known description for each (b) might have a set of samples for each  Reject Class: a generic class for objects not in any of the designated known classes  Classifier: Assigns object to a class based on features
  • 8. 8 Discriminant functions  Functions f(x, K) perform some computation on feature vector x  Knowledge K from training or programming is used  Final stage determines class
  • 9. 9 Classification using nearest class mean  Compute the Euclidean distance between feature vector X and the mean of each class.  Choose closest class, if close enough (reject otherwise)
  • 10. 10 Nearest mean might yield poor results with complex structure  Class 2 has two modes; where is its mean?  But if modes are detected, two subclass mean vectors can be used
  • 12. 12 Nearest Neighbor Classification • Keep all the training samples in some efficient look-up structure. • Find the nearest neighbor of the feature vector to be classified and assign the class of the neighbor. • Can be extended to K nearest neighbors.
  • 13. 13 Receiver Operating Curve ROC  Plots correct detection rate versus false alarm rate  Generally, false alarms go up with attempts to detect higher percentages of known objects
  • 16. 16 Classifiers often used in CV • Decision Tree Classifiers • Artificial Neural Net Classifiers • Bayesian Classifiers and Bayesian Networks (Graphical Models) • Support Vector Machines
  • 17. 17 Decision Trees #holes moment of inertia #strokes #strokes best axis direction #strokes - / 1 x w 0 A 8 B 0 1 2 < t  t 2 4 0 1 0 60 90 0 1
  • 18. 18 Decision Tree Characteristics 1. Training How do you construct one from training data? Entropy-based Methods 2. Strengths Easy to Understand 3. Weaknesses Overtraining
  • 19. 19 Entropy-Based Automatic Decision Tree Construction Node 1 What feature should be used? What values? Training Set S x1=(f11,f12,…f1m) x2=(f21,f22, f2m) . . xn=(fn1,f22, f2m) Quinlan suggested information gain in his ID3 system and later the gain ratio, both based on entropy.
  • 20. 20 Entropy Given a set of training vectors S, if there are c classes, Entropy(S) =  -pi log (pi) Where pi is the proportion of category i examples in S. i=1 c 2 If all examples belong to the same category, the entropy is 0. If the examples are equally mixed (1/c examples of each class), the entropy is a maximum at 1.0. e.g. for c=2, -.5 log .5 - .5 log .5 = -.5(-1) -.5(-1) = 1 2 2
  • 21. 21 Information Gain The information gain of an attribute A is the expected reduction in entropy caused by partitioning on this attribute. Gain(S,A) = Entropy(S) -  ----- Entropy(Sv) v  Values(A) |Sv| |S| where Sv is the subset of S for which attribute A has value v. Choose the attribute A that gives the maximum information gain.
  • 22. 22 Information Gain (cont) Attribute A v1 vk v2 Set S Set S  repeat recursively Information gain has the disadvantage that it prefers attributes with large number of values that split the data into small, pure subsets. S={sS | value(A)=v1}
  • 23. 23 Gain Ratio Gain ratio is an alternative metric from Quinlan’s 1986 paper and used in the popular C4.5 package (free!). GainRatio(S,A) = ------------------ Gain(S,a) SplitInfo(S,A) SplitInfo(S,A) =  - ----- log ------ |Si| |S| |Si| |S| where Si is the subset of S in which attribute A has its ith value. 2 i=1 ni SplitInfo measures the amount of information provided by an attribute that is not specific to the category.
  • 24. 24 Information Content Note: A related method of decision tree construction using a measure called Information Content is given in the text, with full numeric example of its use.
  • 25. 25 Artificial Neural Nets Artificial Neural Nets (ANNs) are networks of artificial neuron nodes, each of which computes a simple function. An ANN has an input layer, an output layer, and “hidden” layers of nodes. . . . . . . Inputs Outputs
  • 26. 26 Node Functions a1 a2 aj an output output = g ( aj * w(j,i) ) Function g is commonly a step function, sign function, or sigmoid function (see text). neuron i w(1,i) w(j,i)
  • 27. 27 Neural Net Learning That’s beyond the scope of this text; only simple feed-forward learning is covered. The most common method is called back propagation. We’ve been using a free package called NevProp. What do you use?
  • 28. 28 Support Vector Machines (SVM) Support vector machines are learning algorithms that try to find a hyperplane that separates the differently classified data the most. They are based on two key ideas: • Maximum margin hyperplanes • A kernel ‘trick’.
  • 29. 29 Maximal Margin 0 0 0 0 1 1 1 1 Margin Hyperplane Find the hyperplane with maximal margin for all the points. This originates an optimization problem Which has a unique solution (convex problem).
  • 30. 30 Non-separable data 0 0 0 0 0 0 0 0 1 1 1 1 1 0 0 1 1 1 1 0 0 What can be done if data cannot be separated with a hyperplane?
  • 31. 31 The kernel trick The SVM algorithm implicitly maps the original data to a feature space of possibly infinite dimension in which data (which is not separable in the original space) becomes separable in the feature space. 0 0 0 0 0 1 1 1 Original space Rk 0 0 0 0 0 1 1 1 Feature space Rn 1 1 Kernel trick
  • 32. 32 Our Current Application • Sal Ruiz is using support vector machines in his work on 3D object recognition. • He is training classifiers on data representing deformations of a 3D model of a class of objects. • The classifiers are starting to learn what kinds of surface patches are related to key parts of the model (ie. A snowman’s face)