ITML U1 Overview
ITML U1 Overview
Dr. N. Kalyani
Professor
Text Book
• Machine Learning, Tom T. Mitchell, McGraw-Hill, 2013
• Ethem Alpaydin, Introduction to Machine Learning, 2nd Edition, MIT
Press, 2010.
Course Objectives
1. To introduce students to the basic concepts of Machine Learning.
2. To become familiar with regression and SVM.
3. To study the Decision tree and Bayesian Classifiers.
4. To understand instance-based learning and clustering techniques
5. To learn Genetic algorithms and reinforcement learning.
Syllabus
UNIT 1:
Introduction to Machine Learning: What is Machine Learning, Examples of Machine
Learning Applications, Types of Machine Learning systems-Supervised learning,
Unsupervised learning, Reinforcement learning, Learning a class from examples.
Concept learning and the general to specific ordering – Introduction, A concept
learning task, Concept learning as search, find-S, Version spaces and the candidate
elimination algorithm.
UNIT 2:
Parametric methods: Introduction, Maximum Likelihood Estimation, Evaluating an
Estimator: Bias and Variance, The Bayes‟ Estimator, Parametric Classification,
Regression.
Support vector machines: Introduction, Optimal hyperplane for linearly separable
patterns, Quadratic optimization for finding the optimal hyperplane, Statistical
properties of the optimal hyperplane, Optimal hyperplane for non-separable patterns,
SVM non-linear regression.
Syllabus
UNIT 3:
Bayesian Learning: Introduction, Bayes theorem, Bayes theorem and concept learning, Bayes
optimal classifier, Naïve Bayes classifier.
Decision Tree Learning: Introduction, Decision tree representation, Appropriate problems for
decision tree learning, The basic decision tree learning algorithm, Hypothesis space search in
decision tree learning, Inductive bias in decision tree learning, Issues in decision tree learning.
UNIT 4:
Instance-based learning: Introduction, KNN learning, Distance weighted NN algorithm,
Remarks on KNN Algorithm.
Unsupervised learning: Introduction, K-means clustering technique, Hierarchical clustering,
Choosing the number of clusters.
UNIT 5:
Reinforcement Learning: Introduction, The learning task, Q learning, Nondeterministic rewards
and actions.
Genetic Algorithm: Motivation, Genetic algorithms, An illustrative example, Hypothesis space
search, Genetic programming.
Outline
• Machine Learning Overview
• Learning Methods
• Types of Machine Learning
Machine Learning Overview
Learner Domain
Performance Goal
Machine
Success Criteria Learning Representation
Rote Learning Memorizing, Save Knowledge, Association based storage, Repeated computation
Issue : Organization, Generalization, Stability
Learning by Methods: Develop sophisticated tools, generating rules from High level abstraction
advice
Process: Request/enquire, Interpret, Operationalize, Integrate, Evaluate
Learning by
parameter Define static evaluation functions using the significant attributes that define the output
adjustment Process: Initialize, modify the weights using input samples
Learning by Identify the similarities and dissimilarities within the data that is not labeled.
Analogy
Types: Transformational Analogy, Derivational Analogy
Learning by
Macro Same as Rote learner
operations Avoid expensive re computation by using macro-operation
8
Types of Learning
Learning
General
h3
h1 h2
Specific
h1=<Round,?,Square,?,?>
x1=<Round,Square,Square,Purple,Yes>
h2=<Round,?,?,?,Yes>
x2=<Round,Square,Round,Green,Yes>
h3=<Round,?,?,?,?>
Hypothesis Space Search by Find-
SInstances X Hypotheses Hh 5
x
2
General
x x h 3,4
1 3
x
5
x
h1,2
4 Specific
h0
h0=< >
x1=<Round,Triangle,Round,Purple,Yes> + h1=<Round,Triangle,Round,Purple,Yes>
x2=<Square,Square,Square,Green,Yes> - h2=<Round,Triangle,Round,Purple,Yes>
x3=<Square,Triangle,Round,Yellow,Yes> + h3=<?,Triangle,Round,?,Yes>
x4=<Round,Triangle,Round,Green,No> - h4=<?,Triangle,Round,?,Yes>
x5=<Square,Square,Round,Yellow,Yes> + h5=<?,?,Round,?,Yes>
General to Specific
ordering
H1 = {Round, ? ,Round, Purple, Yes }
H2 = {?, ?, Round, ?, Yes }
Definition: Let hi and hj be Boolean-valued functions defined over X.
Then hi is more general than or equal to hj
(written hi ≥g hj) if and only if (x X)[(hj(x)=1) →(hi(x)=1)
(Strictly) more general than
(written hi g hj) if and only if (hi ≥g hj) (hj ≥g hi)
Find-SAlgorithm
1. Initialize h to the most specific hypothesis in H
2. For each positive training instance x
For each attribute constraint ai in h
IF the constraint ai in h is satisfied by x THEN do nothing
ELSE
replace ai in h by next more general constraint satisfied by x
3. Output hypothesis h
Training Examples - 2 EnjoySport
23
Representing Hypotheses
Sky Temp Humid Wind Water Forecst EnjoySport
Sunny Warm Normal Strong Warm Same Yes
Sunny Warm High Strong Warm Same Yes
Rainy Cold High Strong Warm Change No
Sunny Warm High Strong Cool Change Yes
S: { <?,Triangle,Round,?,Yes> }
Candidate Elimination Algorithm
1/2
G = maximally general hypotheses in H
S = maximally specific hypotheses in H
S: { <?,Triangle,Round,?,Yes> }
?
Example to find version space
37
What Next Training Example?
38
Remarks on Version Space and Candidate Elimination
39
Inductive Bias
{{{[
3. Find-S
Summary Points
1. Concept learning as search through H
2. General-to-specific ordering over H
3. Version space candidate elimination algorithm
4. S and G boundaries characterize learner’s
uncertainty
5. Learner can generate useful queries
6. Inductive learning possible only if learner is
unbiased
7. Inductive learners can be modeled by equivalent
deductive systems