2_syllabus
2_syllabus
MACHINE LEARNING
Course Code: AD2001-1 Course Type: IPCC
Teaching Hours/Week (L: T: P): 3:0:2 Credits: 04
Total Teaching Hours: 40+0+0 CIE + SEE Marks: 50+50
Prerequisite: AD1102-1, AD2601-1
Course Objectives:
1. Understand the need and basics of machine learning and learn the Decision Tree
model.
2. Learn ANN and Genetic Algorithms along with their applications.
3. Explore the various learning algorithms using Supervised Learning.
4. Understand the important aspects of Analytical Learning and the difference between
Analytical and Inductive Learning Algorithms.
5. Analyze the techniques related to reinforcement learning.
UNIT-I
15 Hours
Foundations of Machine Learning: What is machine learning? Applications of Machine
learning, Understand Data, Types of machine learning: Supervised, Unsupervised,
Reinforcement Learning, Theory of learning: feasibility of learning, error and noise, training
versus testing, theory of generalization, bias, and variance, learning curve, Find S-Version
Spaces and Candidate Elimination Algorithm.
Supervised Learning-I: Linear Regression: Introduction, univariate linear regression,
gradient descent, multivariate linear regression, regularized regression.
UNIT-II
15 Hours
Supervised Learning- II: Logistic regression: classification, Artificial Neural Networks,
Support Vector Machines.
Classification: Introduction, Decision Trees, Linear Discriminant Analysis, K-nearest neighbor
model, Locally Weighted Regression, Bayesian Learning, Naive Bayes Classifier, Introduction
to Hidden Markov Models and deep learning
UNIT-III
10 Hours
Unsupervised Learning: Clustering: Introduction, K-means, Hierarchical clustering
Evaluation Measures and Combining Learners: Evaluation Measures: Cross-validation and
Re-sampling, Measuring Error, Hypothesis Testing.
Reinforcement Learning: Introduction, Learning Task, Q Learning.
165
Curriculum for B.Tech. Artificial Intelligence and Data Science: 2023-27
hypothesis based on a given set of training data samples. Read the training data from
a .CSV file.
2. For a given set of training data examples stored in a .CSV file, implement and
demonstrate the Candidate-Elimination algorithm to output a description of the set of
all hypotheses consistent with the training examples.
3. Develop a program to demonstrate the working of the decision tree based ID3
algorithm. Use an appropriate data set for building the decision tree and apply this
knowledge to classify a new sample.
4. Build an Artificial Neural Network by implementing the Backpropagation algorithm and
test the same using appropriate data sets.
5. Develop a program to implement the naïve Bayesian classifier for a sample training data
set stored as a .CSV file. Compute the accuracy of the classifier, considering few test
data sets.
6. Apply EM algorithm to cluster a set of data stored in a .CSV file. Use the same data set
for clustering using k-Means algorithm. Compare the results of these two algorithms
and comment on the quality of clustering. You can add Java/Python ML library
classes/API in the program.
7. Implement and demonstrate the working of k-Nearest Neighbor algorithm and apply it
to classify the iris data set.
8. Implement the non-parametric Locally Weighted Regression algorithm in order to fit
data points. Select appropriate data set for your experiment and draw graphs.
9. Build a model to classify email as spam or ham. First, download examples of spam and
ham from Apache Spam Assassin’s public datasets and then train a model to classify
email.
PART- B
Mini Project on Machine Learning:
The main goal is to prepare students to apply machine learning algorithms to real-world
tasks, or to leave students well-qualified to start machine learning or AI research. The
mini project is intended to start in these directions.
Students shall carry out one of the following three kinds of projects:
1. Application project. Pick an interesting application and explore how best to apply
learning algorithms to solve it.
2. Algorithmic project. Pick a problem or family of problems, and develop a new
learning algorithm, or a novel variant of an existing algorithm, to solve it.
3. Theoretical project. Prove some interesting/non-trivial properties of a new or an
existing learning algorithm. (This is often quite difficult, and so very few, if any,
projects will be purely theoretical.)
166
Curriculum for B.Tech. Artificial Intelligence and Data Science: 2023-27
TEXTBOOKS:
1. T. M. Mitchell, “Machine Learning”, McGraw Hill, 2017.
2. Joel Grus, “Data Science from Scratch”, 2ndEdition, O’Reilly Publications.
REFERENCE BOOKS:
1. Ethem Alpaydin, “Introduction to Machine Learning”, Second Edition, The MIT Press,
2004.
2. C. M. Bishop, “Pattern Recognition and Machine Learning”, Springer, 2006.
3. R. O. Duda, P. E. Hart and D. G. Stork Pattern Classification, Wiley Publications, 2001.
4. T. Hastie, R. Tibshirani, J. Friedman. The Elements of Statistical Learning, 2e, 2008.
5. P. Flach, “Machine Learning: The art and science of algorithms that make sense of data”,
Cambridge University Press, 2012.
6. K. P. Murphy, “Machine Learning: A probabilistic perspective”, MIT Press, 2012.
7. M. Mohri, A. Rostamizadeh, and A. Talwalkar, “Foundations of Machine Learning”, MIT
Press, 2012.
8. S. Russel and P. Norvig, “Artificial Intelligence: A Modern Approach”, Third Edition,
Prentice Hall, 2009.
E Books / MOOCs/ NPTEL
1. https://nptel.ac.in/courses/106106139
2. https://archive.nptel.ac.in/courses/106/106/106106202/
3. https://www.coursera.org/browse/data-science/machine-learning
167