Machine Learning Techniques Unit-1 (KAI-601)
Machine Learning Techniques Unit-1 (KAI-601)
Mission
• M1 : Developing strong mathematical & computing skill set among the students.
• M2 : Extending the role of computer science and engineering in diverse areas like Internet of Things (IoT),
Artificial Intelligence & Machine Learning and Data Analytics.
• M3 : Imbibing the students with a deep understanding of professional ethics and high integrity to serve the
Nation.
• M4 : Providing an environment to the students for their growth both as individuals and as globally competent
Computer Science professional wit encouragement for innovation & start-up culture.
CO’S TITLE
CO1 Understand the key concerns that are common to all software development processes.
CO2 Select appropriate process models, approaches and techniques to manage a given software development process.
CO3 Able to elicit requirements for a software product and translate these into a documented design.
CO4 Recognize the importance of software reliability and how we can design dependable software, and what measures
are used.
CO5 Understand the principles and techniques underlying the process of inspecting and testing software and making it
free of errors and tolerable.
CO6 Understanding the latest advances and its applications in software engineering and testing.
Unit-II :
REGRESSION:
Linear Regression and Logistic Regression BAYESIAN LEARNING - Bayes theorem, Concept learning, Bayes Optimal
Classifier, Naïve Bayes classifier, Bayesian belief networks, EM algorithm. SUPPORT VECTOR MACHINE: Introduction,
Types of support vector kernel – (Linear kernel, polynomial kernel,and Gaussiankernel), Hyperplane – (Decision
surface), Properties of SVM, and Issues in SVM.
Unit-III :
DECISION TREE LEARNING - Decision tree learning algorithm, Inductive bias, Inductive inference with decision trees,
Entropy and information theory, Information gain, ID-3 Algorithm, Issues in Decision tree learning. INSTANCE-BASED
LEARNING – k-Nearest Neighbour Learning, Locally Weighted Regression, Radial basis function networks, Case-based
learning.
Subject:
Syllabus
Unit-IV :
ARTIFICIAL NEURAL NETWORKS :
Perceptron’s, Multilayer perceptron, Gradient descent and the Delta rule, Multilayer networks, Derivation of
Backpropagation Algorithm, Generalization, Unsupervised Learning – SOM Algorithm and its variant; DEEP LEARNING
- Introduction,concept of convolutional neural network , Types of layers – (Convolutional Layers , Activation function
, pooling , fully connected) , Concept of Convolution (1D and 2D) layers, Training of network, Case study of CNN for
eg on Diabetic Retinopathy, Building a smart speaker, Self-deriving car etc..
Unit-V :
REINFORCEMENT LEARNING:
Introduction to Reinforcement Learning , Learning Task,Example of Reinforcement Learning in Practice, Learning
Models for Reinforcement – (Markov Decision process , Q Learning - Q Learning function, Q Learning Algorithm ),
Application of Reinforcement Learning,Introduction to Deep Q Learning. GENETIC ALGORITHMS: Introduction,
Components, GA cycle of reproduction, Crossover, Mutation, Genetic Programming, Models of Evolution and
Learning, Applications
Subject:
Topics to be Covered
➢ Introduction to ML
➢ Types of Learning
➢ History of ML
• ML is a subset of AI
• Focus on mainly design the system and allow them to learn and make predictions based on some experiences.
• Machine learning is an application of artificial intelligence (AI) that provides systems the ability to
automatically learn and improve from experience without being explicitly programmed.
• Machine learning focuses on the development of computer programs that can access data and use it to learn
for themselves.
Subject: MLT
Introduction
Machine learning (ML) is a field of computer science that gives computers the ability to
automatically learn without being explicitly programmed.
Subject:
Types
Three types of Learning are there :- on the basis of whether or not it needs
I. Supervised Learning
II. Unsupervised Learning
III. Reinforcement Learning
Subject:
Types of ML
Supervised Learning
Example:
A computer program that learns to play checkers might improve its performance as
measured by its ability to win at the class of tasks involving playing checkers games,
through experience obtained by playing games against itself.
A checkers learning problem:
• Task T: playing checkers
• Performance measure P: percent of games won against opponents
• Training experience E: playing practice games against itself
• Information theory
• Philosophy
• Statistics
• Evaluation
V’(b) = w0+w1x1+w2x2+w3x3+w4x4+w5x5+w6x6
• 1950s-1960s
⁻ Samuel’s Algorithm on gaming
⁻ Alan turing proposed turing test
–first algorithm developed based on ideas and from psychology and
statistics.
• 1970s: Practical applications started
⁻ Symbolic concept introduction.
⁻ Image recognition but not accurate
⁻ Natural Language Processing(symbolic)
• 1980s:
⁻ Focus on experimental methodology
⁻ Neural network
2000—Advanced ML and DL
• CNN
• RNN
2001 onwards:
• AlphaGo having millions of moves.
• Image recognition
• Object detection
• Speech recognition like (siri 2010).
• Recommender system (All otp platforms)
• Medical diagnosis and prediction
• Popularity of this field in recent time and the reasons behind that
• New software /algorithms
• Neural networks
• Deep learning
• New hardware
• GPU’s
• Cloud Enabled
• Availability of big data
• Data Quality:
• ML is only as good as the data you provide it and you need a lot of data.
Accuracy of ML is driven by the quality of the data.
• Transparency
• It is often very difficult to make definitive statements on how well a model is
going to generalize in new environments.
• Manpower
• Having data and being able to use it so does not introduce bias into the
model. How organizations change how they think about software
development and how they collect and use data. Make sure they have enough
skillsets in the organization.
• What algorithms exist for learning general target functions from specific training
examples? In what settings will particular algorithms converge to the desired
function, given sufficient training data? Which algorithms perform best for which
types of problems and representations?
• How much training data is sufficient? What general bounds can be found to
relate the confidence in learned hypotheses to the amount of training experience
and the character of the learner's hypothesis space?
➢When and how can prior knowledge held by the learner guide the process of
generalizing from examples? Can prior knowledge be helpful even when it is only
approximately correct?
➢What is the best strategy for choosing a useful next training experience, and how
does the choice of this strategy alter the complexity of the learning problem?
➢What is the best way to reduce the learning task to one or more function
approximation problems? Put another way, what specific functions should the
system attempt to learn? Can this process itself be automated?
• ML The car has to recognize a stop sign using its cameras. We construct a dataset
of millions of photos of street side objects, and train an algorithm to predict
which have stop signs in them.
• AI: Once our car can recognize stop signs, it needs to decide when to take the
action of applying the brakes. It’s dangerous to apply them too early or too late,
and we need it to handle varying road conditions (for example, to recognize on a
slippery road that it is not slowing down quickly enough), which is a problem of
control theory.
Subject: Machine Learning And Techniques
Data Science Vs Machine Learning(Case Study)
• Data science: In street tests we find that the car’s performance isn’t
good enough, with some false negatives in which it drives right by a
stop sign. After analysing the street test data, we gain the insight that
the rate of false negatives depends on the time of day: it is more likely
to miss a stop sign before sunrise or after sunset. We realize that
most of our training data included only objects in full daylight, so we
construct a better dataset including night time images and go back to
the machine learning step.
• Genetic algorithms
• Genetic algorithms are stochastic search algorithms which act on a
population of possible solutions.
• Focus on optimization
• New
• New
• After the environment accepts the action, the state changes, and a reward
is given to the agent.
• Clustering:
• Termination conditions
• Several possibilities, e.g.,
• A fixed number of iterations.
• Doc partition unchanged.
• Centroid positions don’t change.
vertebrate invertebrate