EE353 - 769 06 Intro To ML
EE353 - 769 06 Intro To ML
Amit Sethi
Faculty member, IIT Bombay
(and regularization)
Hyper-parameters
Loss function
• Function fθ(xi)
Parameters θ
Hyper-parameters
Parameters θ
Mathematically speaking…
§ Determine f such that ti=f(xi) and g(T, X) is minimized for unseen set T
and X pairs, where T is the ground truth that cannot be used
9
Key
● Parameters: These are the variable whose values are updated during the
training process of model.
○ Feature coefficient in regression model
○ Weights of a neural network
● Hyperparameters: These are the variables/ parameter whose values are
fixed by model developer before the beginning of learning process.
○ Number of variables in a tree node
○ Height of a tree
○ Number of layers of a neural network
Type of ML problems
● Supervised learning: uses labeled data
○ Classification: Labels are discrete
○ Regression: Labels are continuous
○ Ranking: Labels are ordinal
● Unsupervised learning: uses unlabeled data
○ Clustering: Divide data into discrete groups
○ Dimension reduction: Represent data with fewer numbers
● Somewhere in between: fewer labels than one per example
○ Semi-supervised learning: some examples are labeled
○ Weakly supervised learning: groups of examples are labeled
○ Reinforcement learning: Label (reward) is available after a sequence of steps
Supervised Learning
● Predictor variables/features and a target variable (label)
● Aim: Predict the target variable (label), given the predictor variables
○ Classification: Target variable (y) consists of categories
○ Regression: Target variable is continuous
(Label)
Broad types of ML problems
Dimension
Classification Regression Clustering
reduction
Logistic Linear K-means, PCA, k-PCA,
Vector regression regression Fuzzy C-means, LLE,
SVM, RF, NN DB-SCAN ISOMAP
RNN, LSTM, Transformer,
Series, text
1-D CNN, HMM
Images 2-D CNN, MRF
Video, MRI 3-D CNN, CNN+LSTM, MRF
Recipe for ML training
• Prepare data
• Shortlist ML frameworks
• Temporal order
• Spatial order
• Web of relationships
Images courtesy: Pixabay.com
Model choice and rigorous
validation are very important
Validation
Training
Model complexity
Bias-variance trade-off
Generalization of model is bounded by the two undesirable outcomes high
bias and high variance.
● Underfitting: High bias, Low variance
● Overfitting: Low bias, High variance
Bias occurs when an algorithm has limited flexibility to learn the true signal
from the dataset. High bias can cause an algorithm to miss the relevant
relations between features and target outputs (underfitting).
Variance is an error from sensitivity to small fluctuations in the training set.
High variance can cause an algorithm to model the random noise in the
training data, rather than the intended outputs (overfitting).
Regularization is a key concept in ML
● Regularization means constraining the model
Most ML courses
Relation of ML to other fields
Artificial Intelligence
Machine Learning
Neural Networks
Deep Learning
Relation of ML to other fields
Probability
Machine and Statistics
Learning
Optimization
Programming
Linear Data Science
Algebra