Unit 1 DLT
Unit 1 DLT
Fundamentals
Deep learning is a powerful branch of artificial intelligence that enables
machines to learn from data and make predictions. This presentation delves into
the fundamental concepts of deep learning, covering essential mathematical
foundations, probability distributions, optimization techniques, and key
challenges. We will explore the core data structures, the role of probability in
machine learning, and the importance of optimization algorithms for training
deep learning models.
Linear Algebra: The Foundation of Deep Learning
Scalars Vectors Matrices and Tensors
Scalars are single numbers, representing Vectors are ordered arrays of single Matrices are rectangular arrays of
0th-order tensors. They are denoted by x numbers, representing 1st-order tensors. numbers, representing 2nd-order tensors.
∈ ℝ, indicating a scalar belonging to the They are fragments of vector spaces, Tensors generalize scalars, vectors, and
set of real numbers. which encompass all possible vectors of a matrices, encompassing higher-order
specific length or dimension. entities.
Probability Distributions:
Describing Random
Variables
1 Discrete Variables 2 Continuous Variables
Discrete variables can take on Continuous variables can take
a finite number of values. on any value within a range.
Their probability distributions Their probability distributions
are described using are described using
probability mass functions probability density functions
(PMFs). (PDFs).
Gradient-Based Optimization: Minimizing Loss Functions
1 2 3 4
Underfitting
Underfitting occurs when a model is too simple to capture the underlying patterns in
the data, resulting in poor performance on both training and test data.
Estimators: Quantifying
Guesses in Machine
Learning
Estimator Description
High Bias, Low Variance High Bias, High Variance Bias-Variance Trade-Off
Predictions are consistent but inaccurate on Predictions are inconsistent and inaccurate on Finding a balance between bias and variance is
average, indicating underfitting. average, indicating a model that is not learning crucial for building accurate machine learning
well. models. A model with low bias and low variance
is ideal.
Stochastic Gradient Descent (SGD): Optimizing for
Large Datasets