Chapter 1 - Part1
Chapter 1 - Part1
goal
Safe driving
Comfortable
Reaching on time
Problem taken from Russell Norvig
edition 4
Full joint distribution
,
Summation is always over hidden variables
Summation moves inwards
P(j/A) is independent of P(M/A) so Summation M goes in
Like wise sum of B is independent of P(J/A) and P(M/A) so it moves in
=f4(J)
f1 function is of A and B as summation over E will remove the
variable E as it will be summed up. similarly for f2 and f3
Final step
Inference in Bayesian Network
This is a better model for this dataset but this is definitely not a gaussian distribution as
gaussian will have only one peak,this has 3 peaks.
So we can conclude, this type of data set requires a different model to fit into
Also we can see the data exhibits some kind of clustering.
A new generative model which is a mixture of gaussian will better explains the data,
mechanism which will define the data in a probabilistic way.
Way to model
for all i
Every mixture has its own mean and variance
Say Zi is 5 that means I am going to the fifth mixture and there is mean
Step 2:
Step1
Step2
Latent or unobserved variable
Gaussian Mixture model is a latent variable model because the output
that we assuming not only depends on some parameters that you
want to estimate but also on unobserved variable
What we need to pause and think?
Parameters ]
,)
Total : 2k +k-1
3K-1
The likelihood function is dependent on various parameters µ,sigma
and probability π
As the first step is to select which mixture
The selection of mixture depends on π
Here big pi is product as all events are iids
The data point will always not be closer to the mean as it depends on
the probability of the mixture being picked. if that probability itself is
very low then this may lead to the data point being picked up by
other cluster
Introduction to GMM
If the linear interpolation always gives a value higher than the actual
function itself then such functions are called convex functions