Probabilistic Theory of Deep Learning
Probabilistic Theory of Deep Learning
Learning
This course explores probabilistic deep learning, a robust approach that
embraces uncertainty for more powerful and insightful models.
PRESENTED BY:
S.RANJANI-B.E.(CSE)
IIIrd year
Fundamentals of Probability
Theory
Probability theory is the foundation of probabilistic deep learning. It helps us
understand uncertainty and make informed decisions using data. We'll explore
key ideas like probability distributions, random variables, and Bayes' theorem.
Let’s say:
Data (Observed ( X ))
You flip the coin 3 times and get the following outcomes:
P(Heads) = Y
P(Tails) = 1 - Y
2.Likelihood Function
Let’s try guessing different values of ( Y ) (the probability of heads) and see
which gives us the highest log-likelihood.
From the calculations, we see that the log-likelihood is highest when ( Y = 0.6
), which means the coin is estimated to have a 60% chance of landing heads
based on our observed data.
Bayesian Inference
Bayesian inference helps us learn from data by updating our beliefs about
things we don't know. It's a powerful tool in deep learning, allowing us to build
models that are more reliable and can handle uncertainty better. We'll explore
how Bayesian inference works and how it's used in deep learning, including
Bayesian neural networks and Bayesian optimization.
1 Prior Distribution
The prior distribution represents our initial guess about the
unknown things before we look at any data. It reflects what we
already know or believe about the problem.
2 Likelihood Function
The likelihood function tells us how likely the data is, given a
particular guess about the unknowns. It measures how well
our model fits the observed data.
3 Posterior Distribution
The posterior distribution combines our initial guess (prior)
and the information from the data (likelihood) to give us a
more accurate belief about the unknowns after looking at the
data.
Imagine you have a coin that might be biased. You want to determine the
probability that the coin is biased toward heads after flipping it a few times.
Before flipping the coin, you have a belief about whether it’s fair or biased.
Let's say:
Prior belief: You think there’s a 50% chance the coin is fair (50% heads)
and a 50% chance it is biased (70% heads).
So we define:
P(Fair)=0.5
P(Biased)=0.5
Next, you flip the coin 3 times and get 2 heads and 1 tail. Now you need to
calculate how likely it is to get this result under both hypotheses.
p=0.5
p=0.7
E=Event H=Hypotheses
P(2heads)=P(2heads∣Fair)×P(Fair)+P(2heads∣Biased)×P(Biased)
P(2heads)=(0.375×0.5)+(0.441×0.5)
=0.1875+0.2205=0.408
Now we can use Bayes' Theorem to find the posterior probability that the coin
is biased given that we observed 2 heads:
P(A∣B)=P(B∣A) P(A)/P(B)
P(Biased∣2heads)=P(2heads∣Biased)×P(Biased)/P(2heads)
P(Biased∣2heads)=0.441×0.50/408
=0.22050.408≈0.540P
Step 5: Interpretation
After observing 2 heads in 3 flips, the probability that the coin is biased is
approximately 54%. This means your belief has shifted from an initial 50%
chance to a 54% chance that the coin is biased towards heads.
Optimisation and
Generalization:
Optimization
What it is: Optimization is the process of adjusting the model parameters (like
weights in a neural network) to minimize the difference between the predicted
outputs and the actual outputs (errors).
Why it matters: Effective optimization helps the model learn from the training
data, improving its performance.
In Probabilistic DL:
Generalization
Why it matters: If a model only learns the training data too well (overfitting), it
won’t be effective when faced with new examples. Good generalization means
the model can apply what it learned to different situations.
In Probabilistic DL:
Robotics
Probabilistic approaches enable Cognitive Science
robots to navigate, manipulate Probabilistic models help us
objects, and interact with humans. understand human cognition and
decision-making.