Bayesian Learning¶
Lets compare Bayesian Learning to maximum likelihood learning.
Lets assume the following example:
We model the outcome of a biased coin \(X = \{\text{heads}, \text{tails} \}\).
We tos the coin 10 times and we observe 6 heads. The MLE estimate of the probability of observing heads is:
We toss the coin 100 times and we observe 60 heads. The MLE estimate is the same as before:
Thus we have tossed the coin 100 times but our confidence is the same.
Bayesian learning setup¶
In Bayesian learning we explicitly model the uncertainty over booth X ans \(\theta\), thus we model booth as random variables.
In bayesian learning we define a prior distribution \(p(\theta)\) that encodes our initial belief. (Subjective belief). And we update our belief based on the observed data.
Thus we can incorporate prior knowledge in our models.