Problem 3.8(d) Explanation
Mode vs Mean
- MAP estimates the Mode of the posterior.
- Bayesian Prediction (from part c) uses the Mean of the posterior.
For the Beta distribution :
- Mode =
- Mean =
With Uniform Priors () and data (): The posterior is .
MAP (Mode):
(This is only defined when counts are > 1, strictly speaking, but the limit holds).
Bayes Estimator (Mean):
Why MAP equals ML here?
MAP is ML times Prior. If Prior is flat (multiplication by 1), the "hill" in the landscape is defined entirely by the likelihood. So the peak (Mode) is at the same spot.
Practical Implication
In Machine Learning, we often prefer the Bayesian Mean (or smoothed estimates) because predicting exactly 0 or 1 is dangerous. If you estimate probability 0 for an event, and it happens, your error (log loss) is infinite. The Bayesian estimate naturally safeguards against this by integrating over uncertainty.