-
Analyze the Priors as Beta Distributions:
The Beta distribution is given by Beta(π;α,β)∝πα−1(1−π)β−1.
- For p1(π)=2π∝π1(1−π)0, this matches a Beta distribution with α=2, β=1.
- For p0(π)=2−2π=2(1−π)∝π0(1−π)1, this matches a Beta distribution with α=1, β=2.
-
Calculate MAP Estimate for p1(π)=2π:
The posterior is p1(π∣D)∝p(D∣π)p1(π)=[πs(1−π)n−s]π=πs+1(1−π)n−s.
Take log-posterior: logp1(π∣D)=(s+1)logπ+(n−s)log(1−π)+C.
Differentiate and set to 0:
πs+1−1−πn−s=0
(s+1)(1−π)=(n−s)π⟹s+1−(s+1)π=nπ−sπ⟹(s+1)=(n+1)π
π^MAP,1=n+1s+1
-
Calculate MAP Estimate for p0(π)=2(1−π):
The posterior is p0(π∣D)∝p(D∣π)p0(π)∝[πs(1−π)n−s](1−π)=πs(1−π)n−s+1.
Take log-posterior: logp0(π∣D)=slogπ+(n−s+1)log(1−π)+C.
Differentiate and set to 0:
πs−1−πn−s+1=0
s(1−π)=(n−s+1)π⟹s=(n+1)π
π^MAP,0=n+1s
-
Effective Bayesian Estimates (Predictive Mean):
Using the posterior distributions, which are Beta(s+2,n−s+1) for p1 and Beta(s+1,n−s+2) for p0. The mean of Beta(α,β) is α+βα.
- For p1: π^Bayes,1=(s+2)+(n−s+1)s+2=n+3s+2
- For p0: π^Bayes,0=(s+1)+(n−s+2)s+1=n+3s+1
-
Intuitive Explanation (Virtual Samples):
- Prior p1(π)=2π (Beta(2,1)): The mathematical form π1(1−π)0 is equivalent to having observed 1 virtual "head" prior to the experiment. Hence, the MAP adds 1 to the numerator (heads) and 1 to the denominator (total flips). The Bayesian expectation adds 2 total virtual samples (1 head, 0 tails to the base uniform prior of 1 head, 1 tail).
- Prior p0(π)=2−2π (Beta(1,2)): The form π0(1−π)1 represents observing 1 virtual "tail" before the actual data. The MAP adds 1 to the denominator (total flips) but nothing to the numerator (heads). The Bayesian expectation reflects this by skewing the estimate downward.