Problem 3.8(a) Explanation
The Concept of Likelihood
The function is known as the likelihood function. It tells us how likely the observed data is for a specific value of the parameter .
- If we were observing a single coin flip (), the likelihood would simply be if heads () and if tails ().
- When we observe repeated independent flips, the combined probability is the product of the individual probabilities.
Derivation Logic
The core of the derivation relies on counting outcomes. Since a Bernoulli variable can only be 0 or 1:
- When , the term contributes a factor of .
- When , the term contributes a factor of .
Therefore, if we observe the sequence :
- We have three 's and two 's.
- The probability is .
- Here, and the sum .
- The number of 's is .
- The formula generalizes this counting process.
Sufficient Statistic
The quantity is called a sufficient statistic for . This means that contains all the information in the data that is relevant for estimating . Knowing the exact order of heads and tails (e.g., whether we got HHT or HTH) does not change our estimate of the bias ; only the total number of heads () and total number of flips () matter.
The result basically says the likelihood depends on the data only through .