- The Monte Hall Problem
- Science, Priors, and Prediction
- Statistical Models
- Binomial Data (again)
For two events A and B the conditional probability of \(A\) given \(B\) is defined as \[ P(A|B) = \frac{P(A\cap B)}{P(B)}, \] where \(A\cap B\) denotes the intersection of \(A\) and \(B\). Let \(A^c\) denote the complement of \(A\).
Then Bayes Theorem allows us to compute \(P(A|B)\) from \(P(B|A)\), \(P(B|A^c)\), and \(P(A)\) via \[ P(A|B) = \frac{P(B|A) P(A)}{P(B|A)P(A) + P(B|A^c)P(A^c)}. \]
Direct result of the definition of conditional probability (numerator) and the Law of Total Probability (denominator).
On the television show Let's Make a Deal, hosted by Monte Hall, the grand prize was awarded in the following manner. The prize was placed behind one of three doors. The contestant selected a door. Monte then showed the contestant what was behind one of the other two doors but it was never the grand prize. Finally, the contestant was allowed either to keep their initial choice or switch to the remaining unopened door.
Some people's intuition is that there is a 50/50 chance that the prize is behind either of the two remaining unopened doors, so it would not matter if you switch. In fact, the probability is 2/3 that the prize is behind the other door that Monte did not open. One intuitive way to arrive at this conclusion argues that you already know the prize is not behind one of the two doors you did not select and the fact that Monte showed you it was not behind one of them gives you no additional information. However, by switching from your initial choice, essentially you are being allowed to get both of the other two doors, and thus have a 2/3s chance of getting the prize. This argument is rather inexact, so we now give a careful argument using Bayes' Theorem.
1 | 2 | 3 | |
---|---|---|---|
f(s|1) | 0 | 0.5 | 0.5 |
f(s|2) | 0 | 0.0 | 1.0 |
f(s|3) | 0 | 1.0 | 0.0 |
## Warning: package 'tidyr' was built under R version 3.4.2
Many other items of interest cannot be computed analytically, so we discuss computer simulations of them. These include the median and the probability that \(\theta\) falls into any particular set, e.g., \(P(\theta > .5)\).
With the DAP prior, the posterior mean can be written as a weighted average of the data proportion of successes and the prior proportion of successes with weights being the relative sizes of the actual and prior data sets, that is, \[ E(\theta|y) = \left( \frac{n}{n+a+b} \right) \frac{y}{n} + \left( \frac{a+b}{n+a+b} \right) \frac{a}{a+b}. \]