This page follows on from the Bayes bars page.
In the first part of that page, we found that, in general, we can get the reverse probabilities of the boxes (BOX4, BOX2), given we have seen a red ball, by following three steps:
Get the probability of red given each box — e.g: (0.8, 0.4);
Scale these by the probability of each box — e.g.: (0.3, 0.7);
Divide the results by the sum of the results from step 2 to get the reverse probabilities — e.g. (0.462, 0.538).
We found, in that page, that if the initial probabilities of the boxes were the same — (0.5, 0.5), we could miss out the second step — scaling by the box probabilities.
Let’s see why with a little algebra mixed with code.
As you remember, the calculation works out like this:
# Step 1 - P of reds.
red_ps = np.array([0.8, 0.4])
# Step 2 - scale by P of the boxes
box_ps = np.array([0.3, 0.7])
box_and_red_probs = box_ps * red_ps
# Step 3 - divide each bar by sum of bar heights.
box_given_red_probs = box_and_red_probs / np.sum(box_and_red_probs)
# Show result
box_given_red_probs
Doing the same calculation, but removing the intermediate variables:
We can replace those numbers (0.3, 0.7 etc) with variables, and do the calculation again. The variables are to keep track of the numbers in the discussion below.
Of course we get the same answer, because we are doing the same calculation:
By working through the calculations that are happening above, we find this is the same as:
Now — what happens if the box probabilities are the same, and so b == c
?
Because b == c
, we can replace all the c
s in the calculation above with
b
s, to get:
Then we can take the b
values outside the brackets to get:
The b
s on top and bottom cancel, and so:
We have shown that we can omit the box probabilities from the calculation, when the box probabilities are equal.
You might be able to see that that this will also apply if we have three or more boxes, all with equal probabilities. That is the situation we find in the confidence in bars page.