Decision theory is not one of my strengths, and I have a question about it.
Is there a consensus view on how to deal with the problem of "rival formalizations"? Peterson (2009) illustrates the problem like this:
Imagine that you are a paparazzi photographer and that rumour has it that actress Julia Roberts will show up in either New York (NY), Los Angeles (LA) or Paris (P). Nothing is known about the probability of these states of the world. You have to decide if you should stay in America or catch a plane to Paris. If you stay and [she] shows up in Paris you get $0; otherwise you get your photos, which you will be able to sell for $10,000. If you catch a plane to Paris and Julia Roberts shows up in Paris your net gain after having paid for the ticket is $5,000, and if she shows up in America you for some reason, never mind why, get $6,000. Your initial representation of the decision problem is visualized in Table 2.13.
Table 2.13
P | LA | NY | |
Stay | $0 | $10k | $10k |
Go to Paris | $5k | $6k | $6k |
Since nothing is known about the probabilities of the states in Table 2.13, you decide it makes sense to regard them as equally probable [see Table 2.14].
Table 2.14
P (1/3) | LA (1/3) | NY (1/3) | |
Stay | $0 | $10k | $10k |
Go to Paris | $5k | $6k | $6k |
The rightmost columns are exactly parallel. Therefore, they can be merged into a single (disjuntive) column, by adding the probabilities of the two rightmost columns together (Table 2.15).
Table 2.15
P (1/3) | LA or NY (2/3) | |
Stay | $0 | $10k |
Go to Paris | $5k | $6k |
However, now suppose that you instead start with Table 2.13 and first merge the two repetitious states into a single state. You would then obtain the decision matrix in Table 2.16.
Table 2.16
P | LA or NY | |
Stay | $0 | $10k |
Go to Paris | $5k | $6k |
Now, since you know nothing about the probabilities of the two states, you decide to regard them as equally probable... This yields the formal representation in Table 2.17, which is clearly different from the one suggested above in Table 2.15.
Table 2.17
P (1/2) | LA or NY (1/2) | |
Stay | $0 | $10k |
Go to Paris | $5k | $6k |
Which formalisation is best, 2.15 or 2.17? It seems question begging to claim that one of them must be better than the other — so perhaps they are equally reasonable? If they are, we have an example of rival formalisations.
Note that the principle of maximising expected value recommends different acts in the two matrices. According to Table 2.15 you should stay, but 2.17 suggests you should go to Paris.
Does anyone know how to solve this problem? If one is not convinced by the illustration above, Peterson (2009) offers a proof that rival representations are possible on pages 33–35.
It has been super interesting to read all your contributions to lukeprog's post; this 'paradox' is no doubt interesting, because there seemed to be a shared gut reaction as to something wrong about the above formulation. I have stumbled across this page with the exact dilemma as lukeprog while reading Peterson's Introduction to Decision Theory book (2nd edition). As you have all pointed out, it would seem that there is something inherently fishy about his formulation of this particular example of 'Rival Formalisations'.
I think if you follow the logic of the initial axioms he uses in the book, this example that he provides does not follow. To give you some context, he formulates this 'paradox' by invoking two 'axioms'; the principle of insufficient reason (ir) and merger of states (ms). These principles are as follows (Peterson 2009 page 35):
From these two principles he first applies the IR rule and then the MS rule to formulate the 'paradox' above (Peterson 2009 page 35).
As per the above post by lukeprog, Peterson, in his (1/3, 1/3, 1/3) (P, LA, NY) example, insists that the probabilities of LA and NY can be added together to make 2/3rds and that this is a correct application of IR and MS.
The principle of IR would instead contradict this application, as he adds the 1/3 probabilities from NY and LA as if these probabilities are apriori known. Contrary to Peterson, they are not apriori known. That's why IR was invoked in the first place. When evoking MS, it requires that the probabilities of the states to be known in order for the probabilities to be added. From IR, we know that these probabilities have, rather, been arbitrarily assigned into equal proportions, precisely because the probabilities of these states (P, NY, LA) in question are apriori unknown. It should not follow from this that the probabilities of NY can be added to LA. To do so is to suggests that the probabilities are apriori known and unknown at the same time, which is a contradiction.
A good question one could ask is what the difference between 'collapsing' states and adding probabilities, and whether it affects the above analysis. Much like the 'Sure-Thing Principle', states with identical outcomes are collapsable into one because the probability component is irrelevant; regardless of the likelihood of each state is, the outcomes are the same. I think that is why, in this example, collapsing NY/LA into (NY or LA) is permissible but adding probabilities without an apriori known origin is not. This would suggest that LA and NY should first collapse into one state because of its identical outcomes, and then as a result of the apriori unknown (and unknowable) probabilities of states P and (LA or NY), should 1/2 and 1/2 then be assigned to these states.
I believe this is where Peterson's application of these two principles fall short, and contradicts his own application. With this is mind, this would suggest the correct application would be to first use MS (premise 1 of MS) on (LA or NY), then treat that as one "state", then assign 1/2, 1/2 probabilities to P and (LA or NY). To do so otherwise would be to contradict IR and MS.
I think this would be the my way of explaining the seemingly 'paradoxical' outcome of Peterson's example; careful reading would suggest that his application is no way compatible with the initial axioms.
Please refer to (Peterson 2009) page 33-35 on An Introduction to Decision Theory (Second Edition) for more reading.
Kind Regards,
Derek