AnnaSalamon comments on Decision theory: Why Pearl helps reduce “could” and “would”, but still leaves us with at least three alternatives - Less Wrong

30 Post author: AnnaSalamon 06 September 2009 06:10AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (70)

You are viewing a single comment's thread. Show more comments above.

Comment author: jimmy 06 September 2009 07:01:39PM 3 points [-]

I think a lot of the confusion about these types of decision theory problems has to do with not not everyone thinking about the same problem even when it seems like they are.

For example, consider the problem I'll call 'pseudo newcombs problem'. Omega still gives you the same options, and history has proven a strong correlation between peoples choices and its predictions.

The difference is that instead of simulating the relevant part of your decision algorithm to make the prediction, Omega just looks to see whether you have a blue dot or a red dot on your forehead- since a red dot has been a perfect indicator of a mental dysfunction that makes the response to every query "one box!" and blue dot has been a perfect indicator of a functioning brain. In addition, all people with working brains have chosen two boxes in the past.

If I understand correctly, all decision theories discussed will two box here, and rightly so- choosing one box doesn't cause Omega to choose differently since that decision was determined solely by the color of your dot.

People that say to two box on Newcomblike problems think of this type of Omega, since sufficiently detailed simulations aren't the first thing that come to mind- indicators of broken minds do.

For the one shot PD, it seems like something similar is happening. Cooperating just doesn't 'seem' right to me most of the time, but it's only because I'd have a hard time believing the other guy was running the same algorithm.

I had an interesting dream where I was copied recently, and it made cooperation on one shot PD a lot more intuitive. Even when trued, I'd cooperate with that guy, no question about it.

Comment author: AnnaSalamon 06 September 2009 08:20:38PM *  1 point [-]

If I understand correctly, all decision theories discussed will two box here, and rightly so- choosing one box doesn’t cause Omega to choose differently since that decision was determined solely by the color of your dot.

Depending on the set-up, “innards-CSAs" may one-box here. Innards-CSAs go back to a particular moment in time (or to their creator’s probability distribution) and ask: “if I had been created at that time, with a (perhaps physically transparent) policy that would one-box, would I get more money than if I had been created with a (perhaps physically transparent) policy that would two-box?”

If your Omega came to use the colored dots in its prediction because one-boxing and two-boxing was correlated with dot-colors, and if the innards-CSA in question is programmed to do its its counterfactual innards-swap back before Omega concluded that this was the correlation, and if your innards-CSA ended up copied (perhaps with variations) such that, if it had had different innards, Omega would have ended up with a different decision-rule... then it will one-box.

And “rightly so” in the view of the innards-CSA... because, by reasoning in this manner, the CSA can increase the odds that Omega has decision-rules that favor its own dot-color. At least according to its own notion of how to reckon counterfactuals.

Comment author: jimmy 07 September 2009 06:15:25PM 1 point [-]

I see your point, which is why I made sure to write "In addition, all people with working brains have chosen two boxes in the past."

My point is that you can have situations where there is a strong correlation so that Omega nearly always predicts correctly, but that Omega's prediction isn't caused by the output of the algorithm you use to compute your decisions, so you should two box.

The lack of effort to distinguish between the two cases seems to have generated a lot of confusion (at least it got me for a while).

Comment author: Eliezer_Yudkowsky 06 September 2009 08:30:01PM *  1 point [-]

Depending on your beliefs about what computation Omega did to choose its policy, the TDT counterfactual comes out as either "If things like me one-boxed, then Omega would put $1m into box B on seeing a blue dot" or "If things like me one-boxed, then Omega would still have decided to leave B empty when seeing a blue dot, and so if things like me one-boxed I would get nothing."