You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

orthonormal comments on Decision Theory Paradox: Answer Key - Less Wrong Discussion

6 Post author: orthonormal 05 September 2011 11:13PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (10)

You are viewing a single comment's thread. Show more comments above.

Comment author: orthonormal 06 September 2011 12:41:00PM 0 points [-]

So if I understand you right, even with the short-sighted utility function, there's an echo of Parfit's Hitchhiker here: what TDT decides on these problems actually controls which situations the agent finds itself in, and thus its possible payoff matrices. Since TDT is supposed to get Parfit's Hitchhiker right, therefore, it should give the long-term-winning answer even in this case.

Well, there are some more caveats (it's not clear that agents in the first round would do this, since TDT doesn't win the Counterfactual Mugging, and if agents in round N don't think that way, then what about round N+1...), but you're right that the simple calculation doesn't actually suffice. Drat, and thanks.

Comment author: Vladimir_Nesov 06 September 2011 02:19:50PM *  1 point [-]

Since TDT is supposed to get Parfit's Hitchhiker right, therefore, it should give the long-term-winning answer even in this case.

Its goal is still different though (if we restore some missing pieces): it wants to game the probabilities of encountering certain opponents so that a single round that contains TDT delivers the most reward. It just so happens that getting rid of DefectBots serves this purpose, but if the opponents were CooperateBots, it looks like TDTs would drive themselves to extinction (or farm the opponents) to maximize the number of expected cooperating opponents that they can defect against (for each instance where there's a TDT agent in the round). (I didn't check this example carefully, so could be wrong, but the principle it exemplifies seems to hold.)

Comment author: orthonormal 06 September 2011 11:04:01PM 0 points [-]

That's a seriously sick idea, but there doesn't seem to be a way to both set up such a favorable matchup and exploit it- is there?