You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Andreas_Giger comments on Can anyone explain to me why CDT two-boxes? - Less Wrong Discussion

-12 Post author: Andreas_Giger 02 July 2012 06:06AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (136)

You are viewing a single comment's thread. Show more comments above.

Comment author: Andreas_Giger 03 July 2012 10:19:22AM *  1 point [-]

Given a description of Newcomb's problem, a well-written program will one-box, and a badly-written one will two-box. The difference between the two is not trivial.

I see your point now, and I agree with the quoted statement. However, there's a difference between Newcomb, where you make your decision after Omega made its prediction, and "meta-Newcomb", where you're allowed to precommit before Omega makes its prediction, for example by choosing your programming. In meta-Newcomb, I don't even have to consider being a computer program that can be simulated; I can just give my good friend Epsilon, who always exactly does what he is told, a gun and tell him to shoot me if I lie, then tell Omega I'm going to one-box, and then Omega would make its prediction. I would one-box, get $1,000,000 and, more importantly, not shot.

This is a decision that CDT would make, given the opportunity.

Comment author: Emile 03 July 2012 01:17:30PM *  1 point [-]

there's a difference between Newcomb, where you make your decision after Omega made its prediction, and "meta-Newcomb", where you're allowed to precommit before Omega makes its prediction, for example by choosing your programming.

I agree that meta-Newcomb is not the same problem, and that in meta-newcomb CDT would precommit to one-box.

However, even in normal Newcomb, it's possible to have agents that behave as if they had precommited when they realize precomitting would have been better for them. More specifically, in pseudocode:

function takedecision(informationabout_world, actions):

for each action:

calculate the utility that an agent that always returns that action would have got

return the action that got the highest utility

There are some subtleties, notably about how to take the information about the world into account, but an agent built along this model should one-box on problems like Newcomb's, while two-boxing in cases where Omega decides by flipping a coin.

(such an agent; however, doesn't cooperate with itself in prisonner's dilemma, you need a better agent for that)

Comment author: Andreas_Giger 03 July 2012 01:25:01PM *  0 points [-]

You are 100% correct. However, if you say "it's possible to have agents that behave as if they had precommited", then you are not talking about what's the best decision to make in this situation, but what's the best decision theory to have in this situation, and that is, again, meta-Newcomb, because the decision which decision theory you're going to follow is a decision you have to make before Omega makes its prediction. Switching to this decision theory after Omega makes its prediction doesn't work, obviously, so this is not a solution for Newcomb.

Comment author: wedrifid 03 July 2012 10:48:20AM *  0 points [-]

I can just give my good friend Epsilon, who always exactly does what he is told, a gun and tell him to shoot me if I lie, then tell Omega I'm going to one-box, and then Omega would make its prediction. I would one-box, get $1,000,000 and, more importantly, not shot.

When I first read this I took it literally, as using Epsilon directly as a lie detector. That had some interesting potential side effects (like death) for a CDT agent. On second reading I take it to mean "Stay around with the gun until after everything is resolved and if I forswear myself kill me". As a CDT agent you need to be sure that Epsilon will stay with the gun until you have abandoned the second box. If Epsilon just scans your thoughts, detects whether you are lying and then leaves then CDT will go ahead and take both boxes anyway. (It's mind-boggling to think of agents that couldn't even manage cooperation with themselves with $1m on the line and a truth oracle right there to help them!)

Comment author: Andreas_Giger 03 July 2012 11:40:10AM *  0 points [-]

Yeah, I meant that Epsilon would shoot if you two-box after having said you would one-box. In the end, "Epsilon with a gun" is just a metaphor for / specific instance of precommitting, as is "computer program that can choose its programming".