Ronak comments on [link] Scott Aaronson on free will - Less Wrong

20 Post author: DanielVarga 10 June 2013 11:24PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (109)

You are viewing a single comment's thread.

Comment author: Ronak 13 June 2013 12:55:42PM 0 points [-]

I like his causal answer to Newcomb's problem:

In principle, you could base your decision of whether to one-box or two-box on anything you like: for example, on whether the name of some obscure childhood friend had an even or odd number of letters. However, this suggests that the problem of predicting whether you will one-box or two-box is “you-complete.” In other words, if the Predictor can solve this problem reliably, then it seems to me that it must possess a simulation of you so detailed as to constitute another copy of you (as discussed previously). But in that case, to whatever extent we want to think about Newcomb’s paradox in terms of a freely-willed decision at all, we need to imagine two entities separated in space and time—the “flesh-and-blood you,” and the simulated version being run by the Predictor—that are nevertheless “tethered together” and share common interests. If we think this way, then we can easily explain why one-boxing can be rational, even without backwards-in-time causation. Namely, as you contemplate whether to open one box or two, who’s to say that you’re not “actually” the simulation? If you are, then of course your decision can affect what the Predictor does in an ordinary, causal way.

Comment author: Manfred 15 June 2013 11:09:19PM 3 points [-]

Simple but misleading.

This is because Newcomb's problem is not reliant on the predictor being perfectly accurate. All they need to do is predict you so well that people who one-box walk away with more expected utility than people who two-box. This is easy - even humans can predict other humans this well (though we kinda evolved to be good at it).

So if it's still worth it to one-box even if you're not being copied, what good is an argument that relies on you being copied to work?

Comment author: Ronak 16 June 2013 04:51:59PM *  0 points [-]

In response to this, I want to roll back to saying that while you may not actually be simulated, having the programming to one-box is what causes there to be a million dollars in there. But, I guess that's the basic intuition behind one-boxing/the nature of prediction anyway so nothing non-trivial is left (except the increased ability to explain it to non-LW people).

Also, the calculation here is wrong.

Comment author: Eugine_Nier 16 June 2013 03:04:42AM 0 points [-]

Ok, in that case, am I allowed to roll a dice to determine whether to one box?

Comment author: Manfred 16 June 2013 05:42:53AM *  0 points [-]

Depends on the rules. Who do I look like, Gary Drescher?

What sort of rules would you implement to keep Newcomb's problem interesting in the fact of coins that you can't predict?

Comment author: Eugine_Nier 18 June 2013 12:22:17AM 0 points [-]

Why would I want to keep the problem interesting? I want to solve it.

Comment author: Ronak 18 June 2013 05:55:23PM 3 points [-]

Because the solution to the problem is worthless except to the extent that it establishes your position in an issue it's constructed to illuminate.

Comment author: TheOtherDave 13 June 2013 03:30:55PM 0 points [-]

It seems way simpler to leave out the "freely willed decision" part altogether.

If we posit that the Predictor can reliably predict my future choice based on currently available evidence, it follows that my future choice is constrained by the current state of the world. Given that, what remains to be explained?

Comment author: Ronak 13 June 2013 06:49:06PM *  0 points [-]

Yes, I agree with you - but when you tell some people that the question arises of what is in the big-money box after Omega leaves... and the answer is "if you're considering this, nothing."

A lot of others (non-LW people) I tell this to say it doesn't sound right. The bit just shows you that the seeming closed-loop is not actually a closed loop in a very simple and intuitive way** (oh and it actually agrees with 'there is no free will'), and also it made me think of the whole thing from a new light (maybe other things that look like closed loops can be shown not to be in similar ways).

** Anna Salamon's cutting argument is very good too but a) it doesn't make the closed-loop-seeming thing any less closed-loop-seeming and b) it's hard to understand for most people and I'm guessing it will look like garbage to people who don't default to compatibilist.

Comment author: TheOtherDave 13 June 2013 07:20:35PM 1 point [-]

I suppose.
When dealing with believers in noncompatibilist free will, I typically just accept that on their view a reliable Predictor is not possible in the first place, and so they have two choices... either refuse to engage with the thought experiment at all, or accept that for purposes of this thought experiment they've been demonstrated empirically to be wrong about the possibility of a reliable Predictor (and consequently about their belief in free will).

That said, I can respect someone refusing to engage with a thought experiment at all, if they consider the implications of the thought experiment absurd.

As long as we're here, I can also respect someone whose answer to "Assume Predictor yadda yadda what do you do?" is "How should I know what I do? I am not a Predictor. I do whatever it is someone like me does in that situation; beats me what that actually is."

Comment author: Ronak 14 June 2013 07:11:58PM 0 points [-]

I usually deal with people who don't have strong opinions either way, so I try to convince them. Given total non-compatibilists, what you do makes sense.

Also, it struck me today that this gives a way of one-boxing within CDT. If you naively blackbox prediction, you would get an expected utility table {{1000,0},{1e6+1e3,1e6}} where two-boxing always gives you 1000 dollars more.

But, once you realise that you might be a simulated version, the expected utility of one-boxing is 1e6 but of two-boxing is now is 5e5+1e3. So, one-box.

A similar analysis applies to the counterfactual mugging.

Further, this argument actually creates immunity to the response 'I'll just find a qubit arbitrarily far back in time and use the measurement result to decide.' I think a self-respecting TDT would also have this immunity, but there's a lot to be said for finding out where theories fail - and Newcomb's problem (if you assume the argument about you-completeness) seems not to be such a place for CDT.

Disclaimer: My formal knowledge of CDT is from wikipedia and can be summarised as 'choose A that maximises where D is the desirability function and P the probability function.'