Mitchell_Porter comments on Open Thread September, Part 3 - Less Wrong

2 Post author: LucasSloan 28 September 2010 05:21AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (203)

You are viewing a single comment's thread. Show more comments above.

Comment author: Mitchell_Porter 29 September 2010 04:21:53AM 0 points [-]

I consider it debatable whether these amendments to naive CDT - CDT plus keeping a commitment, CDT plus cooperating with yourself - really constitute a new decision theory. They arise from reasoning about the situation just a little further, rather than importing a whole new method of thought. Do TDT or UDT have a fundamentally different starting point to CDT?

Comment author: Nisan 29 September 2010 08:02:09AM 1 point [-]

Well, I'm not sure what you're asking here. The problem that needs solving is this: We don't have a mathematical formalism that tells us what to do and which also satisfies a bunch of criteria (like one-boxing on Newcomb's problem, etc.) which attempt to capture the idea that "a good decision theory should win".

When we criticize classical CDT, we are actually criticizing the piece of math that can be translated as "do the thing that, if I-here-now did it, would cause the best possible situation to come about". There are lots of problems with this. "Reasoning about the situation" ought to go into formulating a new piece of math that has no problems. All we want is this new piece of math.

Comment author: Mitchell_Porter 29 September 2010 09:32:59AM 1 point [-]

I'm only just learning that (apparently) the standard rival of causal decision theory is "evidential decision theory". So is that the original acausal decision theory, with TDT and UDT just latecomers local to LW? As you can see I am dangerously underinformed about the preexisting theoretical landscape, but I will nonetheless state my impressions.

If I think about a "decision theory" appropriate for real-world decisions, I think about something like expected-utility maximization. There are a number of problems specific to the adoption of a EUM framework. For example, you have to establish a total order on all possible states of the world, and so you want to be sure that the utility function you construct genuinely represents your preferences. But assuming that this has been accomplished, the problem of actually maximizing expected utility turns into a problem of computation, modeling an uncertain world, and so forth.

The problems showing up in these debates about causal vs evidential and causal vs acausal seem to have a very different character. If I am making a practical decision, I expect both to use causal thinking and to rely on evidence. CDT vs EDT then sounds like a debate about which indispensable thing I can dispense with.

Another thing I notice is that the thought experiments which supposedly create problems for CDT all involve extremes that don't actually happen. Newcomb's problem involves a superbeing with a perfect capacity to predict your choice, Parfit's Hitchhiker is picked up by a mind reader who absolutely knows whether you will keep a promise or not, PD against your copy assumes that you and your copy will knowably make exactly the same choice. (At least this last thought experiment is realizable, in miniature, with simple computer programs.) What happens to these problems if you remove the absolutism?

Suppose Omega or Parfit's mindreader is right only 99% of the time. Suppose your copy only makes the same choice as you do, 99% of the time. It seems like a practically relevant decision theory (whether or not you call it CDT) should be able to deal with such situations, because they are only a variation on the usual situation in reality, where you don't have paranormally assured 100% knowledge of other agents, and where everything is a little inferential and a little uncertain. It seem that, if you want to think about these matters, first you should see how your decision theory deals with the "99% case", and then you should "take the limit" to the 100% case which defines the traditional thought experiment, and you should see if the recommended decisions vary continuously or discontinuously.

Comment author: cousin_it 29 September 2010 12:01:45PM *  5 points [-]

All these thought experiments are realizable as simple computer programs, not only PD. In fact the post I linked to shows how to implement Newcomb's Problem.

The 99% case is not very different from the 100% case, it's continuous. If you're facing a 99% Omega (or even a 60% Omega) in Newcomb's Problem, you're still better off being a one-boxer. That's true even if both boxes are transparent and you can see what's in them before choosing whether to take one or two - a fact that should make any intellectually honest CDT-er stop and scratch their head.

No offense, but I think you should try to understand what's already been done (and why) before criticizing it.

Comment author: MatthewW 29 September 2010 05:39:01PM 3 points [-]

To get to the conclusion that against a 60% Omega you're better off to one-box, I think you have to put in a strong independence assumption: that the probability of Omega getting it wrong is independent of the ways of thinking that the player is using to make her choice.

I think that's really the original problem in disguise (it's a 100% Omega who rolls dice and sometimes decides to reward two-boxing instead of one-boxing). The analysis if all you know is that Omega is right 60% of the time would look different.

Comment author: cousin_it 29 September 2010 09:19:16PM *  0 points [-]

The analysis if all you know is that Omega is right 60% of the time would look different.

How exactly different?

Comment author: MatthewW 29 September 2010 10:25:56PM 2 points [-]

It would become a mind game: you'd have to explicitly model how you think Omega is making the decision.

The problem you're facing is to maximise P(Omega rewards you|all your behaviour that Omega can observe). In the classical problem you can substitute the actual choice of one-boxing or two-boxing for the 'all your behaviour' part, because Omega is always right. But in the 'imperfect Omega' case you can't.

Comment author: Mitchell_Porter 29 September 2010 01:54:01PM 2 points [-]

Start at 50% then, with Omega no better than chance. For each thought experiment, start with a null version where there's nothing unusual and where CDT is supposed to work. Then vary the relevant parameter until there's a problem, and understand what has changed.

Comment author: SilasBarta 29 September 2010 07:02:02PM 1 point [-]

That's part of what the people who have been exploring this problem have already done, and why some posters are upset that you're asking this without apparently having tried to get up-to-date on any of this.

Comment author: Mitchell_Porter 30 September 2010 12:20:26AM 0 points [-]

I don't see the bridge from ordinary decision problems to the thought experiments. I see extreme scenarios being constructed, and then complicated solutions being proposed just to deal with those scenarios. I don't consider this a reliable way to arrive at the correct general form of decision theory.

You say that some people have already gone in the other direction, starting with ordinary decision problems and then slowly changing something until ordinary decision theory breaks. If so, great, and I'm sorry I missed it, but where is it? Is it on this site? Somewhere in the literature?

Comment author: wnoise 30 September 2010 06:08:59AM 1 point [-]

Ah, so you don't see the utility of thought experiments about traveling near light speed either then?

Comment author: Mitchell_Porter 30 September 2010 06:22:16AM 0 points [-]

The analogy with relativity had occurred to me. But we could use another analogy from high-energy physics: There are a very large number of theories which have the standard model (the empirically validated part of particle physics) as their low-energy limit. We can't just rely on high-energy thought-experiments to figure out the actual high-energy physics. We need to do some real experiments where we start low, ramp up the energy, and see what happens.

Comment author: wnoise 30 September 2010 05:23:32PM 0 points [-]

We can't just rely on high-energy thought-experiments to figure out the actual high-energy physics.

Right. We can only use it to rule out incoherent or otherwise "clearly wrong" high-energy physics. But in this analogy, we've shown that CDT seems to not be optimal in this extreme case. if we can define a DT that does better than CDT in this case, and no worse in normal cases, we should use it. I don't think TDT has been well enough defined yet to subject to all conceivable tests, but anything that is following the same kinds of principals will reproduce CDT in most cases, and do better in this case.

We need to do some real experiment where we start low, ramp up the energy, and see what happens.

Here's where the analogy falls down -- we only need to start low and ramp up the energy because of the difficulties of doing high-energy experiments. (And theory-wise, we extrapolate down from clear differences between theories at high energies to find signatures of small differences at lower energies.) If the extreme energies are accessible (and not crazily dangerous), we can just go ahead and test in that regime. Game theory is math. In math, unlike physics, there is no difference between thought experiments and real experiments. The question of applicability in everyday life is an applied economics / sociology / psychology one. How close are people or situations that appear to be screwy in this omega-like way to actually being that way?

Comment author: SilasBarta 30 September 2010 02:50:24AM 0 points [-]

See my other reply, or the links any others have given you, or Drescher's handling of acausal means-end links in chapter 7 of Good and Real, which I think I did a good job summarizing here.

Comment author: Mitchell_Porter 30 September 2010 06:00:51AM 0 points [-]

It sounds like I'll have to work through this in my own fashion. As I said, I want to start with a null version, where CDT works - for example, a situation where Omega has no special knowledge and just guesses what your choice was. Obviously two-boxing is the right thing to do in that situation, CDT says so, and I assume that TDT says so too (though it would be nice to see a worked-out derivation in TDT of that conclusion). Then we give Omega some small but nonzero ability to predict what your choice is going to be. At a guess, the optimal strategy here will be a mixed one, one-boxing with probability p and two-boxing with probability (1-p). I think everyone will tell me that CDT always says p should be zero, but is that really so? I'm just not convinced that I need TDT in order to reach the obvious conclusion.

Comment author: cousin_it 30 September 2010 10:33:04AM *  3 points [-]

At a guess, the optimal strategy here will be a mixed one, one-boxing with probability p and two-boxing with probability (1-p).

If Omega's correctness is independent of your thought process, the optimal strategy will be pure, not mixed. As you make Omega more accurate, at some point you switch from pure two-boxing to pure one-boxing.

Comment author: SilasBarta 29 September 2010 07:03:48PM *  2 points [-]

Another thing I notice is that the thought experiments which supposedly create problems for CDT all involve extremes that don't actually happen.

Really? People never decide how to treat you based on estimations of your decision theory (aka your "character")?

Comment author: Mitchell_Porter 30 September 2010 12:29:36AM 1 point [-]

They don't make those decisions with "paranormally assured 100% knowledge" of my decision theory. That's the "extreme that doesn't actually happen". And this is why I won't be adopting any new paradigm of decision theory unless I can start in the middle, with situations that do happen, and move gradually towards the extremes, and see the desirability or necessity of the new paradigm that way.

Comment author: SilasBarta 30 September 2010 02:45:58AM *  2 points [-]

As has been said many times (at least by me, definitely by many others), you don't need 100% accuracy for the argument to hold. If Parfit's mindreader is only 75% accurate, that still justifies choosing the pay/ cooperate / one-box option. One-boxing on newcomblike problems is simply what you get when you have a decision theory that wins in these reasonable cases, and which is continuous -- and then take the limit as all the predicate variables go to what they need to be to make it Newcomb's problem (such as making the predictor 100% accurate).

If it helps, think of the belief in one-boxing as belief in the implied optimal.

It doesn't matter that you'll never be in Newcomb's problem. It doesn't matter that you'll never be in an epistemic state where you can justifiably believe that you are. It's just an implication of having a good decision theory.

Comment author: Mitchell_Porter 30 September 2010 06:05:42AM 0 points [-]

Part of my concern is that I'll end up wasting time, chasing my tail in an attempt to deal with fictitious problems, when I could be working on real problems. I'm still undecided about the merits of acausal decision theories, as a way of dealing with the thought experiments, but I am really skeptical that they are relevant to anything practical, like coordination problems.

Comment author: wedrifid 29 September 2010 04:59:56AM 0 points [-]

Err... the 'C'? 'Causal'.