Occasionally a wrong idea still leads to the right outcome. We know that one-boxing on Newcomb's problem is the right thing to do. Timeless decision theory proposes to justify this action by saying: act as if you control all instances of your decision procedure, including the instance that Omega used to predict your behavior.
But it's simply not true that you control Omega's actions in the past. If Omega predicted that you will one-box and filled the boxes accordingly, that's because, at the time the prediction was made, you were already a person who would foreseeably one-box. One way to be such a person is to be a TDT agent. But another way is to be a quasi-CDT agent with a superstitious belief that greediness is punished and modesty is rewarded - so you one-box because two-boxing looks like it has the higher payoff!
That is an irrational belief, yet it still suffices to generate the better outcome. My thesis is that TDT is similarly based on an irrational premise. So what is actually going on? I now think that Newcomb's problem is simply an exceptional situation where there is an artificial incentive to employ something other than CDT, and that most such situations can be dealt with by being a CDT agent who can self-modify.
Eliezer's draft manuscript on TDT provides another example (page 20): a godlike entity - we could call it Alphabeta - demands that you choose according to "alphabetical decision theory", or face an evil outcome. In this case, the alternative to CDT that you are being encouraged to use is explicitly identified. In Newcomb's problem, no such specific demand is made, but the situation encourages you to make a particular decision - how you rationalize it doesn't matter.
We should fight the illusion that a TDT agent retrocausally controls Omega's choice. It doesn't. Omega's choice was controlled by the extrapolated dispositions of the TDT agent, as they were in the past. We don't need to replace CDT with TDT as our default decision theory, we just need to understand the exceptional situations in which it is expedient to replace CDT with something else. TDT will apply to some of those situations, but not all of them.
I don't have a better decision theory than TDT, but I also don't believe that what you do in the present affects the past. However, the nature of the situation is such as to reinforce, in a TDT-like agent, the illusion that decisions made in the present affect decisions simulated in the past. (That is, assuming it is an agent self-aware enough to have such beliefs.)
One conception of the relationship between CDT and TDT is that it is like the relationship between classical and relativistic mechanics: relativistic mechanics is truer but it reduces to classical mechanics in a certain limit. But I think TDT is more like alphabetical decision theory - though useful in a far wider variety of scenarios: it is not a decision theory that you would want to have, outside of certain peculiar situations which offer an incentive to deviate from CDT.
I need to study UDT more, because sometimes it sounds like it's just CDT in a multiverse context, and yet it's supposed to favor one-boxing.
Would you cooperate in the Prisoner's Dilemma against an almost-copy of yourself (with only trivial differences so that your experiences would be distinguishable)? It can be set up so that neither of you decide within the light-cone of the other's decision, so there's no way your cooperation can physically ensure the other's cooperation.
If you're quite convinced that the reasonable thing is to defect, then pretty obviously you'll get (D,D).
If you're quite convinced that the reasonable thing is to cooperate, then pretty obviously you'll get (C,C).
(OK, you c... (read more)