Vaniver comments on Naive TDT, Bayes nets, and counterfactual mugging - Less Wrong

15 Post author: Stuart_Armstrong 23 October 2012 03:58PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (39)

You are viewing a single comment's thread.

Comment author: Vaniver 23 October 2012 08:14:01PM 2 points [-]

In TDT, as I understand it, we sever you decision node from the history of the universe, and then pick the action that maximises our utility:

This looks wrong; in the picture, you sever the "your decision algorithm" node. Indeed, I think that's the 'difference' between naive TDT and naive CDT- naive CDT supposedly severs "your decision" whereas TDT makes the improvement of severing a step sooner, so it recognizes that it can cause Omega (like a CDTer with a correct map of the problem).

Comment author: Eliezer_Yudkowsky 24 October 2012 09:07:19AM 2 points [-]

Comment author: Stuart_Armstrong 23 October 2012 09:13:42PM 1 point [-]

Typo corrected! And yes, that is the CDT-TDT debate - but not really relevant here.

Comment author: Eliezer_Yudkowsky 24 October 2012 09:07:42AM 7 points [-]

In TDT we don't do any severance! Nothing is uncaused, not our decision, nor our decision algorithm either. Trying to do causal severance is a basic root of paradoxes because things are not uncaused in real life. What we do rather is condition on the start state of our program, thereby screening off the universe (not unlawfully severing it), and factor out our uncertainty about the logical output of the program given its input. Since in real life most things we do to the universe should not change this logical fact, nor will observing this logical fact tell us which non-impossible possible world we are living it, it shouldn't give us any news about the nodes above, once we've screened off the algorithm. It does, however, give us logical news about Omega's output, and of course about which boxes we'll end up with.

Comment author: Vaniver 24 October 2012 07:34:41PM 1 point [-]

My reading of this is that you use influence diagrams, not Bayes nets; you think of your decision as influenced by things preceding it, but not as an uncertainty node. Is that a fair reading, or am I missing something?

Comment author: Stuart_Armstrong 24 October 2012 10:02:03AM *  1 point [-]

I stand corrected, and have corrected it.

Comment author: Armok_GoB 24 October 2012 10:33:49PM 0 points [-]

My instant reaction upon hearing this is to try to come up with cases where they DO change that logical fact. Holding of on proposing solutions for now.

Comment author: Vaniver 23 October 2012 09:32:31PM *  2 points [-]

So, for this issue I would note that for the coinflip to influence the decision algorithm, there needs to be an arrow from the coinflip to the decision algorithm. Consider two situations:

  1. Omega explains the counterfactual mugging deal, learns whether you would pay if the coin comes up tails, and then tells you how the coin came up.

  2. Omega tells you how the coin comes up, explains the counterfactual mugging deal, and then learns whether you would pay if the coin comes up tails.

Those have different Bayes nets and so it can be entirely consistent for TDT to output different strategies in each.