Vladimir_Nesov comments on What is Wei Dai's Updateless Decision Theory? - Less Wrong

37 Post author: AlephNeil 19 May 2010 10:16AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (63)

You are viewing a single comment's thread.

Comment author: Vladimir_Nesov 19 May 2010 11:08:29AM *  8 points [-]

I believe the discussion of UDT is spot on, and a very good summary placing various thought experiments in its context (though reframing Smoking Lesion to get the correct answer seems like cheating).

I have trouble understanding your second point about Sleeping Beauty (and DT-independent probabilities).

Comment author: AlephNeil 19 May 2010 11:35:24AM *  3 points [-]

Thanks very much! I'm especially pleased that you thought it was accurate.

As for the second point - yeah it seems everyone wants to disagree with me on that :-/

What I describe (perhaps unclearly) is a 'standard recipe' for attaching meaning to statements about indexical probabilities (like "I am at the second intersection" in the absent-minded driver problem) which doesn't depend on decision theory (except in way I noted as the 'caveat').

Perhaps it may be objected that there are other recipes. (One such recipe might be 'take a random branch that has at least one player node on it, then take a random player-instance somewhere along that branch'. This of course gives 1/2 as the answer the Sleeping Beauty problem.)

I don't really have any 'absolute justification' for mine, except that it gives the solution to an elegant decision problem: "At every player-instance, try to work out which player-instance you are, so as to minimize -log(subjective probability) at that instance." (With it being implicit that your final utility is the sum of all such 'log(subjective probability)' expressions along the branch.)

Comment author: Vladimir_Nesov 19 May 2010 12:05:19PM *  3 points [-]

You can of course define probability in a way that doesn't refer to any specific decision theory, thus making it "independent" of decision theories. But probability is useful exactly as half-of-decision-theory, where you just add "utility" ingredient to get the correct decisions out. This doesn't work well where indexical uncertainty or mind copying are involved, because "probabilities" you get in those situations (defined in such a way that the resulting decisions are as you'd prefer, as in justification of probability by a bet) depend more on your preference than normally. In simpler situations, maximum entropy at least takes care of situations you don't terminally distinguish in your values, in a way that is independent on further details of your values.

Comment author: Stuart_Armstrong 22 May 2010 08:39:38AM 2 points [-]

Have you seen Full Non-Idexical Conditioning? (http://www.cs.toronto.edu/~radford/ftp/anth.pdf) Though the theory is mathematically incorrect, it's very nearly right, and it's very similar to your sleeping beaty approach...