Lightwave comments on Desirable Dispositions and Rational Actions - Less Wrong

13 Post author: RichardChappell 17 August 2010 03:20AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (180)

You are viewing a single comment's thread. Show more comments above.

Comment author: RichardChappell 18 August 2010 02:02:24AM *  0 points [-]

calling the difference between map and territory "choice"

Eh? That's not what I'm doing. I'm pointing out that there's a respectable (coarse-grained) sense of 'disposition' (i.e. tendency) according to which one can have a disposition to X without this necessarily entailing that one will actually do X. (There's another sense of 'total disposition' where the entailment does hold. N.B. We make choices either way, but it only makes sense to separately evaluate choices from coarse-grained dispositions.)

I take these general dispositions to accurately correspond to real facts in the world -- they're just at a sufficiently high level of abstraction that they allow for various exceptions. (Ceteris paribus laws are not, just for that reason, "inaccurate".)

Comment author: Lightwave 18 August 2010 09:21:46AM *  0 points [-]

My take on this is the following: It's easier to see what is meant by disposition if you look at it in terms of AI. Replace the human with an AI, replace "disposition" with "source code" and replace "change your disposition to do some action X" to "rewrite your source code so that it does action X". Of course it would still want to incorporate the probability of a glitch as someone else already suggested.

If an AI, which is running CDT expects to encounter a newcomb-like problem, it would be rational for it to self-modify (in advance) to use a decision theory which one-boxes (i.e. the AI will change it's disposition).

Comment author: RichardChappell 19 August 2010 12:40:44AM 0 points [-]

Likewise, an AI surrounded by threat-fulfillers would rationally self-modify to become a threat-ignorer. (The debate is not about whether these are desirable dispositions to acquire -- that's common ground.) Do you think it follows from this that the act of ignoring a doomsday threat is also rational?