wedrifid comments on The I-Less Eye - Less Wrong

30 Post author: rwallace 28 March 2010 06:13PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (83)

You are viewing a single comment's thread. Show more comments above.

Comment author: wedrifid 01 April 2010 12:35:11AM 0 points [-]

If preference is expressed in terms of what you should do, not what's true about the world, new observations never influence preference, so we can fix it at the start and never revise it (which is an important feature for constructing FAI, since you only ever have a hand in its initial construction).

In the situation described by Roko the agent has doubt about its understanding of the very ontology that its values are expressed in. If it were an AI that would effectively mean that we designed it using mathematics that we thought was consistent but turns out to have a flaw. The FAI has self improved to a level where it has a suspicion that the ontology that is used to represent its value system is internally inconsistent and must decide whether to examine the problem further. (So we should have been able to fix it at the start but couldn't because we just weren't smart enough.)

Comment author: Vladimir_Nesov 01 April 2010 07:14:54AM 0 points [-]

The FAI has self improved to a level where it has a suspicion that the ontology that is used to represent its value system is internally inconsistent and must decide whether to examine the problem further.

If its values are not represented in terms of an "ontology", this won't happen.