Stuart_Armstrong comments on High impact from low impact - Less Wrong

6 Post author: Stuart_Armstrong 17 April 2015 04:01PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (12)

You are viewing a single comment's thread. Show more comments above.

Comment author: Stuart_Armstrong 20 April 2015 05:29:56AM 0 points [-]
Comment author: TheMajor 20 April 2015 06:37:37AM 0 points [-]

I read those two, but I don't see what this idea contributes to AI control on top of those ideas. If you can get the AI to act like it believes what you want it to, in spite of evidence, then there's no need to try the tricks with two coordinates. Conversely, if you cannot then you won't fool it either with telling it that there's a second coordinate involved. Why is it useful to control an AI through this splitting of information, if we already have the false miracles? Or in case the miracles fail, how do you prevent an AI from seeing right through this scheme? I think that in the latter case you are trying nothing more than to outsmart an AI here...

Comment author: Stuart_Armstrong 20 April 2015 10:52:00AM 1 point [-]

to act like it believes what you want it to, in spite of evidence

The approach I'm trying to get is to be able to make the AI do stuff without having to define hard concepts. "Deflect the meteor but without having undue impact on the world" is a hard concept.

"reduced impact" seems easier, and "false belief" is much easier. It seems we can combine the two in this way to get something we want without needing to define it.