Stuart_Armstrong comments on Reduced impact AI: no back channels - Less Wrong

13 Post author: Stuart_Armstrong 11 November 2013 02:55PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (41)

You are viewing a single comment's thread. Show more comments above.

Comment author: MugaSofer 24 November 2013 07:39:33PM *  -1 points [-]

I just read through the comments, and no-one seems to have said this yet. So either I missed something in the OP or it was just too obvious to mention? Regardless:

How is this elaborate proposal superior to bog-standard AI boxing?

Why bother to write in an elaborate function telling it not to affect the outside world except through a channel, when you could simply only give it one channel? It can either escape through the specified channel, or it can't. Whether the channel is connected to an airgapped computer running the AI or an unlocked cell containing a sophisticated robot running the AI seems immaterial.

The example of the laser co-ordinates is relatively secure, although it might be possible to do stuff with your one shot - the situation isn't that clearly specified because it's unrealistic in any case. But that's a property of the output mechanism, not the AI design. Isn't it?

Comment author: Stuart_Armstrong 25 November 2013 02:54:39PM 0 points [-]

How is this elaborate proposal superior to bog-standard AI boxing?

For the moment, there is little difference between a really well boxed AI and a reduced impact one (though consider the "a few extra paperclips" AI). But this is an epistemic boxing, not a physical one, which is interesting/useful, and the idea may develop into something more.