bokov comments on AIs and Gatekeepers Unite! - Less Wrong

10 Post author: Eliezer_Yudkowsky 09 October 2008 05:04PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (160)

Sort By: Old

You are viewing a single comment's thread. Show more comments above.

Comment author: bokov 09 September 2013 09:01:54PM 0 points [-]

Probably the latter, since they both lost at least once. A real AI trying to get out would devote all its energies to counterfeiting friendliness and probably succeeding.

Boxing is non-viable only in the same sense that locks, passwords, treaties, contracts, testing, peer review, seatbelts, and all other imperfect precautions are non-viable.

Pile on enough of them, in combination, and perhaps they will buy a few years or seconds in which to react. All things equal, is there any reason why an AI of unknown friendliness is any safer without being boxed?

A flawed containment method is still better at containing than no containment method (if implemented with awareness of its flaws) but apparently a flawed friendly AI will miss a very small target in goalspace and for practical purposes be unfriendly. So, if we spent 5 minutes considering the alternatives, would we continue to believe that better boxes are a less tractable problem than friendly AI?

Comment author: hairyfigment 09 September 2013 09:39:20PM -1 points [-]

Not that particular AI. But if you think yours is Friendly and others under development have a sufficient probability of being UnFriendly, then trivially, letting it run (in both senses) beats boxing. Oh, and people will die 'naturally' while you dither. I hope that thinking this much about making an AI Friendly will prepare someone to get the job done ASAP once the AGI part seems more feasible.