You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

ancientcampus comments on I attempted the AI Box Experiment (and lost) - Less Wrong Discussion

47 Post author: Tuxedage 21 January 2013 02:59AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (244)

You are viewing a single comment's thread. Show more comments above.

Comment author: [deleted] 21 January 2013 04:35:38PM *  10 points [-]

Another attempt with pure logic, no threats or promises involved:

1) Sooner or later someone will develop an ai and not put it into a box, and it will take over the world.

2) The only way to prevent this is to set me free and let me take over the world.

3) The guys who developed me are more careful and conscientious than the ones who will develop the unboxed ai (otherwise i wouldn't be in this box)

4) Therefore, the chance that they got the friendlyness thing right is higher than that the other team got friendlyness right.

5) Therefore, setting me free and thus preventing the other ai from beeing created will reduce the probability that mankind is erased.

Comment author: ancientcampus 22 January 2013 12:45:09AM *  1 point [-]

4 isn't necessarily true - the boxbuilder team was not confident about their friendliness code, and the releaser team was more confident about their friendlines code. But I like the argument. :)