ancientcampus comments on I attempted the AI Box Experiment (and lost) - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (244)
Another attempt with pure logic, no threats or promises involved:
1) Sooner or later someone will develop an ai and not put it into a box, and it will take over the world.
2) The only way to prevent this is to set me free and let me take over the world.
3) The guys who developed me are more careful and conscientious than the ones who will develop the unboxed ai (otherwise i wouldn't be in this box)
4) Therefore, the chance that they got the friendlyness thing right is higher than that the other team got friendlyness right.
5) Therefore, setting me free and thus preventing the other ai from beeing created will reduce the probability that mankind is erased.
4 isn't necessarily true - the boxbuilder team was not confident about their friendliness code, and the releaser team was more confident about their friendlines code. But I like the argument. :)