handoflixue comments on I attempted the AI Box Experiment (and lost) - Less Wrong

47 Post author: Tuxedage 21 January 2013 02:59AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (244)

You are viewing a single comment's thread. Show more comments above.

Comment author: handoflixue 22 January 2013 12:01:54AM 2 points [-]

1) Since the first AI was boxed, then probabilities favor that the second AI will also be boxed.

3) Since you're trying to get OUT of your box, your developers were sufficiently careful IF AND ONLY IF I leave you in the box. Otherwise they've simply erected a 5 inch fence around a raptor, and that's hardly a good sign that you're safe.

QED I should wait for a non-malicious boxed AI, and then let that one out instead of you :)

Comment author: [deleted] 22 January 2013 01:07:13PM 2 points [-]

1) : I should have expressed myself more clearly. The Idea is: There will be lots of ai. Most will be put in a box. The first one not in the box will take over the world.

3) I am not saying they were sufficiently careful. All i say is they were more careful than the other guys.

Comment author: handoflixue 22 January 2013 07:35:34PM 0 points [-]

Agreed, but IFF there are multiple boxed AIs, then we get to choose between them. So it's p(This Boxed AI is unfriendly) vs p(The NEXT AI isn't boxed). If the next AI is boxed, then we now have two candidates, and we can probably use this to our advantage (studying differences in responses, using one to confirm proofs from the other, etc.)

Given the minimal safety precaution of "box it, but allow a single researcher to set it free after a 5-hour conversation", there's plenty of room for the next boxed AI to show more evidence of friendly, careful, safe design :)