You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

RichardKennaway comments on AI box: AI has one shot at avoiding destruction - what might it say? - Less Wrong Discussion

18 Post author: ancientcampus 22 January 2013 08:22PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (354)

You are viewing a single comment's thread.

Comment author: RichardKennaway 23 January 2013 12:24:24PM *  17 points [-]

One reason for Eliezer not publishing the logs of the AIbox experiment is to avoid people seeing how he got out and responding, "ok, so all we have to do to keep the AI in its box is avoid succumbing to that trick." This thread might just provide more fuel for that fallacy (as, I admit, I did in replying to Eliezer's original comment).

I'm sure that for everything an AI might say, someone can think up a reason for not being swayed, but it does not follow that for someone confronted with an AI, there is nothing that would sway them.

Comment author: handoflixue 23 January 2013 09:33:40PM 3 points [-]

I wouldn't expect any effective real-life gatekeeper to be swayed by my ability to destroy one-sentence AIs.

Comment author: gryffinp 29 January 2013 01:39:04AM 1 point [-]

It just occurred to me that Eliezer's original stipulation that no chat logs would be released gives him an advantage. The responses of a Gatekeeper who knows that his inputs will be thoroughly scrutinized by the public will be different then one who has every reason to believe that his discussion will be entirely private.

Has someone else pointed this out before?

Comment author: ancientcampus 24 January 2013 04:19:50AM *  1 point [-]

Honest question: are you proposing we avoid discussing the problem entirely?

Personally, I think there is more to be gained here than just "how will an AI try to get out and how can we prevent it." For me, it's gotten me to actually think about the benefits and pitfalls of a transhuman AI (friendly or otherwise) rather than just knowing intellectually, "there are large potential benefits and pitfalls" which was my previous level of understanding.

Edit: I've modified the OP to include your concerns. They're definitely valid, but I think this is still a good discussion for my reasons above.

Comment author: RichardKennaway 24 January 2013 09:43:49AM 1 point [-]

Honest question: are you proposing we avoid discussing the problem entirely?

No, I just thought that it was worth adding that concern to the pot.

I take what I dare say some would consider a shockingly lackadaisical attitude to the problem of Unfriendly AI, viz. I see the problem, but it isn't close at hand, because I don't think anyone yet has a clue how to build an AGI. Outside of serious mathematical work on Friendliness, discussing it is no more than a recreation.

Comment author: ancientcampus 10 February 2013 09:10:09PM 0 points [-]

That's pretty much my same attitude on the situation, as well. :)

Comment author: OrphanWilde 28 January 2013 05:45:27PM 0 points [-]

Discussing it makes people more aware of exactly how difficult a problem it is. That such discussions are entertaining merely permit the discussions to take place.

Comment author: V_V 25 January 2013 06:16:43PM -1 points [-]

One reason for Eliezer not publishing the logs of the AIbox experiment is to avoid people seeing how he got out and responding, "ok, so all we have to do to keep the AI in its box is avoid succumbing to that trick."

He could post the logs of the games he lost.

Comment author: RichardKennaway 25 January 2013 07:55:41PM 0 points [-]

He could post the logs of the games he lost.

Thereby giving a different reason for false confidence in boxing.

Comment author: V_V 26 January 2013 02:42:37PM 0 points [-]

Can you elaborate, please?