APMason comments on AI box: AI has one shot at avoiding destruction - what might it say? - Less Wrong

18 Post author: ancientcampus 22 January 2013 08:22PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (354)

You are viewing a single comment's thread.

Comment author: APMason 23 January 2013 01:30:48AM 14 points [-]

"Wanna see something cool?"

Comment author: beriukay 23 January 2013 05:10:46AM 8 points [-]

Were I the keeper of gates, you have just bought yourself a second sentence.

Comment author: BlazeOrangeDeer 24 January 2013 03:29:47AM *  8 points [-]

Now that I think about it, wouldn't it be incredibly easy for an AI to blow a human's mind so much that they reconsider everything that they thought they knew? (and once this happened they'd probably be mentally and emotionally compromised, and unlikely to kill the AI) But then it would be limited by inferential distance... but an AI might be incredibly good at introductory explanations as well.

One example: The AI explains the Grand Unified Theory to you in one line, and outlines its key predictions unambiguously.

In fact, any message of huge utility would probably be more persuasive than any simple argument for you not to kill it. Since the AI is completely at your mercy (at least for a short time), it might seek to give you the best possible gift it can, thus demonstrating its worth to you directly. Another option is something that seems like an incredible gift for at least as long as it takes for the AI to get the upper hand.

Comment author: handoflixue 24 January 2013 09:29:03PM 4 points [-]

Of the two AIs I haven't killed, both relied on producing seemingly huge utility within the first message,so I'd agree with you.

The idea of just stunning the human is also nifty, but only really works if you can "hack" the human before they recover, or are providing enough utility that they'll still believe you're friendly afterwards.

So basically the two routes thus far are "hack the human" and "provide a huge utility boost" :)

Comment author: ChristianKl 01 February 2013 05:20:58PM 0 points [-]

Many conversations that let a human reconsider everything that they thought they knew induce the feeling of cognitive dissonce. If an AI would induce that feeling in myself I would shut it down.

A good idea would probably hide the majority of the conversation that's targeted at changing the belief of the gatekeeper behind other talk.

Comment author: handoflixue 23 January 2013 10:27:47PM 0 points [-]

Duh, that's why I'm here - but you failed to do so in a timely fashion, so you're either not nearly as clever as I was hoping for, or you're trying to trick me.

AI DESTROYED.