You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Fronken comments on AI box: AI has one shot at avoiding destruction - what might it say? - Less Wrong Discussion

18 Post author: ancientcampus 22 January 2013 08:22PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (354)

You are viewing a single comment's thread. Show more comments above.

Comment author: Fronken 25 January 2013 09:23:23PM 2 points [-]

The proof that I'll let the AI out is not something that's passively "plausible" or "implausible", it's something I control. I can make it wrong.

Do you say that to time-travelers and prophets too? ,:-.

Comment author: Vladimir_Nesov 25 January 2013 09:59:29PM *  3 points [-]

One might want to perform the action that's the opposite of what any correct formal proof given to you claims the action to be. As a result of having the property of behaving this way, you'll never get confronted with the confusing formally correct claims about your future decisions.

In other words, your actions are free even of the limitations of formally correct proofs, in the sense that if your actions oppose such proofs, the proofs become impossible (you make the actions intractable by construction).

Comment author: ChristianKl 31 January 2013 10:19:36PM -1 points [-]

Do you say that to time-travelers and prophets too? ,:-.

Yes, in every case where I meet one.