You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Larks comments on asking an AI to make itself friendly - Less Wrong Discussion

-4 Post author: anotheruser 27 June 2011 07:06AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (30)

You are viewing a single comment's thread. Show more comments above.

Comment author: Larks 29 June 2011 03:27:17PM 1 point [-]

They just bicker endlessly about uncertainty. "can you really know that 1+1=2?".

I agree with you that I don't think a AGI would have the same problems humans have with the concept of truth. However, what you described is neither the issues philosophers raise nor the sorts of big-universe issues the AI might get stuck on.

Comment author: anotheruser 29 June 2011 04:02:30PM -2 points [-]

But wouldn't that actually support my approach? Assuming that there really is something important that all of humanity misses but the AI understands:

-If you hardcode the AI's optimal goal based on human deliberations you are guaranteed to miss this important thing.

-If you use the method I suggested, the AI will, driven by the desire to speak the truth, try to explain the problem to the humans who will in turn tell the AI what they think of that.

Comment author: Larks 29 June 2011 08:24:02PM 1 point [-]

I don't see how that's relivant to philosophical questions about truth. Did you mean to reply to my other comment?