Kaj_Sotala comments on Open Thread: November 2009 - Less Wrong

3 [deleted] 02 November 2009 01:18AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (539)

You are viewing a single comment's thread.

Comment author: Kaj_Sotala 08 November 2009 12:31:46PM 0 points [-]

What's a brief but effective way to respond to the "an AI, upon realizing that it's programmed in a way its designer didn't intend to, would reprogram itself to be like the designer intended" fallacy? (Came up here: http://xuenay.livejournal.com/325292.html?thread=1229996#t1229996 )

Comment author: Vladimir_Nesov 08 November 2009 02:46:14PM 3 points [-]

I hope I'm not misinterpreting again, but this is a Giant cheesecake fallacy. The problem is that AI's decisions depend on its motive. "An AI, upon realizing that it's programmed in a way its designer didn't intend to, would try to convince the programmer that what the AI turned out to be is exactly what he intended in the first place", "An AI, upon realizing that it's programmed in a way its designer didn't intend to, would print a string "Styggron" to the console".

Comment author: Kaj_Sotala 08 November 2009 07:31:25PM 0 points [-]

Thanks, that's a good one. I'll try it.

Comment author: Cyan 08 November 2009 01:33:06PM 0 points [-]

How about: an AI can be smart enough to realize all of those things, and it still won't change its utility function. Then link Eliezer's short story about that exact scenario. (Can't find it in two minutes, but it's the one where the dude wakes up with a construct designed to be his perfect mate, and he rejects her because she's not his wife.)

Comment author: Eliezer_Yudkowsky 08 November 2009 07:54:18PM 2 points [-]