jacob_cannell comments on Brainstorming additional AI risk reduction ideas - Less Wrong

12 Post author: John_Maxwell_IV 14 June 2012 07:55AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (37)

You are viewing a single comment's thread. Show more comments above.

Comment author: jacob_cannell 14 June 2012 09:38:06PM 1 point [-]

This may have some value, but probably not towards actually making AI more moral/friendly on average. Conversing about morality can demonstrate knowledge of morality, but does little to demonstrate evidence of being moral/friendly. Example: a psychopath would not necessarily have any difficulty passing this Moral Turing Test.

Comment author: Viliam_Bur 15 June 2012 08:50:07AM *  3 points [-]

On the other hand, a machine could fail a morality test simply by saying something controversial, or just failing to signal properly. For example atheism could be considered immoral by religious people; they could conclude that the machine is missing a part of human utility function. Or if some nice and correct belief has bad consequences, but humans compartmentalize it away and the machine would point it out explicitly, that could be percieved as a moral failure.

If the machine is allowed to lie, passing this test could just mean the machine is a skilled psychopath. If the machine is not allowed to lie, failing this test could just mean humans confuse signalling with the real thing.

Comment author: Emile 14 June 2012 09:52:41PM 0 points [-]

I agree, the goal is to get humans to think about programming some forms of moral reasoning, even if it's far from sufficient (and it's far from being the hardest part of FAI).