Pablo_Stafforini comments on Open Thread, January 1-15, 2013 - Less Wrong

5 Post author: OpenThreadGuy 01 January 2013 06:09AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (333)

You are viewing a single comment's thread. Show more comments above.

Comment author: Pablo_Stafforini 19 January 2013 09:37:29PM *  1 point [-]

Tim, your comment highlights the potential conflict between CEV and FAI that I also mentioned previously. FAI is by definition not hostile to human beings, whereas CEV might permit, or even require, the extinction of all humanity. This may happen, for instance, if the process of coherent extrapolation shows that humans value certain superior beings more than they value themselves, and if the coexistence of humans and these beings is impossible.

When I pointed out this problem, both Kaj Sotala and Michael Anissimov replied that CEV can never condone hostile actions towards humanity because FAI is "defined as 'human-benefiting, non-human harming'". However, this reply just proves my point, namely that there is a potential internal inconsistency between CEV and FAI.

Comment author: TimS 20 January 2013 03:46:53AM 0 points [-]

Don't look at me to resolve that conflict. I think moral extrapolation is unlikely to output anything coherent if the reference class is sufficiently large to avoid the objections I raised above. And I can't think of any other plausible candidate to produce Friendly instructions for an AI.