TheAncientGeek comments on LessWrong's attitude towards AI research - Less Wrong

8 Post author: Florian_Dietz 20 September 2014 03:02PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (49)

You are viewing a single comment's thread. Show more comments above.

Comment author: TheAncientGeek 22 September 2014 12:50:11PM 2 points [-]

Tht isn't a fact. MIRI assumes goal stability is desirable for safety, but at the same time, MIRIs favourite UFAI is only possible with goal stability.

Comment author: [deleted] 22 September 2014 12:55:59PM *  3 points [-]

MIRIs favourite UFAI is only possible with goal stability.

A paperclip maximizer wouldn't become that much less scary if it accidentally turned itself into a paperclip-or-staple maximizer, though.

Comment author: [deleted] 22 September 2014 03:46:57PM 1 point [-]

What if it decided making paperclips was boring, and spent some time in deep meditation formulating new goals for itself?

Comment author: ChristianKl 22 September 2014 02:32:51PM 1 point [-]

Paperclip maximizers serve as illustration of a principle. I think that most MIRI folks consider UFAI to be more complicated than simple paperclip maximizers.

Goal stability also get's harder the more complicated the goal happens to be. A paperclip maximizer can have a off switch but at the same time prevent anyone from pushing that switch.