ChristianKl comments on LessWrong's attitude towards AI research - Less Wrong

8 Post author: Florian_Dietz 20 September 2014 03:02PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (49)

You are viewing a single comment's thread. Show more comments above.

Comment author: ChristianKl 22 September 2014 11:53:50AM 2 points [-]

Goals are standardly regarded as immune self modification, so an off switch, in my sense, would be too.

No. Part of what making an FAI is about is to produce agents that keeps their values constant under self modification. It's not something where you expect that someone accidently get's it right.

Comment author: TheAncientGeek 22 September 2014 12:50:11PM 2 points [-]

Tht isn't a fact. MIRI assumes goal stability is desirable for safety, but at the same time, MIRIs favourite UFAI is only possible with goal stability.

Comment author: [deleted] 22 September 2014 12:55:59PM *  3 points [-]

MIRIs favourite UFAI is only possible with goal stability.

A paperclip maximizer wouldn't become that much less scary if it accidentally turned itself into a paperclip-or-staple maximizer, though.

Comment author: [deleted] 22 September 2014 03:46:57PM 1 point [-]

What if it decided making paperclips was boring, and spent some time in deep meditation formulating new goals for itself?

Comment author: ChristianKl 22 September 2014 02:32:51PM 1 point [-]

Paperclip maximizers serve as illustration of a principle. I think that most MIRI folks consider UFAI to be more complicated than simple paperclip maximizers.

Goal stability also get's harder the more complicated the goal happens to be. A paperclip maximizer can have a off switch but at the same time prevent anyone from pushing that switch.