ChristianKl comments on LessWrong's attitude towards AI research - Less Wrong

8 Post author: Florian_Dietz 20 September 2014 03:02PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (49)

You are viewing a single comment's thread. Show more comments above.

Comment author: JoshuaMyer 21 September 2014 04:30:53PM 1 point [-]

Very thoughtful response. Thank you for taking the time to respond even though its clear that I am painfully new to some of the concepts here.

Why on earth would anyone build any "'tangible object' maximizer"? That seems particularly foolish.

AI boxing ... fantastic. I agree. A narrow AI would not need a box. Are there any tasks an AGI can do that a narrow AI cannot?

Comment author: ChristianKl 21 September 2014 08:44:07PM 2 points [-]

Why on earth would anyone build any "'tangible object' maximizer"? That seems particularly foolish.

Stock market computer programs are created in a way to maximize profits. In many domains computer programs are used to maximize some variable.

A narrow AI would not need a box.

What do you mean with "narrow"?

Comment author: TheAncientGeek 22 September 2014 11:04:43AM 1 point [-]

It's foolish to build things without off switches, which translates to building flexible iinteligences that only pursue one goal.

Comment author: ChristianKl 22 September 2014 11:11:55AM 2 points [-]

Nobody said something about no off switches. Off-switches mean that you need to understand that the program is doing something wrong to switch it off. A complex AGI that acts in complex ways might produce damage that you can't trace. Furthermore self modification might destroy an off switch.

Comment author: TheAncientGeek 22 September 2014 11:51:10AM 0 points [-]

By an off switch I mean a backup goal.

I know nobody mentioned it. The point is that Clippie has one main goal, any no backup goal, so off switches, in my sense, are being IMPLICITLY omitted.

Goals are standardly regarded as immune self modification, so an off switch, in my sense, would be too.

Comment author: ChristianKl 22 September 2014 11:53:50AM 2 points [-]

Goals are standardly regarded as immune self modification, so an off switch, in my sense, would be too.

No. Part of what making an FAI is about is to produce agents that keeps their values constant under self modification. It's not something where you expect that someone accidently get's it right.

Comment author: TheAncientGeek 22 September 2014 12:50:11PM 2 points [-]

Tht isn't a fact. MIRI assumes goal stability is desirable for safety, but at the same time, MIRIs favourite UFAI is only possible with goal stability.

Comment author: [deleted] 22 September 2014 12:55:59PM *  3 points [-]

MIRIs favourite UFAI is only possible with goal stability.

A paperclip maximizer wouldn't become that much less scary if it accidentally turned itself into a paperclip-or-staple maximizer, though.

Comment author: [deleted] 22 September 2014 03:46:57PM 1 point [-]

What if it decided making paperclips was boring, and spent some time in deep meditation formulating new goals for itself?

Comment author: ChristianKl 22 September 2014 02:32:51PM 1 point [-]

Paperclip maximizers serve as illustration of a principle. I think that most MIRI folks consider UFAI to be more complicated than simple paperclip maximizers.

Goal stability also get's harder the more complicated the goal happens to be. A paperclip maximizer can have a off switch but at the same time prevent anyone from pushing that switch.

Comment author: warbo 22 September 2014 12:13:17PM 1 point [-]

By an off switch I mean a backup goal. Goals are standardly regarded as immune self modification, so an off switch, in my sense, would be too.

This is quite a subtle issue.

If the "backup goal" is always in effect, eg. it is just another clause of the main goal. For example, "maximise paperclips" with a backup goal of "do what you are told" is the same as having the main goal "maximise paperclips while doing what you are told".

If the "backup goal" is a separate mode which we can switch an AI into, eg. "stop all external interaction", then it will necessarily conflict with the the AI's main goal: it can't maximise paperclips if it stops all external interaction. Hence the primary goal induces a secondary goal: "in order to maximise paperclips, I should prevent anyone switching me to my backup goal". These kind of secondary goals have been raised by Steve Omohundro.

Comment author: TheAncientGeek 22 September 2014 12:46:22PM 1 point [-]

You haven't dealt with the case where the safety goals are the primary ones.

These kinds of primary goals have been raised by Isaac Asimov.

Comment author: FeepingCreature 22 September 2014 04:13:52PM *  1 point [-]

The question of "what are the right safety goals" is what FAI research is all about.