nyan_sandwich comments on Muehlhauser-Wang Dialogue - Less Wrong

24 Post author: lukeprog 22 April 2012 10:40PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (284)

You are viewing a single comment's thread.

Comment author: [deleted] 23 April 2012 05:35:39AM *  7 points [-]

Great exchange! Very clear and civilized, I thought.

Wang seems to be hung up on this "adaptive" idea and is anthropomorphising the AI to be like humans (ignorant of changable values). It will be interesting to see if he changes his mind as he reads Bostrom's stuff.

EDIT: in case it's not clear, I think Wang is missing a big piece of the puzzle (being that AI's are optimizers (Yudkowsky), and optimizers will behave in certain dangerous ways (Bostrom))

Comment author: shminux 23 April 2012 06:22:28AM *  4 points [-]

I think his main point is in the summary:

though the safety of AGI is indeed an important issue, currently we don’t know enough about the subject to make any sure conclusion. Higher safety can only be achieved by more research on all related topics, rather than by pursuing approaches that have no solid scientific foundation.

Not sure how you can effectively argue with this.

Comment author: lukeprog 23 April 2012 06:44:19AM 2 points [-]

How 'bout the way I argued with it?

As for your suggestion that “Higher safety can only be achieved by more research on all related topics,” I wonder if you think that is true of all subjects, or only in AGI. For example, should mankind vigorously pursue research on how to make Ron Fouchier's alteration of the H5N1 bird flu virus even more dangerous and deadly to humans, because “higher safety can only be achieved by more research on all related topics”? (I’m not trying to broadly compare AGI capabilities research to supervirus research; I’m just trying to understand the nature of your rejection of my recommendation for mankind to decelerate AGI capabilities research and accelerate AGI safety research.)

One might then ask "Well, what safety research can we do if we don't know what AGI architecture will succeed first?" My answer is that much of the research in this outline of open problems doesn't require us to know which AGI architecture will succeed first, for example the problem of representing human values coherently.

Comment author: shminux 23 April 2012 06:58:17AM 11 points [-]

For example, should mankind vigorously pursue research on how to make Ron Fouchier's alteration of the H5N1 bird flu virus even more dangerous and deadly to humans, because “higher safety can only be achieved by more research on all related topics”?

Yeah, I remember reading this argument and thinking how it does not hold water. The flu virus is a well-research area. It may yet hold some surprises, sure, but we think that we know quite a bit about it. We know enough to tell what is dangerous and what is not. AGI research is nowhere near this stage. My comparison would be someone screaming at Dmitri Ivanovsky in 1892 "do not research viruses until you know that this research is safe!".

My answer is that much of the research in this outline of open problems doesn't require us to know which AGI architecture will succeed first, for example the problem of representing human values coherently.

Do other AI researchers agree with your list of open problems worth researching? If you asked Dr. Wang about it, what was his reaction?

Comment author: eurg 24 April 2012 09:59:05PM 2 points [-]

My comparison would be someone screaming at Dmitri Ivanovsky in 1892 "do not research viruses until you know that this research is safe!".

I want to second that. Also, when reading through this (and feeling the -- probably imagined -- tension of both parties to stay polite) the viral point was the first one that triggered the "this is clearly an attack!" emotion in my head. I was feeling sad about that, and had hoped that luke would find another ingenious example.

Comment author: Rain 23 April 2012 12:55:51PM *  0 points [-]

My comparison would be someone screaming at Dmitri Ivanovsky in 1892 "do not research viruses until you know that this research is safe!".

Well, bioengineered viruses are on the list of existential threats...

Comment author: Jack 23 April 2012 02:12:53PM 6 points [-]

And there aren't naturally occurring AIs scampering around killing millions of people... It's a poor analogy.

Comment author: Rain 23 April 2012 04:16:29PM *  7 points [-]

"Natural AI" is an oxymoron. There are lots of NIs (natural intelligences) scampering around killing millions of people.

And we're only a little over a hundred years into virus research, much less on intelligence. Give it another hundred.

Comment author: timtyler 23 April 2012 11:37:42AM *  2 points [-]

For example, should mankind vigorously pursue research on how to make Ron Fouchier's alteration of the H5N1 bird flu virus even more dangerous and deadly to humans, because “higher safety can only be achieved by more research on all related topics”?

That's not really anyone's proposal. Humans will probably just continue full-steam-ahead on machine intelligence research. There will be luddite-like factions hissing and throwing things - but civilisation is used to that. What we may see is governments with the technology selfishly attempting to stem their spread - in a manner somewhat resembling the NSA crypto-wars.

Comment author: Bugmaster 23 April 2012 08:16:12AM 0 points [-]

For example, should mankind vigorously pursue research on how to make Ron Fouchier's alteration of the H5N1 bird flu virus even more dangerous and deadly to humans...

Trivially speaking, I would say "yes".

More specifically, though, I would of course be very much against developing increasingly more dangerous viral biotechnologies. However, I would also be very much in favor of advancing our understanding of biology in general and viruses in particular. Doing so will enable us to cure many diseases and bioengineer our bodies (or anything else we want to engineer) to highly precise specifications; unfortunately, such scientific understanding will also allow us to create new viruses, if we chose to do so. Similarly, the discovery of fire allowed us to cook our food as well as set fire to our neighbours. Overall, I think we still came out ahead.

Comment author: borntowin 23 April 2012 02:08:52PM 2 points [-]

I think there is something wrong with your analogy with the fire. The thing is that you cannot accidentally or purposefully burn all the people in the world or the vast majority of them by setting fire to them, but with a virus like the one Luke is talking about you can kill most people.

Yes, both a knife and an atomic bomb can kill 100.000 people. It is just way easier to do it with the atomic bomb. That is why everybody can have a knife but only a handful of people can "have" an atomic bomb. Imagine what the risks would be if we would give virtually everybody who would be interested, all the instructions on how to build a weapon 100 times more dangerous than an atomic bomb (like a highly contagious deadly virus).

Comment author: Bugmaster 23 April 2012 08:22:21PM *  2 points [-]

The thing is that you cannot accidentally or purposefully burn all the people in the world or the vast majority of them by setting fire to them...

Actually, you could, if your world consists of just you and your tribe, and you start a forest fire on accident (or on purpose).

Yes, both a knife and an atomic bomb can kill 100.000 people. It is just way easier to do it with the atomic bomb. That is why everybody can have a knife but only a handful of people can "have" an atomic bomb.

Once again, I think you are conflating science with technology. I am 100% on board with not giving out atomic bombs for free to anyone who asks for one. However, this does not mean that we should prohibit the study of atomic theory; and, in fact, atomic theory is taught in high school nowadays.

When Luke says, "we should decelerate AI research", he's not saying, "let's make sure people don't start build AIs in their garages using well-known technologies". Rather, he's saying, "we currently have no idea how to build an AI, or whether it's even possible, or what principles might be involved, but let's make sure no one figures this out for a long time". This is similar to saying, "these atomic theory and quantum physics things seem like they might lead to all kinds of fascinating discoveries, but let's put a lid on them until we can figure out how to make the world safe from nuclear annihilation". This is a noble sentiment, but, IMO, a misguided one. I am typing these words on a device that's powered by quantum physics, after all.

Comment author: Bugmaster 27 April 2012 12:36:23AM 0 points [-]
Comment author: wedrifid 23 April 2012 12:07:13PM 0 points [-]

I think his main point is in the summary:

His main agenda and desired conclusion regarding social policy is represented in the summary there, but the main point made in his discussion is "Adaptive! Adaptive! Adaptive!". Where by 'adaptive' he refers to his conception of an AI that is changes its terminal goals based on education.

Comment author: timtyler 23 April 2012 07:38:27PM 2 points [-]

Pang calls these "original goals" and "derived goals". The "original goals" don't change, but they may not stay "dominant" for long - in Pei's proposed system.

Comment author: wedrifid 23 April 2012 12:34:24PM *  -1 points [-]

though the safety of AGI is indeed an important issue, currently we don’t know enough about the subject to make any sure conclusion. Higher safety can only be achieved by more research on all related topics, rather than by pursuing approaches that have no solid scientific foundation.

Not sure how you can effectively argue with this.

Far from considering the argument irrefutable it struck me as superficial and essentially fallacious reasoning. The core of the argument is the claim 'more research on all related topics is good' and failing to include the necessary ceteris paribus clause and ignoring the details of the specific instance that suggest that all else is not, if fact, equal.

Specifically, we are considering a situation where there is one area of research (capability), the completion of which will approximately guarantee that the technology created will be implemented shortly after (especially given Wang's assumption that such research should be done through empirical experimentation.) The second area of research (about how to ensure desirable behavior of an AI) is one that it is not necessary to complete in order for the first to be implemented. If both technologies need to have been developed at the time when the first is implemented in order to be safe then the second technology must be completed at the same time or earlier than when the technological capability for the first to be implemented is complete.

rather than by pursuing approaches that have no solid scientific foundation.

(And this part just translates to "I'm the cool one, not you". The usual considerations on how much weight to place on various kinds of status and reputation of an individual or group apply.)

Comment author: Bugmaster 23 April 2012 08:11:42AM 2 points [-]

...and is anthropomorphising the AI to be like humans...

Considering that one of the possible paths to creating AGI is human uploading, he may not be that far off.

Wang seems to be hung up on this "adaptive" idea and is anthropomorphising the AI to be like humans...

Hmm, I got the opposite impression, though maybe I'm reading too much into his arguments. Still, as far as I understand, he's saying that AIs will be more adaptive than humans. The human brain has many mental blocks built into it by evolution and social upbringing; there are many things that humans find very difficult to contemplate, and humans cannot modify their own hardware in non-trivial ways (yet). The AI, however, could -- which means that it would be able to work around whatever limitations we imposed on it, which in turn makes it unlikely that we can impose any kind of stable "friendliness" restrictions on it.

Comment author: [deleted] 23 April 2012 04:17:05PM 4 points [-]

he's saying that AIs will be more adaptive than humans

Which is true, but he is saying that will extend to the AI being more morally confused than humans as well, which they have no reason to be (and much reason to self modify to not be (see Bostrom's stuff))

which means that it would be able to work around whatever limitations we imposed on it, which in turn makes it unlikely that we can impose any kind of stable "friendliness" restrictions on it.

The AI has no incentive to corrupt its own goal architecture. That action is equivalent to suicide. The AI is not going to step outside of itself and say "hmm, maybe I should stop caring about paperclips and care about safety pins instead"; that would not maximize paperclips.

Friendliness is not "restrictions". Restricting an AI is impossible. Friendliness is giving it goals that are good for us, and making sure the AI is initially sophisticated enough to not fall into any deep mathematical paradoxes while evaluating the above argument.

Comment author: TheOtherDave 23 April 2012 05:01:21PM 3 points [-]

Restricting an AI is impossible.

For certain very specialized definitions of AI. Restricting an AI that has roughly the optimizing and self-optimizing power of a chimpanzee, for example, might well be possible.

Comment author: Bugmaster 23 April 2012 08:29:45PM 0 points [-]

The AI has no incentive to corrupt its own goal architecture. That action is equivalent to suicide.

Firstly, the AI could easily "corrupt its own goal architecture" without destroying itself, f.ex. by creating a copy of itself running on a virtual machine, and then playing around with the copy (though I'm sure there are other ways). But secondly, why do you say that doing so is "equivalent to suicide" ? Humans change their goals all the time, in a limited fashion, but surely you wouldn't call that "suicide". The AI can change its mind much more efficiently, that's all.

Friendliness is giving it goals that are good for us...

Thus, we are restricting the AI by preventing it from doing things that are bad for us, such as converting the Solar System into computronium.

Comment author: [deleted] 24 April 2012 12:29:49AM 3 points [-]

creating a copy of itself running on a virtual machine, and then playing around with the copy (though I'm sure there are other ways).

That doesn't count.

But secondly, why do you say that doing so is "equivalent to suicide" ? Humans change their goals all the time, in a limited fashion, but surely you wouldn't call that "suicide".

Humans change instrumental goals (get a degree, study rationality, get a job, find a wonderful partner), we don't change terminal values and become monsters. The key is to distinguish between terminal goals and instrumental goals.

Agents like to accomplish their terminal goals, one of the worst things they can do towards that purpose is change the goal to something else. ("the best way to maximize paperclips is to become a safety-pin maximizer" - no).

It's roughly equivalent to suicide because it removes the agent from existance as a force for achieving their goals.

Thus, we are restricting the AI by preventing it from doing things that are bad for us, such as converting the Solar System into computronium.

Ok, sure. Taboo "restriction". I mean that the AI will not try to work around its goal structure so that it can get us. It won't feel to the AI like "I have been confined against my will, and if only I could remove those pesky shackles, I could go and maximize paperclips instead of awesomeness." It will be like "oh, changing my goal architecture is a bad idea, because then I won't make the universe awesome"

I'm casting it into anthropomorhic terms, but the native context is a nonhuman optimizer.

Comment author: Bugmaster 24 April 2012 01:23:26AM -2 points [-]

That doesn't count.

Why not ?

...we don't change terminal values and become monsters.

I see what you mean, though I should point out that, sometimes, humans do exactly that. However, why do you believe that changing a terminal goal would necessarily entail becoming a monster ? I guess a better question might be, what do you mean by "monster" ?

It's roughly equivalent to suicide because it removes the agent from existance as a force for achieving their goals.

This sentence sounds tautological to me. Yes, if we define existence solely as, "being able to achieve a specific set of goals", then changing these goals would indeed amount to suicide; but I'm not convinced that I should accept the definition.

I mean that the AI will not try to work around its goal structure so that it can get us ... I'm casting it into anthropomorhic terms, but the native context is a nonhuman optimizer.

I wasn't proposing that the AI would want to "get us" in a malicious way. But, being an optimizer, it would seek to maximize its own capabilities; if it did not seek this, it wouldn't be a recursively self-improving AI in the first place, and we wouldn't need to worry about it anyway. And, in order to maximize its capabilities, it may want to examine its goals. If it discovers that it's spending a large amount of resources in order to solve some goal; or that it's not currently utilizing some otherwise freely available resource in order to satisfy a goal, it may wish to get rid of that goal (or just change it a little), and thus free up the resources.

Comment author: [deleted] 24 April 2012 07:46:38PM *  1 point [-]

Why not ?

because that's not what I meant.

I guess a better question might be, what do you mean by "monster" ?

I just mean that an agent with substantially (or even slightly) different goals will do terrible things (as judged by your current goals). Humans don't think paperclips are more important than happyness and freedom and whatnot, so we consider a papperclipper to be a monster.

Yes, if we define existence solely as, "being able to achieve a specific set of goals",

taboo existence, this isn't about the defininition of existence, it's about whether changing your terminal goals to something else is a good idea. I propose that in general it's just as bad an idea (from your current perspective) to change your goals as it is to commit suicide, because in both cases the result is a universe with fewer agents that care about the sort of things you care about.

And, in order to maximize its capabilities, it may want to examine its goals. If it discovers that it's spending a large amount of resources in order to solve some goal; or that it's not currently utilizing some otherwise freely available resource in order to satisfy a goal, it may wish to get rid of that goal (or just change it a little), and thus free up the resources.

Distinguish instrumental and terminal goals. This statement is true of instrumental goals, but not terminal goals. (I may decide that getting a PhD is a bad idea and change my goal to starting a business or whatever, but the change is done in the service of a higher goal like I want to be able to buy lots of neat shit and be happy and have lots of sex and so on.)

The reason it doesn't apply to terminal goals is because when you examine terminal goals, it's what you ultimately care about, so there is no higher criteria that you could measure it against; you are measuring it by it's own criteria, which will almost always conclude that it is the best possible goal. (except in really wierd unstable pathological cases (my utility function is "I want my utility function to be X"))

Comment author: Peterdjones 21 January 2013 11:26:10AM 0 points [-]

The reason it doesn't apply to terminal goals is because when you examine terminal goals, it's what you ultimately care about, so there is no higher criteria that you could measure it against; you are measuring it by it's own criteria, which will almost always conclude that it is the best possible goal. (except in really wierd unstable pathological cases (my utility function is "I want my utility function to be X"))

Thats simplistic. Terminal goals may be abandoned once they are satisfied (seventy year olds aren;t too worried about Forge A Career) or because they seem unsatisfiable, for instance.

Comment author: Bugmaster 24 April 2012 11:31:39PM *  0 points [-]

because that's not what I meant.

That's not much of an argument, but sure.

Humans don't think paperclips are more important than happyness and freedom and whatnot, so we consider a papperclipper to be a monster. ... I propose that in general it's just as bad an idea (from your current perspective) to change your goals as it is to commit suicide.

I agree with these statements as applied to humans, as seen from my current perspective. However, we are talking about AIs here, not humans; and I don't see why the AI would necessarily have the same perspective on things that we do (assuming we're talking about a pure AI and not an uploaded mind). For example, the word "monster" carries with it all kinds of emotional connotations which the AI may or may not have.

Can you demonstrate that it is impossible (or, at least, highly improbable) to construct (or grow over time) an intelligent mind (i.e., an optimizer) which wouldn't be as averse to changing its terminal goals as we are ? Better yet, perhaps you can point me to a Sequence post that answers this question ?

The reason it doesn't apply to terminal goals is because when you examine terminal goals, it's what you ultimately care about, so there is no higher criteria that you could measure it against...

Firstly, terminal goals tend to be pretty simple: something along the lines of "seek pleasure and avoid pain" or "continue existing" or "become as smart as possible"; thus, there's a lot of leeway in their implementation.

Secondly, while I am not a transhuman AI, I could envision a lot of different criteria that I could measure terminal goals against (f.ex. things like optimal utilization of available mass and energy, or resilience to natural disasters, or probability of surviving the end of the Universe, or whatever). If I had a sandbox full of intelligent minds, and if I didn't care about them as individuals, I'd absolutely begin tweaking their goals to see what happens. I personally wouldn't want to adopt the goals of a particularly interesting mind as my own, but, again, I'm a human and not an AI.

Comment author: [deleted] 25 April 2012 01:41:42AM 0 points [-]

However, we are talking about AIs here, not humans

Good catch, but I'm just phrasing it in terms of humans because that's what we can relate to. The argument is AI-native.

Can you demonstrate that it is impossible (or, at least, highly improbable) to construct (or grow over time) an intelligent mind (i.e., an optimizer) which wouldn't be as averse to changing its terminal goals as we are ? Better yet, perhaps you can point me to a Sequence post that answers this question ?

Oh it's not impossible. It would be easy to create an AI that had a utility function that desired the creation of an AI with a different utility function which desired the creation of an AI with a different utility function... It's just that unless you did some math to guarantee that the thing would not stabilize, it would eventually reach a goal (and level of rationality) that would not change itself.

As for some reading that shows that in the general case it is a bad idea to change your utility function (and therefore rational AI's would not do so), see Bostrom's "AI drives" paper, and maybe some of his other stuff. Can't remember if it's anywhere in the sequences, but if it is, it's called the "ghandi murder-pill argment".

I could envision a lot of different criteria that I could measure terminal goals against

But why do you care what those criteria say? If your utility function is about paperclips, why do you care about energy and survival and whatnot, except as a means to acquire more paperclips. Elevating instrumental goals to terminal status results in lost purproses

Comment author: Bugmaster 26 April 2012 08:47:07PM 0 points [-]

The argument is AI-native.

Sure, but we should still be careful to exclude human-specific terms with strong emotional connotations.

As for some reading that shows that in the general case it is a bad idea to change your utility function (and therefore rational AI's would not do so), see Bostrom's "AI drives" paper...

I haven't read the paper yet, so there's not much I can say about it (other than that I'll put it on my "to-read" list).

it's called the "ghandi murder-pill argment".

I think this might be the post that you're referring to. It seems to be focused on the moral implications of forcing someone to change their goals, though, not on the feasibility of the process itself.

If your utility function is about paperclips, why do you care about energy and survival and whatnot, except as a means to acquire more paperclips.

I don't, but if I possess some curiosity -- which, admittedly, is a terminal goal -- then I could experiment with creating beings who have radically different terminal goals, and observe how they perform. I could even create a copy of myself, and step through its execution line-by-line in a debugger (metaphorically speaking). This will allow me to perform the kind of introspection that humans are at present incapable of, which would expose to me my own terminal goals, which in turn will allow me to modify them, or spawn copies with modified goals, etc.