You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

David_Gerard comments on Google may be trying to take over the world - Less Wrong Discussion

22 [deleted] 27 January 2014 09:33AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (133)

You are viewing a single comment's thread. Show more comments above.

Comment author: David_Gerard 27 January 2014 07:35:06PM *  4 points [-]

Um, surely if you take (a) people with a track record of successful achievement in an area (b) people without a track record of success but who think they know a lot about the area, the presumption that (a) is more likely to know what they're talking about should be the default presumption. It may of course not work out that way, but that would surely be the way to bet.

Comment author: fortyeridania 27 January 2014 07:44:20PM -2 points [-]

Yes, I agree, but that is only part of the story, right?

What if autodidacts, in their untutored excitability, are excessively concerned about a real risk? Or if a real risk has nearly all autodidacts significantly worried, but only 20% of actual experts significantly worried? Wouldn't that falsify /u/private_messaging's assertion? And what's so implausible about that scenario? Shouldn't we expect autodidacts' concerns to be out of step with real risks?

Comment author: V_V 27 January 2014 08:34:19PM *  4 points [-]

What if autodidacts, in their untutored excitability, are excessively concerned about a real risk?

If autodidacts are excessively concerned, then why would it be worth for experts to listen to them?

Comment author: fortyeridania 27 January 2014 10:42:03PM 0 points [-]

It may not be. I was not taking issue with the claim "Experts need not listen to autodidacts." I was taking issue with the claim "Given a real risk, experts are more likely to be concerned than autodidacts are."

Comment author: V_V 27 January 2014 11:50:29PM *  5 points [-]

I would assume that experts are likely to be concerned to an extent more appropriate to the severity of the risk than autodidacts are.

There can be exceptions, of course, but when non-experts make widely more extreme claims than experts do on some issue, especially a strongly emotively charged issue (e.g. the End of the World), unless they can present really compelling evidence and arguments, Dunning–Kruger effect seems to be the most likely explanation.

Comment author: fortyeridania 28 January 2014 12:24:13AM 0 points [-]

I would assume that experts are likely to be concerned to an extent more appropriate to the severity of the risk than autodidacts are.

That is exactly what I would assume too. Autodidacts' risk estimates should be worse than experts'. It does not follow that autodidacts' risk estimates should be milder than experts', though. The latter claim is what I meant to contest.

Comment author: private_messaging 27 January 2014 11:46:18PM *  5 points [-]

"Autodidacts" was in quotes for a reason.

Let's talk about some woo that you're not interested in. E.g. health risks of thymerosal and vaccines in general. Who's more likely to notice it, some self proclaimed "autodidacts", or normal biochemistry experts? Who noticed the possibility of a nuke, back-then conspiracy theorists or scientists? Was Semmelweis some weird outsider, or was he a regular medical doctor with medical training? And so on and so forth.

Right now, experts are concerned with things like nuclear war, run-away methane releases, epidemics, and so on, while various self proclaimed existential risk people (mostly philosophers) seem to be to greater or lesser extent neglecting said risks in favor of movie plot dangers such as runaway self improving AI or perhaps totalitarian world government. (Of course if you listen to said x-risk folks, they're going to tell you that it's because the real experts are wrong.)

Comment author: fortyeridania 28 January 2014 12:29:26AM -1 points [-]

Who's more likely to notice it, some self proclaimed "autodidacts", or normal biochemistry experts? Who noticed the possibility of a nuke, back-then conspiracy theorists or scientists? Was Semmelweis some weird outsider, or was he a regular medical doctor with medical training?

All are good and relevant examples, and they all support the claim in question. Thanks!

But your second paragraph supports the opposite claim. (Again, the claim in question is: Experts are more likely to be concerned over risks than autodidacts are.) In the second paragraph, you give a couple "movie plot" risks, and note that autodidacts are more concerned about them than experts are. Those would therefore be cases of autodidacts being more concerned about risks than experts, right?

If the claim were "Experts have more realistic risk estimates than autodidacts do," then I would readily agree. But you seem to have claimed that autodidacts' risk estimates aren't just wrong--they are biased downward. Is that indeed what you meant to claim, or have I misunderstood you?

Comment author: private_messaging 28 January 2014 12:50:39AM *  2 points [-]

What I said was that "autodidacts" (note the scare quotes) are more likely to fail to notice some genuine risk, than the experts are. E.g. if there's some one specific medication that poses risk for a reason X, those anti vaxers are extremely unlikely to spot that, due to the lack of necessary knowledge and skills.

By "autodidacts" in scare quotes I mean interested and somewhat erudite laymen who may have read a lot of books but clearly did very few exercises from university textbooks (edit: or any other feedback providing exercises at all).

Comment author: fortyeridania 28 January 2014 01:10:04AM 0 points [-]
  1. I understand the scare quotes.

  2. I agree that autodidacts "are more likely to fail to notice some genuine risk, than experts are."

  3. But autodidacts are also more likely to exaggerate other genuine risks than experts are, are they not?

  4. If (3) is true, then doesn't that undermine the claim "Experts are more likely to be concerned over risks than autodidacts are"?

Comment author: private_messaging 28 January 2014 01:17:04AM *  1 point [-]

What I said was:

Does that mean there is a terrible ignored risk? No, when there is a real risk, the brightest people of extreme and diverse intellectual accomplishment are the ones most likely to be concerned about it (and various "autodidacts" are most likely to fail to notice the risk).

Besides, being more concerned is not the same as being more likely to be concerned. Just as being prone to panic doesn't automatically make you better at hearing danger.

Comment author: fortyeridania 28 January 2014 05:07:42AM 0 points [-]

being more concerned is not the same as being more likely to be concerned

True, and I see that this distinction undercuts one of the ways there could be more autodidact concern than expert concern. But there is at least one more way, which I suggested earlier.

Imagine a world populated by a hundred experts, a hundred autodidacts, and a risk. Let E be the number of experts concerned about the risk, and A be the number of concerned autodidacts.

I interpret you as saying that E is greater than A. Is this a correct interpretation?

To the claim that E > A, I am saying "not necessarily." Here is how.

Since the risk is a genuine risk, we assume that nearly all the experts are concerned. So we set E = 95. Now suppose those without formal training all suffer from the same common pitfalls, and so tend to make errors in the same direction. Suppose that due to these errors, autodidacts with their little learning are even more likely to be concerned. If they were all better trained, they would all relax a bit, and some would relax enough to cross the line into "not concerned" territory.

The above scenario seems perfectly plausible to me; is there some problem with it that I have missed? Does it miss the point? It is not the most likely scenario, but it's far from impossible, and you seem to have cavalierly ruled it out. Hence my original request for a source.

Comment author: private_messaging 27 January 2014 11:28:02PM *  5 points [-]

To clarify, I have nothing anything against self educated persons. Some do great things. The "autodidacts" was specifically in quotes.

What is implausible, is this whole narrative where you have a risk obvious enough that people without any relevant training can see it (by the way of that paperclipping argument), yet the relevant experts are ignoring it. Especially when the idea of an intelligence turning against it's creator is incredibly common in fiction, to the point that nobody has to form that idea on their own.

Comment author: [deleted] 28 January 2014 03:54:03PM 3 points [-]

In general, current AGI architectures work via reinforcement learning: reward and punishment. Relevant experts are worried about what will happen when an AGI with the value-architecture of a pet dog finds that it can steal all the biscuits from the kitchen counter without having to do any tricks.

They are less worried about their current creations FOOMing into god-level superintelligences, because current AI architectures are not FOOMable, and it seems quite unlikely that you can create a self-improving ultraintelligence by accident. Except when that's exactly what they plan for them to do (ie: Shane Legg).

Juergen Schmidhuber gave an interview on this very website where he basically said that he expects his Goedel Machines to undergo a hard takeoff at some point, with right and wrong being decided retrospectively by the victors of the resulting Artilect War. He may have been trolling, but it's a bit hard to tell.

Comment author: private_messaging 28 January 2014 04:18:03PM 1 point [-]

I'd need to have links and to read it by myself.

With regards to reinforcement learning, one thing to note is that the learning process is in general not the same thing as the intelligence that is being built by the learning process. E.g. if you were to evolve some ecosystem of programs by using "rewards" and "punishments", the resulting code ends up with distinct goals (just as humans are capable of inventing and using birth control). Not understanding this, local genuises of the AI risk been going on about "omg he's so stupid it's going to convert the solar system to smiley faces" with regards to at least one actual AI researcher.

Comment author: [deleted] 28 January 2014 04:31:05PM 1 point [-]

I'd need to have links and to read it by myself.

Here is his interview. It's very, very hard to tell if he's got his tongue firmly in cheek (he refers to minds of human-level intelligence and our problems as being "small"), or if he's enjoying an opportunity to troll the hell out of some organization with a low opinion of his work.

With regards to reinforcement learning, one thing to note is that the learning process is in general not the same thing as the intelligence that is being built by the learning process.

With respect to genetic algorithms, you are correct. With respect to something like neural networks (real world stuff) or AIXI (pure theory), you are incorrect. This is actually why machine-learning experts differentiate between evolutionary algorithms ("use an evolutionary process to create an agent that scores well on X") versus direct learning approaches ("the agent learns to score well on X").

Not understanding this, local genuises of the AI risk been going on about "omg he's so stupid it's going to convert the solar system to smiley faces" with regards to at least one actual AI researcher.

What, really? I mean, while I do get worried about things like Google trying to take over the world, that's because they're ideological Singulatarians. They know the danger line is there, and intend to step over it. I do not believe that most competent Really Broad Machine Learning (let's use that nickname for AGI) researchers are deliberately, suicidally evil, but then again, I don't believe you can accidentally make a dangerous-level AGI (ie: a program that acts as a VNM-rational agent in pursuit of an inhumane goal).

Accidental and evolved programs are usually just plain not rational agents, and therefore pose rather more limited dangers (crashing your car, as opposed to killing everyone everywhere).

Comment author: private_messaging 28 January 2014 05:03:15PM *  2 points [-]

With respect to something like neural networks (real world stuff)

Well, the neural network in my head doesn't seem to want to maximize the reward signal itself, but instead is more interested in maximizing values imprinted into it by the reward signal (which it can do even by hijacking the reward signal or even by administering "punishments"). Really, reward signal is not utility, period. Teach the person to be good, and they'll keep themselves good by punishing/rewarding themselves.

or AIXI (pure theory), you are incorrect.

I don't think it's worth worrying about the brute force iteration over all possible programs. Once you stop iterating over the whole solution space in the learning method itself, the learning method faces the problem that it can not actually ensure that the structures constructed by the learning method don't have separate goals (nor is it desirable to ensure such, as you would want to be able to teach values to an agent using the reward signal).

Comment author: [deleted] 28 January 2014 05:28:35PM 1 point [-]

Well, the neural network in my head doesn't seem to want to maximize the reward signal itself, but instead is more interested in maximizing values imprinted into it by the reward signal (which it can do even by hijacking the reward signal or even by administering "punishments"). Really, reward signal is not utility, period. Teach the person to be good, and they'll keep themselves good by punishing/rewarding themselves.

Firstly, I was talking about artificial neural networks, which do indeed function as reinforcement learners, by construction and mathematical proof.

Secondly, human beings often function as value learners ("learn what is good via reinforcement, but prefer a value system you're very sure about over a reward that seems to contradict the learned values") rather than reinforcement learners. Value learners, in fact, are the topic of a machine ethics paper from 2011, by Daniel Dewey.

the learning method faces the problem that it can not actually ensure that the structures constructed by the learning method don't have separate goals (nor is it desirable to ensure such, as you would want to be able to teach values to an agent using the reward signal).

Sorry, could you explain this better? It doesn't match up with how the field of machine learning usually works.

Yes, any given hypothesis a learner has about a target function is only correct to within some probability of error. But that probability can be very small.

Comment author: private_messaging 28 January 2014 06:05:19PM *  3 points [-]

With the smiley faces, I am referring to disagreement with Hibbard, summarized e.g. here on wikipedia

Secondly, human beings often function as value learners ("learn what is good via reinforcement, but prefer a value system you're very sure about over a reward that seems to contradict the learned values") rather than reinforcement learners. Value learners, in fact, are the topic of a machine ethics paper from 2011, by Daniel Dewey.

You're speaking as if value learners were not a subtype of reinforcement learners.

For a sufficiently advanced AI, i.e. one that learns to try different counter-factual actions on a world model, it is essential to build a model of the reward, which is to be computed on the counter-factual actions. It's this model of the reward that is specifying which action gets chosen.

Yes, any given hypothesis a learner has about a target function is only correct to within some probability of error. But that probability can be very small.

Looks like presuming a super-intelligence from the start.

Comment author: [deleted] 28 January 2014 06:33:36PM *  0 points [-]

With the smiley faces, I am referring to disagreement with Hibbard, summarized e.g. here on wikipedia

Right, and that wikipedia article refers to stuff Eliezer was writing more than ten years ago. That stuff is nowhere near state-of-the-art machine ethics.

(I think this weekend I might as well blog some decent verbal explanations of what is usually going on in up-to-date machine ethics on here, since a lot of people appear to confuse real, state-of-the-art work with either older, superseded ideas or very intuitive fictions.

Luckily, it's a very young field, so it's actually possible for some bozo like me to know a fair amount about it.)

You're speaking as if value learners were not a subtype of reinforcement learners.

That's because they are not. These are precise mathematical terms being used here, and while they are similar (for instance, I'd consider a Value Learner closer to a reinforcement learner than to a fixed direct-normativity utility function), they're not identical, neither is one a direct supertype of the other.

For a sufficiently advanced AI, i.e. one that learns to try different counter-factual actions on a world model, it is essential to build a model of the reward, which is to be computed on the counter-factual actions. It's this model of the reward that is specifying which action gets chosen.

This intuition is correct, regarding reinforcement learners. It is slightly incorrect regarding value learners, but how precisely it is incorrect is at the research frontier.

Looks like presuming a super-intelligence from the start.

No, I didn't say the target function was so complex as to require superintelligence. If I have a function f(x) = x + 1, a learner will be able to learn that this is the target function to within a very low probability of error, very quickly, precisely because of its simplicity.

The simpler the target function, the less training data needed to learn it in a supervised paradigm.