Below is part 2 of the first draft of my book Ethics and Superintelligence. Your comments and constructive criticisms are much appreciated.

This is not a book for a mainstream audience. Its style is that of contemporary Anglophone philosophy. Compare to, for example, Chalmers' survey article on the singularity.

Bibliographic references and links to earlier parts are provided here.

Part 2 is below...

 

 

 

 

 

***

Late in the Industrial Revolution, Samuel Butler (1863) worried about what might happen when machines become more capable than the humans who designed them:

…we are ourselves creating our own successors; we are daily adding to the beauty and delicacy of their physical organisation; we are daily giving them greater power and supplying by all sorts of ingenious contrivances that self-regulating, self-acting power which will be to them what intellect has been to the human race. In the course of ages we shall find ourselves the inferior race.

…the time will come when the machines will hold the real supremacy over the world and its inhabitants…

By the time of the computer, Alan Turing (1950) realized that machines will one day be capable of genuine thought:

I believe that at the end of the century…  one will be able to speak of machines thinking without expecting to be contradicted.

Turing (1951/2004) concluded:

…it seems probable that once the machine thinking method has started, it would not take long to outstrip our feeble powers... At some stage therefore we should have to expect the machines to take control…

All-powerful machines are a staple of science fiction, but one of the first serious arguments that such a scenario is likely came from the statistician I.J. Good (1965):

Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an “intelligence explosion”, and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make.

Vernor Vinge (1993) called this future event the “technological singularity.” Though there are several uses of the term “singularity” in futurist circles (Yudkowky 2007), I will always use the term to refer to Good’s predicted intelligence explosion.

David Chalmers (2010) introduced another terminological convention that I will borrow:

Let us say that AI is artificial intelligence of human level or greater (that is, at least as intelligent as an average human). Let us say that AI+ is artificial intelligence of greater than human level (that is, more intelligent than the most intelligent human). Let us say that AI++ (or superintelligence) is AI of far greater than human level (say, at least as far beyond the most intelligent human as the most intelligent human is beyond a mouse).

With this in place, Chalmers formalized Good’s argument like so:

1.     There will be AI (before long, absent defeaters).

2.     If there is AI, there will be AI+ (soon after, absent defeaters).

3.     If there is AI+, there will be AI++ (soon after, absent defeaters).

4.     Therefore, there will be AI++ (before too long, absent defeaters).

I will defend Chalmers’ argument in greater detail than he has, using “before long” to mean “within 150 years,” using “soon after” to mean “within two decades,” and using “before too long” to mean “within two centuries.” My definitions here are similar to Chalmers’ definitions, but more precise.

Following Chalmers, by “defeaters” I mean “anything that prevents intelligent systems (human or artificial) from manifesting their capacities to create intelligent systems.” Defeaters include “disasters, disinclination, and active prevention.”

Disasters include catastrophic events that would severely impede scientific progress, such as supervolcano eruption, asteroid impact, cosmic rays, climate change, pandemic, nuclear war, biological warfare, an explosion of nanotechnology, and so on. The risk of such disasters and others are assessed in Bostrom & Cirkovic (2008).

Disinclination refers to a lack of interest in developing AI of human-level general intelligence. Given the enormous curiosity of the human species, and the power that human-level AI could bring its creators, I think long-term disinclination is unlikely.

Active prevention of the development of human-level artificial intelligence has already been advocated by Thomas Metzinger (2004), though not because of the risk to humans. Rather, Metzinger is concerned about the risk to artificial agents. Early AIs will inevitably be poorly designed, which could lead to enormous subjective suffering for them that we cannot predict. One might imagine an infant from near Cherynobl whose parts are so malformed by exposure to nuclear radiation during development that its short existence is a living hell. In working toward human-level artificial intelligence, might we be developing millions of internally malformed beings that suffer horrible subjective experiences but are unable to tell us so?

It is difficult to predict the likelihood of the active prevention of AI development, but the failure of humanity to halt the development of ever more powerful nuclear weapons (Norris & Kristensen 2009) – even after tasting their destructive power – does not inspire optimism.

Later, we will return to consider these potential defeaters again. For now, let us consider the premises of Chalmers’ argument.

***

 

New to LessWrong?

New Comment
16 comments, sorted by Click to highlight new comments since: Today at 11:27 AM

With this in place, Chalmers formalized Good’s argument like so:

I recommend substituting "like this" or "as follows". Apart from the fact that I personally dislike the phrase "like so" (and similar goofy-sounding expressions such as "yay high"), my suggested alternatives are a better fit with the style of written academic prose.

I'd be curious to hear what people think of Metzinger's argument about the morality of developing artificial intelligences. Has it been discussed on Less Wrong before?

Eliezer discussed something like that in Nonsentient Optimizers.

Excellent, thanks!

Disasters include catastrophic events that would severely impede scientific progress, such as supervolcano eruption, asteroid impact, cosmic rays, climate change, pandemic, nuclear war, biological warfare, an explosion of nanotechnology, and so on. The risk of such disasters and others are assessed in Bostrom & Cirkovic (2008).

A bit as a side note, but I don't think going into the details here adds much to your argument; Bostrom's discussion of global catastrophic risks is interesting but it's a distraction here.

Yup. That's why I pass the buck to Global Catastrophic Risks.

Where does 150 years and 200 years come from? Just long enough to sound conservative?

Since Turing humans have been overly optimistic about when we would have machine intelligence. There is nothing I see in the study of either machine intelligence or biological intellitence to suggest to me that we are anywhere but on a very steep part of a very complicated learning curve that we have been on for 50 years already. I don't know how I would quantify where the peak of that learning curve might be.

Even with that, I like your book!

It is difficult to predict the likelihood of the active prevention of AI development, but the failure of humanity to halt the development of ever more powerful nuclear weapons (Norris & Kristensen 2009) – even after tasting their destructive power – does not inspire optimism.

"Optimism" being code for "preventing machine intelligence from being invented"?

You may want to spell this out - or perhaps substitute the word "pessimism".

Agreed. Thanks.

Lol, why was this downvoted?

While in context of your text it can be inferred that "optimism" refers to prevention of AI development, Tim's suggestion that "pessimism" could be substituted instead is a disingenuous way of proselytizing his own position, unrelated to what he was commenting on since this option would be doubly confusing if actually implemented.

Huh. Well, anyway, I think it will help to clarify what I meant by optimism.

proselytizing his own position

Which is, for the uninitiated (link preferred) ? thx

Efforts to halt the development of increasingly-powerful machine intelligences seem likely to be made by a faction of humanity labelled by the rest of the humans as terrorists.

So far there has mostly only been the unabomber - but there may be more - and it may not be very much fun.

I repudiate the accusation of Machiavellian manipulation.

I suppose you could just raise your right hand.

I didn't downvote Tim, but I considered your usage standard English in referring to the first clause of the sentence, not a coded value judgement. I read it as "...Does not inspire optimism [that active prevention of AI is likely]."

Compare to, "It is difficult to predict the likelihood of John Dillinger's gang successfully robbing first national, but the addition of 8 armed guards To the bank's security plan does not inspire optimism."