So far, of course, the sampling here is very locally biased.
Think about nuclear technology. It evolved in a time of war... The probability that nuclear technology was going to arise at a time when we use it well rather than [for] destruction was low...
So far the nuclear record has turned out reasonably well though, I'm not sure what the argument is here.
Unfortunately, Hanson does not explain his reasons for rejecting Miller's analysis.
One reason to discount arms race worries is increasing global peace and cooperation, as covered at length by Steven Pinker.
and that the efforts of MIRI are critical to the planet's survival.
I would certainly disclaim this as a likely possibility, speaking for myself.
Former GiveWell researcher Jonah Sinick has expressed optimism on the issue:
- In the Manhattan project, the "will bombs ignite the atmosphere?" question was analyzed and dismissed without much (to our knowledge) double-checking.
Argh. This is a myth. (It is especially frustrating in its most dire form, when it's implied that the concern was chemical ignition, i.e. ordinary fire. The concern was actually about nuclear ignition. Other retellings often say "set fire to" which implies ordinary fire.)
See Wikipedia's article on the Trinity test, which links to Report LA-602, "Ignition of the Atmosphere With Nuclear Bombs", where this was carefully analyzed.
I had totally forgotten about that post. (Probably because I learned about the history of the Manhattan Project long before reading anything from EY.) Thanks for the reminder.
There were calculations done before the bomb was tested, which confirmed people's strong priors against an atmospheric ignition effect. But the report is dated August 1946, after the first nuclear tests and the destruction of Hiroshima and Nagasaki. Presumably the report is elaborating on the earlier calculations, but the analysis before the first nuclear detonations is more important than the analysis after.
I suspect, without any evidence, that the analysis was carried out to a sufficient extent to convince all of the physicists involved, saving the formal writeup for later. There was a war on, you know.
Nowadays, as I understand it, most areas of science are carried out through informal circulation of preprints long before papers are formally published for the record. I imagine the same thing went on at Los Alamos, especially given the centralization of that community.
I suspect, without any evidence, that the analysis was carried out to a sufficient extent to convince all of the physicists involved, saving the formal writeup for later.
Based on the evidence of reading about the Teller story, yes the calculations were enough to settle the issue internally. Indeed, others thought it a bit silly in the first place. The procedure clearly had a very low expected failure rate in general, and the danger was low prior. OTOH, it's not clear what p "convincing" translated into.
In any case, those calculations eliminated most of the subjective expected value of atmospheric nuclear ignition risk, as asteroid searches have eliminated most of the expected value of asteroid extinction risk.
(It is especially frustrating in its most dire form, when it's implied that the concern was chemical ignition, i.e. ordinary fire. The concern was actually about nuclear ignition. Other retellings often say "set fire to" which implies ordinary fire.)
I'm peeved by the fact that stellar nucleosynthesis processes are usually called “burning” rather than “fusion”, BTW.
Stars are awesome (in the old-school non-diluted sense) which naturally makes it tempting to use more evocative language when talking about them. And you could think of 'burning' in such usage referring more to incandescence rather than rapid oxidation.
I was going to object to the idea that “fusion” isn't evocative enough, but I guess that whoever first named stellar nucleosynthesis “burning” hadn't been exposed to Dragon Ball Z, Gillette advertising and repeated claims that “thirty years from now” fusion power will solve all of our problems.
(And, of course, the real question here is whether Jews are allowed to operate fusion reactors on Shabbat. ;-))
Does Eliezer assign lots of probability mass to a particular failure mode or does he have his probability mass fairly evenly spread across many failure modes? His answer seems a bit overconfident to me for a question that involves the actions of squishy and unpredictable humans.
Sergey Brin, an apparently smart person who has met politicians (unlike anyone quoted here?), says the ones he has met are "invariably thoughtful, well-meaning people" whose main problem is the fact that "90% of their effort seems to be focused on how to stick it to the other party". So it could matter a lot how the issue ends up getting framed. What are the issues that the government seems to deal with most intelligently, and how can we make FAI end up getting treated like those issues?
Nate Silver's book discusses the work of government weather forecasters, earthquake researchers, and disease researchers and seems to give them positive reviews.
Some publicly funded universities do important and useful research.
My dad told me a story about a group of quants hired by the city of New York to develop a model of what buildings needed visits from the city fire inspector, with impressive results. Here's an article I found while trying to track down that anecdote. (Oh wait, maybe this is it?)
I like the Bureau of Labor Statistics' Occupational Outlook Handbook, but it's hard to know how accurate it is.
Does anyone have more examples? Note that I've reframed the problem from "elites and AI" to "government and AI"; not sure if that's a good reframing.
One guess: make the part of the government concerned with AI be a boring government department staffed with PhDs in a technical subject where one of the best careers one can get given a PhD in this subject is to work for this government department (math?). IMO, the intelligence of government workers probably matters more than the fact that they are government workers, and there are factors that determine this. This strategy (a variant of "can't beat 'em? join 'em") would probably end up working better in a scenario where there is no "AI Sputnik moment". (BTW, can we expect politicians to weigh the opinions of government experts over those of experts who have private-sector or nonprofit jobs?)
Here's an interesting government department. I wonder how hard it would be for a few of us Less Wrong users to get jobs there?
Scott Aaronson seems to be notoriously optimistic and has reiterated his opinion fairly recently.
Where does he express optimism about elites' handling of AGI? In that post, he seems to just be saying "AGI is probably many centuries away, and I don't see much we can knowably do about it so far in advance."
Right, and if it's that slow, there is plenty of time for it to be noticed and mitigated, see Adams Law of Slow-Moving Disasters
In past conversations he has been rather pessimistic about climate change collapsing civilization via nuclear war.
There seems to be the implicit assumption that superhuman AI will be some sort of sudden absolute thing.
why?
If I were to guess I'd say that the most likely course is one of gradual improvement for quite some time, more similar to the development of airplanes than the development of the atomic bomb.
if you handed modern bombers and the tech to support them to one of the sides in the first world war then you can be sure they've have won pretty quickly. And there was investment in flight and it was useful. but early planes were slow, fault prone, terrible as weapons platforms etc.
We might very well see AI develop slowly with roadblocks every few years or decades which halt or slow advancement for a while until some sollution is found.
I guess it's down to whether you assume that the difficulty of increasing intelligence is exponential or linear.
If each additional IQ point(for want of a better measure) gets harder to add than the last then even with a cycle of self improvement you're not automatically going to get a god.
We might even see intelligence augmentation keeping pace with AI development for quite some time.
"What kind of competitive or political system would make fragmented squabbling AIs safer than an attempt to get the monolithic approach right?" by saying "the answer is, 'None.' It's like asking how you should move your legs to walk faster than a jet plane" — again, implying extreme skepticism that political elites will manage AI properly.
It seems like a bit of a non sequitur to go from competing AIs being unsafe to elites not managing AI properly. Was there meant to be an additional clause explaining that elites would tend to favor multiple competing AIs over a single monolithic one? (Perhaps this was an artifact of the deletion of the quote referenced in the footnote?)
If you've read Guy Mcpherson's best articles you will know AI/robots/machinery is pure fantasy at this point because of climate change alone. http://guymcpherson.com/2013/01/climate-change-summary-and-update/
Previously, I asked "Will the world's elites navigate the creation of AI just fine?" My current answer is "probably not," but I think it's a question worth additional investigation.
As a preliminary step, and with the help of MIRI interns Jeremy Miller and Oriane Gaillard, I've collected a few stated opinions on the issue. This survey of stated opinions is not representative of any particular group, and is not meant to provide strong evidence about what is true on the matter. It's merely a collection of quotes we happened to find on the subject. Hopefully others can point us to other stated opinions — or state their own opinions.
MIRI researcher Eliezer Yudkowsky is famously pessimistic on this issue. For example, in a 2009 comment, he replied to the question "What kind of competitive or political system would make fragmented squabbling AIs safer than an attempt to get the monolithic approach right?" by saying "the answer is, 'None.' It's like asking how you should move your legs to walk faster than a jet plane" — again, implying extreme skepticism that political elites will manage AI properly.1
Cryptographer Wei Dai is also quite pessimistic:
Stanford philosopher Ken Taylor has also expressed pessimism, in an episode of Philosophy Talk called "Turbo-charging the mind":
Here, Taylor seems to express the view that humans are not yet morally and rationally advanced enough to be trusted with powerful technologies. This general view has been expressed before by many others, including Albert Einstein, who wrote that "Our entire much-praised technological progress... could be compared to an axe in the hand of a pathological criminal."
In response to Taylor's comment, MIRI researcher Anna Salamon (now Executive Director of CFAR) expressed a more optimistic view:
Economist James Miller is another voice of pessimism. In Singularity Rising, chapter 5, he worries about game-theoretic mechanisms incentivizing speed of development over safety of development:
In chapter 6, Miller expresses similar worries about corporate incentives and AI:
Miller expanded on some of these points in his chapter in Singularity Hypotheses.
In a short reply to Miller, GMU economist Robin Hanson wrote that
Unfortunately, Hanson does not explain his reasons for rejecting Miller's analysis.
Sun Microsystems co-founder Bill Joy is famous for the techno-pessimism of his Wired essay "Why the Future Doesn't Need Us," but that article's predictions about elites' likely handling of AI are actually somewhat mixed:
Former GiveWell researcher Jonah Sinick has expressed optimism on the issue:
Paul Christiano is another voice of optimism about elites' handling of AI. Here are some snippets from his "mainline" scenario for AI development:
Christiano is no Polyanna, however. In the same document, he outlines "what could go wrong," and what we might do about it.
Notes
1 I originally included another quote from Eliezer, but then I noticed that other readers on Less Wrong had elsewhere interpreted that same quote differently than I had, so I removed it from this post.