MIRI AMA plus updates
MIRI is running an AMA on the Effective Altruism Forum tomorrow (Wednesday, Oct. 11): Ask MIRI Anything. Questions are welcome in the interim!
Nate also recently posted a more detailed version of our 2016 fundraising pitch to the EA Forum. One of the additions is about our first funding target:
We feel reasonably good about our chance of hitting target 1, but it isn't a sure thing; we'll probably need to see support from new donors in order to hit our target, to offset the fact that a few of our regular donors are giving less than usual this year.
The Why MIRI's Approach? section also touches on new topics that we haven't talked about in much detail in the past, but plan to write up some blog posts about in the future. In particular:
Loosely speaking, we can imagine the space of all smarter-than-human AI systems as an extremely wide and heterogeneous space, in which "alignable AI designs" is a small and narrow target (and "aligned AI designs" smaller and narrower still). I think that the most important thing a marginal alignment researcher can do today is help ensure that the first generally intelligent systems humans design are in the “alignable” region. I think that this is unlikely to happen unless researchers have a fairly principled understanding of how the systems they're developing reason, and how that reasoning connects to the intended objectives.
Most of our work is therefore aimed at seeding the field with ideas that may inspire more AI research in the vicinity of (what we expect to be) alignable AI designs. When the first general reasoning machines are developed, we want the developers to be sampling from a space of designs and techniques that are more understandable and reliable than what’s possible in AI today.
In other news, we've uploaded a new intro talk on our most recent result, "Logical Induction," that goes into more of the technical details than our previous talk.
See also Shtetl-Optimized and n-Category Café for recent discussions of the paper.
[Recommendation] Steven Universe & cryonics
I've been watching Steven Universe with my fiancee (a children's cartoon on Cartoon Network by Rebecca Sugar), and it wasn't until I got to Season 3 that I realized there's been a cryonics metaphor running in the background since the very first episode. If you want to introduce your kids to the idea of cryonics, this series seems like a spectacularly good way to do it.
If you don't want any spoilers, just go watch it, then come back.
Otherwise, here's the metaphor I'm seeing, and why it's great:
-
In the very first episode, we find out that the main characters are a group called the Crystal Gems, who fight 'gem monsters'. When they defeat a monster, a gem is left behind, which they lock in a bubble-forcefield and store in their headquarters.
-
One of the Crystal Gems is injured in a training accident, and we find out that their bodies are just projections; each Crystal Gem has a gem located somewhere on their body, which contains their minds. So long as their gem isn't damaged, they can project a new body after some time to recover. So we already have the insight that minds and bodies are separate.
-
This is driven home by a second episode where one of the Crystal Gems has their crystal cracked; this is actually dangerous to their mind, not just body, and is treated as a dire emergency instead of merely an inconvenience.
-
Then we eventually find out that the gem monsters are actually corrupted members of the same species as the Crystal Gems. They are 'bubbled' and stored in the temple in hopes of eventually restoring them to sanity and their previous forms.
-
An attempt is made to cure one of the monsters, which doesn't fully succeed, but at least restores them to sanity. This allows them to remain unbubbled and to be reunited with their old comrades (who are also corrupted). This was the episode where I finally made the connection to cryonics.
-
The Crystal Gems are also revealed to be over 5000 years old, and effectively immortal. They don't make a big deal out of this; for them, this is totally normal.
-
This also implies that they've made no progress in curing the gem monsters in 5000 years, but that doesn't stop them from preserving them anyway.
-
Finally, a secret weapon is revealed which is capable of directly shattering gems (thus killing the target permanently), but the use of it is rejected as unethical.
So, all in all, you have a series where when someone is hurt or sick in a way that you can't help, you preserve their mind in a safe way until you can figure out a way to help them. Even your worst enemy deserves no less.
Also, Steven Universe has an entire episode devoted to mindfulness meditation.
The University of Cambridge Centre for the Study of Existential Risk (CSER) is hiring!
The University of Cambridge Centre for the Study of Existential Risk (CSER) is recruiting for an Academic Project Manager. This is an opportunity to play a shaping role as CSER builds on its first year's momentum towards becoming a permanent world-class research centre. We seek an ambitious candidate with initiative and a broad intellectual range for a postdoctoral role combining academic and project management responsibilities.
The Academic Project Manager will work with CSER's Executive Director and research team to co-ordinate and develop CSER's projects and overall profile, and to develop new research directions. The post-holder will also build and maintain collaborations with academic centres, industry leaders and policy makers in the UK and worldwide, and will act as an ambassador for the Centre’s research externally. Research topics will include AI safety, bio risk, extreme environmental risk, future technological advances, and cross-cutting work on governance, philosophy and foresight. Candidates will have a PhD in a relevant subject, or have equivalent experience in a relevant setting (e.g. policy, industry, think tank, NGO).
Application deadline: November 11th. http://www.jobs.cam.ac.uk/job/11684/
Astrobiology III: Why Earth?
After many tribulations, my astrobiology bloggery is back up and running using Wordpress rather than Blogger because Blogger is completely unusable these days. I've taken the opportunity of the move to make better graphs for my old posts.
"The Solar System: Why Earth?"
https://thegreatatuin.wordpress.com/2016/10/03/the-solar-system-why-earth/
Here, I try to look at our own solar system and what the presence of only ONE known biosphere, here on Earth, tells us about life and perhaps more importantly what it does not. In particular, I explore what aspects of Earth make it special and I make the distinction between a big biosphere here on Earth that has utterly rebuilt the geochemistry and a smaller biosphere living off smaller amounts of energy that we probably would never notice elsewhere in our own solar system given the evidence at hand.
Commentary appreciated.
Previous works:
Space and Time, Part I
https://thegreatatuin.wordpress.com/2016/09/25/space-and-time-part-i
Space and Time, Part II
https://thegreatatuin.wordpress.com/2016/09/25/space-and-time-part-ii
Fermi paradox of human past, and corresponding x-risks
Based on known archaeological data, we are the first technological and symbol-using civilisation on Earth (but not the first tool-using species).
This leads to an analogy that fits Fermi’s paradox: Why are we the first civilisation on Earth? For example, flight was invented by evolution independently several times.
We could imagine that on our planet, many civilisations appeared and also became extinct, and based on mediocre principles, we should be somewhere in the middle. For example, if 10 civilisations appeared, we have only a 10 per cent chance of being the first one.
The fact that we are the first such civilisation has strong predictive power about our expected future: it lowers the probability that there will be any other civilisations on Earth, including non-humans or even a restarting of human civilisation from scratch. It is because, if there will be many civiizations, we should not find ourselves to be the first one (It is some form of Doomsday argument, the same logic is used in Bostrom's article “Adam and Eve”).
If we are the only civilisation to exist in the history of the Earth, then we will probably become extinct not in mild way, but rather in a way which will prevent any other civilisation from appearing. There is higher probability of future (man-made) catastrophes which will not only end human civilisation, but also prevent any existence of any other civilisations on Earth.
Such catastrophes would kill most multicellular life. Nuclear war or pandemic is not that type of a catastrophe. The catastrophe must be really huge: such as irreversible global warming, grey goo or black hole in a collider.
Now, I will list possible explanations of the Fermi paradox of human past and corresponding x-risks implications:
1. We are the first civilisation on Earth, because we will prevent the existence of any future civilisations.
If our existence prevents other civilisations from appearing in the future, how could we do it? We will either become extinct in a very catastrophic way, killing all earthly life, or become a super-civilisation, which will prevent other species from becoming sapient. So, if we are really the first, then it means that "mild extinctions" are not typical for human style civilisations. Thus, pandemics, nuclear wars, devolutions and everything reversible are ruled out as main possible methods of human extinction.
If we become a super-civilisation, we will not be interested in preserving biosphera, as it will be able to create new sapient species. Or, it may be that we care about biosphere so strongly, that we will hide very well from new appearing sapient species. It will be like a cosmic zoo. It means that past civilisations on Earth may have existed, but decided to hide all traces of their existence from us, as it would help us to develop independently. So, the fact that we are the first raises the probability of a very large scale catastrophe in the future, like UFAI, or dangerous physical experiments, and reduces chances of mild x-risks such as pandemics or nuclear war. Another explanation is that any first civilisation exhausts all resources which are needed for a technological civilisation restart, such as oil, ores etc. But, in several million years most such resources will be filled again or replaced by new by tectonic movement.
2. We are not the first civilisation.
2.1. We didn't find any traces of a previous technological civilisation, yet based on what we know, there are very strong limitations for their existence. For example, every civilisation makes genetic marks, because it moves animals from one continent to another, just as humans brought dingos to Australia. It also must exhaust several important ores, create artefacts, and create new isotopes. We could be sure that we are the first tech civilisation on Earth in last 10 million years.
But, could we be sure for the past 100 million years? Maybe it was a very long time ago, like 60 million years ago (and killed dinosaurs). Carl Sagan argued that it could not have happened, because we should find traces mostly as exhausted oil reserves. The main counter argument here is that cephalisation, that is the evolutionary development of the brains, was not advanced enough 60 millions ago, to support general intelligence. Dinosaurian brains were very small. But, bird’s brains are more mass effective than mammalians. All these arguments in detail are presented in this excellent article by Brian Trent “Was there ever a dinosaurian civilisation”?
The main x-risks here are that we will find dangerous artefacts from previous civilisation, such as weapons, nanobots, viruses, or AIs. And, if previous civilisations went extinct, it increases the chances that it is typical for civilisations to become extinct. It also means that there was some reason why an extinction occurred, and this killing force may be still active, and we could excavate it. If they existed recently, they were probably hominids, and if they were killed by a virus, it may also affect humans.
2.2. We killed them. Maya civilisation created writing independently, but Spaniards destroy their civilisation. The same is true for Neanderthals and Homo Florentines.
2.3. Myths about gods may be signs of such previous civilisation. Highly improbable.
2.4. They are still here, but they try not to intervene in human history. So, it is similar to Fermi’s Zoo solution.
2.5. They were a non-tech civilisation, and that is why we can’t find their remnants.
2.6 They may be still here, like dolphins and ants, but their intelligence is non-human and they don’t create tech.
2.7 Some groups of humans created advanced tech long before now, but prefer to hide it. Highly improbable as most tech requires large manufacturing and market.
2.8 Previous humanoid civilisation was killed by virus or prion, and our archaeological research could bring it back to life. One hypothesis of Neanderthal extinction is prionic infection because of cannibalism. The fact is - several hominid species went extinct in the last several million years.
3. Civilisations are rare
Millions of species existed on Earth, but only one was able to create technology. So, it is a rare event.Consequences: cyclic civilisations on earth are improbable. So the chances that we will be resurrected by another civilisation on Earth is small.
The chances that we will be able to reconstruct civilisation after a large scale catastrophe, are also small (as such catastrophes are atypical for civilisations and they quickly proceed to total annihilation or singularity).
It also means that technological intelligence is a difficult step in the evolutionary process, so it could be one of the solutions of the main Fermi paradox.
Safety of remains of previous civilisations (if any exist) depends on two things: the time distance from them and their level of intelligence. The greater the distance, the safer they are (as the biggest part of dangerous technology will be destructed by time or will not be dangerous to humans, like species specific viruses).
The risks also depend on the level of intelligence they reached: the higher intelligence the riskier. If anything like their remnants are ever found, strong caution is recommend.
For example, the most dangerous scenario for us will be one similar to the beginning of the book of V. Vinge “A Fire upon the deep.” We could find remnants of a very old, but very sophisticated civilisation, which will include unfriendly AI or its description, or hostile nanobots.
The most likely place for such artefacts to be preserved is on the Moon, in some cavities near the pole. It is the most stable and radiation shielded place near Earth.
I think that based on (no) evidence, estimation of the probability of past tech civilisation should be less than 1 per cent. While it is enough to think that they most likely don’t exist, it is not enough to completely ignore risk of their artefacts, which anyway is less than 0.1 per cent.
Meta: the main idea for this post came to me in a night dream, several years ago.
Is Politics the Mindkiller? An Inconclusive Test
Or is the convention against discussing politics here silly?
I propose a test. I'm going to try to lay down some rules on voting on comments for the test here (not that I can force anybody to abide by them):
1.) Top-level comments should introduce arguments (or ridicule me and/or this test); responses should be responses to those arguments.
2.) Upvote and downvote based on whether or not you find an argument convincing in the context in which it was raised. This means if it's a good argument against the argument it is responding to, not whether or not there's a good/obvious counterargument to it; if you have a good counterargument, raise it. If it's a convincing argument, and the counterargument is also convincing, upvote both. If both arguments are unconvincing, downvote both.
3.) Try not to downvote particular comments excessively, if they're legitimate lines of argument. A faulty line of argument provides opportunity for rebuttal, and so for our test has value even then; that is, I want some faulty lines of argument here. If you disagree, please downvote me, instead of the faulty comments, because this post is what you want less of, not those comments. This necessarily implies, for balance, that we not excessively upvote comments. I'd suggest fairly arbitrary limits of 3/-3?
Edit: 4.) A single argument per comment would be ideal; as MixedNuts points out here, it's otherwise hard to distinguish between one good and one bad argument, which makes the upvoting/downvoting difficult to evaluate. (My apologies about missing this, folks.)
I'm going to try really hard not to get personally involved, except to lay down a leading comment posing an argument against abortion, a position I don't hold, for the record. The core of the argument isn't disingenuous, and I hold that this argument is true, it just doesn't lead to my opposing abortion. I do not hold the moral axiom by which I extend the basic argument to argue against abortion, however; I'm playing the devil's advocate to try to help me from getting sucked into the argument while providing an initial point of discussion.
Which leads me to the next point: If you see a hole in an argument, even if it's an argument for a perspective you agree with, poke through it. The goal is to see whether we can have a constructive political argument here.
The fact that this is a test, and known to be a test, means this isn't a blind study. Uh, try to act as if you're not being tested?
After it's gone on a little while, if this post hasn't been hopelessly downvoted and ridiculed (and thus the premise and test discarded as undesirable to begin with), we can put up a poll to see whether people found the political debates helpful, not helpful, and so on.
Open thread, Sep. 26 - Oct. 02, 2016
If it's worth saying, but not worth its own post, then it goes here.
Notes for future OT posters:
1. Please add the 'open_thread' tag.
2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)
3. Open Threads should start on Monday, and end on Sunday.
4. Unflag the two options "Notify me of new top level comments on this article" and "
Purchase Fuzzies and Utilons Separately
Previously in series: Money: The Unit of Caring
Yesterday:
There is this very, very old puzzle/observation in economics about the lawyer who spends an hour volunteering at the soup kitchen, instead of working an extra hour and donating the money to hire someone...
If the lawyer needs to work an hour at the soup kitchen to keep himself motivated and remind himself why he's doing what he's doing, that's fine. But he should also be donating some of the hours he worked at the office, because that is the power of professional specialization and it is how grownups really get things done. One might consider the check as buying the right to volunteer at the soup kitchen, or validating the time spent at the soup kitchen.
I hold open doors for little old ladies. I can't actually remember the last time this happened literally (though I'm sure it has, sometime in the last year or so). But within the last month, say, I was out on a walk and discovered a station wagon parked in a driveway with its trunk completely open, giving full access to the car's interior. I looked in to see if there were packages being taken out, but this was not so. I looked around to see if anyone was doing anything with the car. And finally I went up to the house and knocked, then rang the bell. And yes, the trunk had been accidentally left open.
Under other circumstances, this would be a simple act of altruism, which might signify true concern for another's welfare, or fear of guilt for inaction, or a desire to signal trustworthiness to oneself or others, or finding altruism pleasurable. I think that these are all perfectly legitimate motives, by the way; I might give bonus points for the first, but I wouldn't deduct any penalty points for the others. Just so long as people get helped.
But in my own case, since I already work in the nonprofit sector, the further question arises as to whether I could have better employed the same sixty seconds in a more specialized way, to bring greater benefit to others. That is: can I really defend this as the best use of my time, given the other things I claim to believe?
Altruistic Kidney Donation
People are dying because their kidneys have failed, and you probably have an extra one: should you donate it? A dialysis machine can do most of the work of a kidney, but it's not as good. A donated kidney gives someone an average of 10 additional years. [1] (Some of them reject it right away, others live for years before having problems.) This is pretty good: if you volunteer to be a kidney donor and are selected as a match, you give someone an average of another 10 years of life. [2]
A kidney donation is a surgery, however, and does have some risk for you. It's hard to calculate the risk because people who are selected for kidney donation are a weird group: they're heavily screened (so they tend to be healthier) but they're likely to be related to the person they're donating to (so they tend to be at risk for kidney problems themself). An ideal study on the dangers of kidney donation would compare donors with people who were going to donate and then didn't for some nearly-random reason. I can't find such a study, but I've heard people give odds of around 1:3000 for dying due to donating. You also spend several days in the hospital, miss some weeks of work, and take several months to fully recover. [3] This sounds like a lot, but giving someone 10 years of life is a lot too.
The question isn't so much "does this do good," however, but "does this maximize the good I can do for the amount of sacrifice I'm willing to accept?" GiveWell estimates that each $1600 given to the AMF for antimalarial net distribution prevents the death of someone with an average life expectancy of maybe 40 years [4]. So another way to give a stranger 10 more years of life is to give $400 to the AMF. Kidney donation is in the same ballpark as this, in terms of the ratio of benefit (to others) to cost (to you) [5], but even ignoring the risk of death [6] I would pay that $400 just to avoid the hospital and recovery time.
Update 2012-06-22: BDan and Alexander Berger have convinced me I was wrong to write off chains. At least right now there are more possible chains than kidneys to start them, and they can be quite long. My updated estimate is that this gets you a 3x multiplier on the value of a donated kidney, bringing the $400 charity equivalent to $1200. Which is still less than the $10-$100K at which I think I would sell a kidney, but substantially higher.
(I also posted this on my blog)
[1] Comparison of mortality in all patients on dialysis, patients on dialysis awaiting transplantation, and recipients of a first cadaveric transplant. (1999), table 3. This is from a longitudinal study on 23,275 people who got kidney transplants from cadavers and had never had one before. In our case we're interested in live donation, but I would guess it's about the same.
[2] It's possible that you could do better than this if your donation to a stranger set off a chain of donations. The idea is that there are lots of people who would give their kidney to a friend of family member if they could, but are for whatever reason incompatible. You can make donor-donee-donor-donee chains where it just takes one altruistic person to set the whole thing off. (Cycles are also possible, but they don't need someone to get them started.) The question is, though, how likely are these chains to happen anyway? And how long a chain can you reasonably expect? My understanding is that these are quite rare, with most altruistic kidney donors not starting one. I'm going to ignore chains.
[3] Someone who gave their kidney wrote in a Guardian article: "I was in hospital for four days--the worst part of the whole ordeal was removing the catheter. In just two weeks I was back at work (the full recovery is supposed to be three months)."
[4] I would love to find a GiveWell estimate of what life expectancy for the people who don't die of malaria because of a net, but I can't find one. So this "40 years" is a guess after looking at life expectancies for different ages in the countries where the AMF works.
[5] There are huge costs to both dialysis and kidney transplant, so as a health intervention distributing antimalarial nets [makes much more sense]. From the perspective of a prospective donor, however, I don't think this matters.
[6] You might think that if you're someone who gives a substantial fraction of their earnings to effective charity, a 1:3000 risk of death would be pretty bad because you'd be able to help other people much less. But 3000 is really big: if you're giving $20K/year and have an expected additional 30 years of working life, a 1:3000 chance of not being able to do that has an undiscounted expected value of only $200 less donated. [7]
[7] $20K * 30 years * 1:3000 = $200
View more: Next
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)