One possible answer to the argument "attempting to build FAI based on Eliezer's ideas seems infeasible and increases the risk of UFAI without helping much to increase the probability of a good outcome, and therefore we should try to achieve a positive Singularity by other means" is that it's too early to decide this. Even if our best current estimate is that trying to build such an FAI increases risk, there is still a reasonable chance that this estimate will turn out to be wrong after further investigation. Therefore, the counter-argument goes, we ought to mount a serious investigation into the feasibility and safety of Eliezer's design (as well as other possible FAI approaches), before deciding to either move forward or give up.
(I've been given to understand that this is a standard belief within SI, except possibly for Eliezer, which makes me wonder why nobody gave this counter-argument in response to my post linked above. ETA: Carl Shulman did subsequently give me a version of this argument here.)
This answer makes sense to me, except for the concern that even seriously investigating the feasibility of FAI is risky, if the team doing so isn't fully rational. For example they may be overconfident about their abilities and thereby overestimate the feasibility and safety, or commit sunken cost fallacy once they have developed lots of FAI-relevant theory in the attempt to study feasibility, or become too attached to their status and identity as FAI researchers, or some team members may disagree with a consensus of "give up" and leave to form their own AGI teams and take the dangerous knowledge developed with them.
So the question comes down to, how rational is such an FAI feasibility team likely to be, and is that enough for the benefits to exceed the costs? I don't have a lot of good ideas about how to answer this, but the question seems really important to bring up. I'm hoping this post this will trigger SI people to tell us their thoughts, and maybe other LWers have ideas they can share.
Thanks, that's good to know.
I guess I would describe my overall view as being around 50/50 uncertain about whether the Singularity will be Yudkowsky-style (fast local FOOM) or Hanson-style (slower distributed FOOM). Conditioning on Yudkowsky-style Singularity, I agree with Eliezer that the default outcome is probably a paperclipper-style UFAI, and disagree with him on how hard the FAI problems are (I think they are harder). Conditioning on Hanson-style Singularity, I agree with Hanson's competitive evolution / burning of the cosmic commons scenario, but disagree with him in that I think that would be a terrible outcome rather than an ok outcome. Paul seems to be similarly uncertain about the speed and locality of the intelligence explosion, but apparently much more optimistic than me (or Eliezer and Robin) about the outcome of both scenarios. I'm not entirely sure why yet.
Disagree in which direction?
You got 11 comments here versus none at MIRI's blog. Seems like a no-brainer to me...
I'm not so much worried about Eliezer's own FAI attempts specifically (although I'm not as confident about Eliezer's rationality as you are), as worried about MIRI making a lot of people interested in various problems that it thinks are safe for people to work on in public. But since such safety is hard to predict in advance and FAI problems are intimately connected with AGI problems, and those people will publish interesting technical papers and draw even more people in who might be interested purely in the technical problems, you'll have a lot of people working on FAI/AGI problems, that wouldn't exist if MIRI hadn't gotten the snowball started.
Suppose at some point MIRI decides that P(eAI | FAI attempted) is too high and FAI shouldn't be attempted, or that people are doing a lot of unwanted AGI capability work as a result of your current activities, how do you get everyone to stop?
I find it relatively easy to write down my strategic thoughts in blog form, and they tend to be well-received on LW, so they seem to be a counterexample to your "difficulty of purchasing new strategic insights". Unless all my ideas/arguments were already thought of before and taken into account by "insiders", but just never written down somewhere public? But if that's the case, why haven't they been written down? Again, it wasn't that hard for me to write down my thoughts in a way that's understandable to more than just "insiders".
My guess was that you think the things on this list are more probable and/or bad than I do, especially the first three. But to keep things focused, I suggest we not add those possible disagreements to the current thread.
In part, I'm also trying to see whether we can cause discussion to happen on MIRI's blog. If not, we'll have to do it on LW. But that'll be better when we've had enough development resources go into LW that we can have subre... (read more)