At the FHI, we are currently working on a project around whole brain emulations (WBE), or uploads. One important question is if getting to whole brain emulations first would make subsequent AGI creation
- more or less likely to happen,
- more or less likely to be survivable.
If you have any opinions or ideas on this, please submit them here. No need to present an organised overall argument; we'll be doing that. What would help most is any unusual suggestion, that we might not have thought of, for how WBE would affect AGI.
EDIT: Many thanks to everyone who suggested ideas here, they've been taken under consideration.
From a utilitarian point of view, once you can create a VR, why would you want to create an FAI? On the other hand, if you create an FAI, wouldn't you still want to create WBE and VR? If you had WBE and VR, would you want to create more existential risks by making an AGI? The motive would have to be non-utilitarian. "Because they can" seems about right.
Also, about the bioconservatives objecting and libertarians becoming more interested. A Seastead would be the perfect retreat for WBE researchers.