Wei_Dai comments on Modest Superintelligences - Less Wrong

20 Post author: Wei_Dai 22 March 2012 12:29AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (88)

You are viewing a single comment's thread. Show more comments above.

Comment author: Wei_Dai 23 March 2012 05:58:57AM 6 points [-]

why not simply push to acquire financial resources and hire brilliant people to do the work you think is necessary

The point is to obtain an insurmountable lead on WBE tech, otherwise you'll just spur competition and probably end up with Robin Hanson's Malthusian scenario. (If intelligence explosion were possible, you could win the WBE race by a small margin and translate that into a big win, but for this post I'm assuming that intelligence explosion isn't possible, so you need to win the race by a large margin.)

[comment from the heart, rather than from the head: your description of MSI-1 sounds kind of, well, totalitarian. Don't you think that's a little peculiar?]

In that case you're in for a surprise when you find out what I was referring to by "WBE-enabled institutional controls" for MSI-2. Read Carl Shulman's Whole Brain Emulation and the Evolution of Superorganisms.

Comment author: Vladimir_Nesov 01 April 2012 10:48:02PM 0 points [-]

(If intelligence explosion were possible, you could win the WBE race by a small margin and translate that into a big win, but for this post I'm assuming that intelligence explosion isn't possible, so you need to win the race by a large margin.)

Since exploiting intelligence explosion still requires FAI, and FAI could be very difficult, you might still need a large enough margin to perform all the necessary FAI research before your competition stumbles on an AGI.