You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Dr_Manhattan comments on Andrew Ng dismisses UFAI concerns - Less Wrong Discussion

3 Post author: OphilaDros 06 March 2015 05:26AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (22)

You are viewing a single comment's thread. Show more comments above.

Comment author: Dr_Manhattan 07 March 2015 03:19:22PM 1 point [-]

My impression was that imminence is a point of contention, much less orthogonality. Who specifically do you have in mind?

Comment author: Vaniver 07 March 2015 03:38:05PM 3 points [-]

My impression was that imminence is a point of contention, much less orthogonality.

This article is a good place to start in clarifying the MIRI position. Since their estimate for imminence seems to boil down to "we asked the community what they thought and made a distribution," I don't see that as contention.

There is broad uncertainty about timelines, but the MIRI position is "uncertainty means we should not be confident we have all the time we need," not "we're confident it will happen soon," which is where someone would need to be for me to say they're "for imminence."

Comment author: Dr_Manhattan 07 March 2015 04:07:41PM 0 points [-]

Interesting. I considered imminence more of a point of contention b/c the most outspoken "AI risk is overhyped" people are mostly using it as an argument (and I consider this bunch way more serious than Searle and Brooks: Yann LeCun, Yoshua Bengio, Andrew Ng).