You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

danieldewey comments on [Link] Will Superintelligent Machines Destroy Humanity? - Less Wrong Discussion

1 Post author: roystgnr 27 November 2014 09:48PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (5)

You are viewing a single comment's thread. Show more comments above.

Comment author: danieldewey 29 November 2014 12:36:20PM 4 points [-]

That's what I thought at first, too, but then I looked at the paper, and their figure looks right to me. Could you check my reasoning here?

On p.11 of Vincent's and Nick's survey, there's a graph "Proportion of experts with 10%/50%/90% confidence of HLMI by that date". At around the the 1 in 10 mark of proportion of experts -- the horizontal line from 0.1 -- the graph shows that 1 in 10 experts thought there was a 50% chance of HLAI by 2020 or so (the square-boxes-line), and 1 in 10 thought there was a 90% chance of HLAI by 2030 or so (the triangles-line). So, maybe 1 in 10 researchers think there's a 70% chance of HLAI by 2025 or so, which is roughly in line with the journalist's remark.

Did I do that right? Do you think the graph is maybe incorrect? I haven't checked the number against other parts of the paper.

There's a good chance that the reviewer got the right number by accident, I think, but it doesn't seem far enough away to call out.

Comment author: ciphergoth 29 November 2014 12:50:21PM 1 point [-]

Gosh, you might be right! I'm surprised by that. Would be good to have the data to get into it in more detail.