polymathwannabe comments on Open Thread May 2 - May 8, 2016 - Less Wrong

5 Post author: Elo 02 May 2016 02:43AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (149)

You are viewing a single comment's thread. Show more comments above.

Comment author: Daniel_Burfoot 04 May 2016 02:02:56PM 8 points [-]

Though I enthusiastically endorse the concept of rationality, I often find myself coming to conclusions about Big Picture issues that are quite foreign to the standard LW conclusions. For example, I am not signed up for cryonics even though I accept the theoretical arguments in favor of it, and I am not worried about unfriendly AI even though I accept most of EY's arguments.

I think the main reason is that I am 10x more pessimistic about the health of human civilization than most other rationalists. I'm not a cryonicist because I don't think companies like Alcor can survive the long period of stagnation that humanity is headed towards. I don't worry about UFAI because I don't think our civilization has the capability to achieve AI. It's not that I think AI is spectacularly hard, I just don't think we can do Hard Things anymore.

Now, I don't know whether my pessimism is more rational than others' optimism. LessWrong, and rationalists in general, probably have a blind spot relative to questions of civilizational inadequacy because those questions relate to political issues, and we don't talk about politics. Is there a way we can discuss civilizational issues without becoming mind-killed? Or do we simply have to accept that civilizational issues are going to create a large error bar of uncertainty around our predictions?

Comment author: polymathwannabe 04 May 2016 02:47:52PM 3 points [-]

Is there a way we can discuss civilizational issues without becoming mind-killed?

A LWer created Omnilibrium for that.

Comment author: Viliam 05 May 2016 08:41:19AM 4 points [-]

Any results? (I am personally unimpressed by the few random links I have seen.)