bentarm comments on Open thread, October 2011 - Less Wrong

5 Post author: MarkusRamikin 02 October 2011 09:05AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (308)

You are viewing a single comment's thread. Show more comments above.

Comment author: Thomas 02 October 2011 12:46:05PM -1 points [-]

I estimate, that a currently working and growing superintelligence has a probability in a range of 1/million to 1/1000. I am at least 50% confident that it is so.

Not a big probability but given the immense importance of such an object, it is already a significant event to consider. The very near term birth of a superintelligence is something to think about. It wouldn't be just another Sputnik launched by some other people you thought they are unable to make it, but they sure were. We know that well, it wouldn't be just a minor blow for a pride as Sputnik was for some, and a triumph for others who conceived it and launched it.

No, that could be a check mate in a first move.

Non the less, people are dismissive of any short term success in the field. I am not and I want to express it in an open thread.

Comment author: wedrifid 02 October 2011 01:15:03PM 5 points [-]

I estimate, that a currently working and growing superintelligence has a probability in a range of 1/million to 1/1000. I am at least 50% confident that it is so.

The probability is already just an expression of your own uncertainty. Giving a confidence interval over the probability does not make sense.

Comment author: bentarm 04 October 2011 01:33:41PM *  1 point [-]

I basically agree that the part of the original comment that you quote doesn't make any sense at all, and am not attempting to come to the defence of confidence intervals over probabilities, but it does feel like there should be some way of giving statements of probability and indicating how sure one is about the statement at the same time. I think, in some sense, I want to be able to say how likely I think it is that I will get new information that will cause me to update away from my current estimate, or give a second-derivative of my uncertainty, if you will.

Let's say we have two bags, one contains 1 million normal coins, one contains 500,000 2-headed coins and 500,000 2-tailed coins. Now, I draw a coin from the first bag and toss it - I have a 50% chance of getting a head. I draw a coin from the second bag and toss it - I also have a 50% chance of getting a head, but it does feel like there's some meaningful difference between the two situations. I will admit, though, that I have basically no idea how to formalise this - I assume somebody, somewhere, does.