Sewing-Machine comments on Best career models for doing research? - Less Wrong

27 Post author: Kaj_Sotala 07 December 2010 04:25PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (999)

You are viewing a single comment's thread. Show more comments above.

Comment author: [deleted] 08 December 2010 02:02:04PM 1 point [-]

assuming that the entire 80 billion galaxies' fate is decided then.

What's your P of "the fate of all 80 billion galaxies will be decided on Earth in the next 100 years"?

Comment author: FormallyknownasRoko 08 December 2010 02:15:18PM 0 points [-]

Some complexities regarding "decided" since physics is deterministic, but hand waving that aside, I'd say 50%.

Comment author: [deleted] 09 December 2010 12:50:42AM 1 point [-]

With high probability, many of those galaxies are already populated. Is that irrelevant?

Comment author: FormallyknownasRoko 09 December 2010 12:24:19PM 0 points [-]

I disagree. I claim that the probability of >50% of the universe being already populated (using the space of simultaneity defined by a frame of reference comoving with earth) is maybe 10%.

Comment author: [deleted] 09 December 2010 01:33:40PM 0 points [-]

"Already populated" is a red herring. What's the probability that >50% of the universe will ever be populated? I don't see any reason for it to be sensitive to how well things go on Earth in the next 100 years.

Comment author: FormallyknownasRoko 09 December 2010 06:32:33PM 1 point [-]

I think it is likely that we are the only spontaneously-created intelligent species in the entire 4-manifold that is the universe, space and time included (excluding species which we might create in the future, of course).

Comment author: [deleted] 09 December 2010 06:58:44PM 1 point [-]

I'm curious to know how likely, and why. But do you agree that aliens are relevant to evaluating astronomical waste?

Comment author: timtyler 09 December 2010 06:37:12PM 0 points [-]

That seems contrary to the http://en.wikipedia.org/wiki/Self-Indication_Assumption

Do you have a critique - or a supporting argument?

Comment author: FormallyknownasRoko 09 December 2010 06:38:43PM *  3 points [-]

Yes, I have a critique. Most of anthropics is gibberish. Until someone makes anthropics work, I refuse to update on any of it. (Apart from the bits that are commonsensical enough to derive without knowing about "anthropics", e.g. that if your fising net has holes 2 inches big, don't expect to catch fish smaller then 2 inches wide)

Comment author: timtyler 09 December 2010 07:59:08PM 2 points [-]

I don't think you can really avoid anthropic ideas - or the universe stops making sense. Some anthropic ideas can be challenging - but I think we have got to try.

Anyway, you did the critique - but didn't go for a supporting argument. I can't think of very much that you could say. We don't have very much idea yet about what's out there - and claims to know such things just seem over-confident.

Comment author: FormallyknownasRoko 09 December 2010 08:18:16PM 1 point [-]

Basically Rare Earth seems to me to be the only tenable solution to Fermi's paradox.

Comment author: [deleted] 09 December 2010 06:55:14PM 0 points [-]

There are strained applications of anthropics, like the doomsday argument. "What happened here might happen elsewhere" is much more innocuous.

Comment author: [deleted] 09 December 2010 06:58:45PM 1 point [-]

There are some more practical and harmless applications as well. In Nick Bostrom's Anthropic Bias, for example, there is an application of the Self-Sampling Assumption to traffic analysis.

Comment author: Vladimir_Nesov 09 December 2010 06:40:13PM 0 points [-]

I agree.

Comment author: [deleted] 09 December 2010 06:46:12PM *  2 points [-]

Even Nick Bostrom, who is arguably the leading expert on anthropic problems, rejects SIA for a number of reasons (see his book Anthropic Bias). That alone is a pretty big blow to its credibility.

Comment author: timtyler 09 December 2010 07:49:53PM *  0 points [-]

That is curious. Anyway, the self-indication assumption seems fairly straight-forwards (as much as any anthropic reasoning is, anyway). The critical material from Bostrom on the topic I have read seems unpersuasive. He doesn't seem to "get" the motivation for the idea in the first place.

Comment author: Kevin 09 December 2010 02:28:22PM *  0 points [-]

If you think there is a significant probability that an intelligence explosion is possible or likely, then that question is sensitive to how well things go on Earth in the next 100 years.

Comment author: [deleted] 09 December 2010 03:06:06PM *  3 points [-]

However likely they are, I expect intelligence explosions to be evenly distributed through space and time. If 100 years from now Earth loses by a hair, there are still plenty of folks around the universe who will win or have won by a hair. They'll make whatever use of the 80 billion galaxies that they can--will they be wasting them?

If Earth wins by a hair, or by a lot, we'll be competing with those folks. This also significantly reduces the opportunity cost Roko was referring to.

Comment author: Vladimir_Nesov 08 December 2010 02:31:26PM *  0 points [-]

About 10% (if we ignore existential risk, which is a way of resolving the ambiguity of "will be decided"). Multiply that by opportunity cost of 80 billion galaxies.

Comment author: David_Gerard 08 December 2010 02:58:15PM 1 point [-]

Could you please detail your working to get to this 10% number? I'm interested in how one would derive it, in detail.

Comment author: Vladimir_Nesov 08 December 2010 03:20:26PM *  0 points [-]

I restored the question as asking about probability that we'll be finishing an FAI project in the next 100 years. Dying of engineered virus doesn't seem like an example of "deciding the fate of 80 billion galaxies", although it's determining that fate.

FAI looks really hard. Improvements in mathematical understanding to bridge comparable gaps in understanding can take at least many decades. I don't expect a reasonable attempt at actually building a FAI anytime soon (crazy potentially world-destroying AGI projects go in the same category as engineered viruses). One possible shortcut is ems, that effectively compress the required time, but I estimate that they probably won't be here for at least 80 more years, and then they'll still need time to become strong enough and break the problem. (By that time, biological intelligence amplification could take over as a deciding factor, using clarity of thought instead of lots of time to think.)

Comment author: [deleted] 09 December 2010 01:00:46AM 0 points [-]

My question has only a little bit to do with the probability that an AI project is successful. It has mostly to do with P(universe goes to waste | AI projects are unsuccessful). For instance, couldn't the universe go on generating human utility after humans go extinct?

Comment author: ata 09 December 2010 01:05:09AM 1 point [-]

For instance, couldn't the universe go on generating human utility after humans go extinct?

How? By coincidence?

(I'm assuming you also mean no posthumans, if humans go extinct and AI is unsuccessful.)

Comment author: [deleted] 09 December 2010 01:23:45AM 2 points [-]

Aliens. I would be pleased to learn that something amazing was happening (or was going to happen, long "after" I was dead) in one of those galaxies. Since it's quite likely that something amazing is happening in one of those 80 billion galaxies, shouldn't I be pleased even without learning about it?

Of course, I would be correspondingly distressed to learn that something horrible was happening in one of those galaxies.