nshepperd comments on Efficient Charity: Do Unto Others... - Less Wrong

130 Post author: Yvain 24 December 2010 09:26PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (318)

You are viewing a single comment's thread. Show more comments above.

Comment author: Will_Sawin 26 December 2010 03:25:08AM 0 points [-]

So we need to formalize this, obviously.

Method 1: Exponential discounting.

Problem: You don't care very much about future people.

Method 2: Taking the average over all time (specifically the limit as t goes to infinity of the integral of utility from 0 to t, divided by t)

Conclusion which may be problematic: If humanity does not live forever, nothing we do matters.

Caveat: Depending on our anthropics, we can argue that the universe is infinite in time or space with probability 1, in which case there are an infinite number of copies of humanity, and so we can always calculate the average. This seems like the right approach to me. (In general, using the same math for your ethics and your anthropics has nice consequences, like avoiding most versions of Pascal's Mugging.)

Comment author: nshepperd 26 December 2010 05:44:21AM 2 points [-]

How about this:

If you have a set E = {X, Y, Z...} of possible actions, A (in E) is the utility-maximising action iff for all other B in E, the limit

is greater than zero, or approaches zero from the positive side. Caveat: I have no evidence this doesn't implode in some way, perhaps by the limit being undefined. This is just a stupid idea to consider. A possibly equivalent formulation is

The inequality being greater or equal allows for two or more actions being equivalent, which is unlikely but possible.

Comment author: DSimon 27 December 2010 06:59:42PM 4 points [-]

Side comment: that math equation image generator you used is freakin' excellent. The image itself is generated based from the URL, so you don't have to worry about hosting. Editor is here.

Comment author: nshepperd 28 December 2010 03:06:43AM *  4 points [-]

I prefer this one, which automatically generates the link syntax to paste into a LW comment. There's a short discussion of all this on the wiki.

Comment author: Will_Sawin 26 December 2010 12:28:12PM 1 point [-]

Functions whose limit is +infinity and -infinity can be distinguished, so your good there.

I think it's the same as my second: As long as the probability given both actions of a humanity lasting forever is nonzero, and the differences of expected utilities far in the future is nonzero, nothing that happens in the first million billion years matters.

Comment author: nshepperd 27 December 2010 09:30:13AM *  0 points [-]

The difference in expected utility would have to decrease slow enough (slower than exponential?) to not converge, not just be nonzero. [Which would be why exponential discounting "works"...]

However I would be surprised to see many decisions with that kind of lasting impact. The probability of an action having some effect at time t in the future "decays exponentially" with t (assuming p(Effect_t | Effect_{t-1}, Action) is approximately constant), so the difference in expected utility will in general fall off exponentially and therefore converge anyway. Exceptions would be choices where the utilities of the likely effects increase in magnitude (exponentially?) as t increases.

Anyway I don't see infinities as an inherent problem under this scheme. In particular if we don't live forever, everything we do does indeed matter. If we do live forever, what we do does matter, excepts how it affects us might not if we anticipate causing "permanant" gain by doing something.

Comment author: Sniffnoy 26 December 2010 06:04:15AM 0 points [-]

Can't think about the underlying idea right now due to headache, but instead of talking about any sort of limit, just say that it's eventually positive, if that's what you mean.