I haven't read the entire post but I believe I solved "infinite ethics" in http://lesswrong.com/lw/jub/updateless_intelligence_metrics_in_the_multiverse/ (by sticking to a bounded utility function with discounts of particular asymptotics resulting from summing over a Solomonoff ensemble).
Thanks. As per theorem 3.2 above you can't have both Pareto and an anonymity constraint. Finite anonymity would add a constant factor to the complexity of the utility vector and hence shouldn't affect the prior, so I assume your method follows the finite anonymity constraint.
As a result, you must be disobeying Pareto? It's not obvious to me why your solution results in this, so I'm bringing it up in case it wasn't obvious to you either. (Or it could be that I'm completely misunderstanding what you are trying to do. Or maybe that you don't think Pareto is actually a reasonable requirement. In any case I think at least one of us is misunderstanding what's going on.)
It seems to me there is another principle that needs to be considered in practical ethics.
When we are confronted with a situation with two mutually exclusive options, standard utility calculation normally allows that they may be of equal value. Standard economics agrees, allowing two goods to be of equal value. But an agent always chooses one or the other. Even when dealing with fungible commodities — say, two identical five-pound bags of rice on a shelf — we always do, in practice, end up choosing one or the other. Even if we flip a coin or use some other random decision procedure, ultimately one good ends up in the shopping cart and the other good ends up staying on the shelf.
To avoid Buridan's Ass situations, we must always end up ranking one good above the other. Otherwise we end up with neither good, which is worse than choosing arbitrarily.
"Should two courses be judged equal, then the will cannot break the deadlock, all it can do is to suspend judgement until the circumstances change, and the right course of action is clear." — Jean Buridan
If choice A has utility 1, choice B also has utility 1, and remaining in a state of indecision has utility 0, then we can't allow the equality between A and B to result in us choosing the lower utility option.
This is especially important when time comes into play. Consider choosing between two mutually exclusive activities, each of which produces 1 util per unit time. The longer you spend trying to decide, the less time you have to do either one.
One solution to this seems to be to deny equality. Any time we perform a comparison between two utilities, it always returns < or >, never =. Any two options are ranked, not numerically evaluated.
In this system, utilities do not behave like real numbers; they are ordered but do not have equality.
(Crossposted from the EA forum.)
Summary: The universe may very well be infinite, and hence contain an infinite amount of happiness and sadness. This causes several problems for altruists; for example: we can plausibly only affect a finite subset of the universe, and an infinite quantity of happiness is unchanged by the addition or subtraction of a finite amount of happiness. This would imply that all forms of altruism are equally ineffective.
Like everything in life, the canonical reference in philosophy about this problem was written by Nick Bostrom. However, I found that an area of economics known as "sustainable development" has actually made much further progress on this subject than the philosophy world. In this post I go over some of what I consider to be the most interesting results.
NB: This assumes a lot of mathematical literacy and familiarity with the subject matter, and hence isn't targeted to a general audience. Most people will probably prefer to read my other posts:
1. Summary of the most interesting results
2. Definitions
In general, we consider a population to consist of an infinite utility vector (u0,u1,…) where ui is the aggregate utility of the generation alive at time i. Utility is a bounded real number (the fact that economists assume utility to be bounded confused me for a long time!). Our goal is to find a preference ordering over the set of all utility vectors which is in some sense “reasonable”. While philosophers have understood for a long time that finding such an ordering is difficult, I will present several theorems which show that it is in fact impossible.
Due to a lack of latex support I’m going to give English-language definitions and results instead of math-ey ones; interested people should look at the papers themselves anyway.
3. Impossibility Results
3.1 Definitions
3.2 Diamond-Basu-Mitra Impossibility Result1
3.3 Zame's impossibility result2
4. Possibility Results
We’ve just shown that it’s impossible to construct or even prove the existence of any useful ethical system. But not all hope is lost!
The important idea here is that of a “subrelation”: < is a subrelation to <’ if x<y implies x<’y.
Our arguments will work like this:
Suppose we could extend utilitarianism to the infinite case. (We don't, of course, know that we can extend utilitarianism to the infinite case. But suppose we could.) Then A, B and C must follow.
Technically: suppose utilitarianism is a subrelation of <. Then < must have properties A, B and C.
Everything in this section comes from (3), which is a great review of the literature.
4.1 Definition
4.2 Theorem
4.3 Definition
4.4 Theorem
4.5 Definition
4.6 Theorem
4.7 Definition
4.8 Theorem
4.9 Definition
4.10 Theorem
5. Conclusion
I hope I've convinced you that there's a "there" there: infinite ethics is something that people can make progress on, and it seems that most of the progress is being made in the field of sustainable development.
Fun fact: the author of the last theorem (the one which defined "sustainable") was one of the lead economists on the Kyoto protocol. Who says infinite ethics is impractical?
6. References