Nice article. I think it's a mistake for Harsanyi to argue for average utilitarianism. The view has some pretty counterintuitive implications:
If the ancient Egyptians were very happy, my child would bring down the average, and so having the child would be wrong. If the ancient Egyptians were unhappy, my child would bring up the average, and so having the child would be right.
I do prefer total utilitarianism to average utilitarianism,[1] but one thing that pulls me to average utilitarianism is the following case.
Let's suppose Alice can choose either (A) create 1 copy at 10 utils, or (B) create 2 copies at 9 utils. Then average utilitarianism endorses (A), and total utilitarianism endorses (B). Now, if Alice knows she's been created by a similar mechanism, and her option is correlated with the choice of her ancestor, and she hasn't yet learned her own welfare, then EDT endorses picking (A). So that matches average utilitarianism.[2]
Basically, you'd be pleased to hear that all your ancestors were average utility maximisers, rather than total utility maximisers, once you "update on your own existence" (whatever that means). But also, I'm pretty confused by everything in this anthropics/decision theory/population ethics area. Like, the egyptology thing seems pretty counterintuitive, but acausal decision theories and anthropic considerations imply all kind of weird nonlocal effects, so idk if this is excessively fishy.
I think aggregative principles are generally better than utilitarian ones. I'm a fan of LELO in particular, which is roughly somewhere between total and average utilitarianism, leaning mostly to the former.
Maybe this also requires SSA??? Not sure.
Yeah I think correlations and EDT can make things confusing. But note that average utilitarianism can endorse (B) given certain background populations. For example, if the background population is 10 people each at 1 util, then (B) would increase the average more than (A).
1. Introduction
Utilitarianism is the view that a social planner should choose options which maximise the social utility of the resulting social outcome. The central object in utilitarianism is the social utility function u:S→R which assigns a real value u(s)∈R to each social outcome s∈S. This function typically involves variables such as the well-being, preferences, and mental states of individuals, distributional factors like inequality, and other relevant factors such as justice, social cohesion, and freedoms. Utilitarianism is a broad class of social choice principles, one corresponding to each function u:S→R.
In my previous article, I introduced aggregative principles, which state that a social planner should make decisions as if they will face the aggregated personal outcomes of every individual in the population. The central object in aggregativism is the function ζ:S→P, represented with the the greek letter zeta, which assigns a personal outcome ζ(s)∈P to each social outcome s∈S. This function typically aggregates the collection of personal outcomes facing the entire population into a single personal outcome. Aggregativism is a broad class of social choice principles, one corresponding to each function ζ:S→P.
We examined three well-known aggregative principles:
I'm interested in aggregative principles because they avoid many theoretical pitfalls of utilitarian principles. Unlike utilitarianism, aggregativism doesn't require specifying a social welfare function, which is notoriously intractable. Moreover, it seems less prone to counterintuitive conclusions such as the repugnant conclusion or the violation of moral side constraints.[1] In this article, I will show that, under natural conditions of human rationality, aggregative principles approximate utilitarian principles. Therefore, even though aggregativism avoids these theoretical pitfalls, we should nonetheless expect aggregativism to generate roughly-utilitarian recommendations in practical social contexts, and thereby retain the most appealing insights from utilitarianism.
The rest of the article is organized as follows. Section 2 formalises social choice principles as functions of type (X→S)→P(X). Section 3 demonstrates the structural similarity between two strategies to specifying such principles, namely the aggregative and utilitarian strategies. Section 4 proves that under natural conditions about human rationality, the aggregative and utilitarian principles are mathematically equivalent. This theorem is the key contribution of the article. Sections 5, 6, and 7 applies the theorem to LELO, HL, and ROI respectively.
2. Social choice principles
Suppose you are a social planner choosing from a set of options X={x1,…,xn}. The set X might be the set of available tax rates, environmental policies, military actions, political strategies, neural network parameters, or whatever else is being chosen by the social planner. Now, your choice will presumably depend on the social consequences of the options, even if you also consider non-consequentialist factors. We can model the social consequences with a function f:X→S, where S is the set of social outcomes. In particular, if you choose an option x∈X, then the resulting social outcome would be f(x)∈S.
We call f:X→S the "social context". As a concrete example, suppose the options are different tax rates (say 10%, 20%, and 30%), and the social outcomes are characterized by variables like total tax revenue, income inequality, and unemployment rate. Then the social context is the function f:[0,1]→S which maps each tax rate x∈[0,1] to the resulting values of these social outcome variables.
A social choice principle should say, for each social context, which options are acceptable. Formally, a social choice principle is characterised by some function Ψ:(X→S)→P(X), which takes a social context f:X→S as input and returns a subset of the options Ψ(f)⊆X as output. Specifically, Ψ(f)⊆X consists of exactly those options which satisfy the principle in the social context f:X→S.
Note that (X→S) denotes the set of all functions from X to S, so Ψ is a higher-order function, meaning it receives another function as input. Additionally P(X) denotes the powerset of X, i.e. the set of subsets of X. We use the powerset P to allow for the fact that multiple options may satisfy a principle: if a principle Ψ permits only options x1 and x2 in a context f:X→S then Ψ(f)={x1,x2}. Finally, the powerset P(X) includes the empty set ∅, which allows for the case Ψ(f)=∅. Informally, Ψ(f)=∅ means that the social planner, following principle Ψ and faced with context f:X→S, has no acceptable options, which allows for principles that aren't universally satisfiable.
Here are some examples of social choice principle:
Let X0⊆X be any fixed subset of the options, and consider the principle Ψ:f↦X0, which returns X0 regardless of the input f. Whether an option x∈X satisfies this principle depends only on whether x∈X0, and is independent of the social context. At one extreme, there's a trivial principle Ψ:f↦X which never constrains the social planner, and at the other extreme, there's a principle Ψ:f↦∅ which is always unsatisfiable. When X0={x0} consists of a single option, the principle Ψ states that the social planner must choose x0∈X regardless of the context.[2]
Let Starget⊆S be any fixed subset of the social outcomes, whose elements we'll call targets. There is a principle which says that the social planner should choose an option which achieves a target. This principle is characterised by the function Ψ:f↦f−1(Starget), where f−1(Starget)⊆X denotes the preimage of Starget, i.e. f−1(Starget):={x∈X:f(x)∈Starget}. Note that if f(x)∉Starget for all x∈X then Ψ(f)=∅, i.e. if no option would achieve a target then the principle says all options are unacceptable.
We can also characterise more unusual principles as functions (X→S)→P(X). For example, consider the principle that says a social planner should choose an option if most other options would've led to the same social outcome. Intuitively, this captures some notion of impact minimisation. Formally, this principle is characterised by the function Ψ(f):={x∈X∣#[x]f>#X2}, where #S denote the cardinality of a set and [x]f⊆X denotes the f-equivalency class of x∈X, i.e. [x]f={y∈X∣f(x)=f(y)}.
These examples illustrate the diversity of conceivable social choice principles. The key point is that they can all be represented by functions Ψ:(X→S)→P(X). I've found this a productive way to think about principles of decision-making, and agency more generally.[3] Finding compelling social choice principle is the central problem in social ethics, and different normative frameworks will propose different principles.
3. Two strategies for specifying principles
3.1. Utilitarian principles
Utilitarianism and aggregativism are two strategies for specifying a social choice principle Ψ:(X→S)→P(X). The utilitarian strategy specifies a social choice principle using two components:
Given the social utility function u:S→R and the operator argmaxX:(X→R)→P(X), the utilitarian principle is defined by Ψ(f):=argmaxX(u∘f). Note that if f:X→S is the social context, then the composition u∘f:X→R calculates the social utility resulting from each option, thereby providing a real-valued function r:X→R. The utilitarian principle f↦argmaxX(u∘f) says that the social planner should choose an option that maximizes this function.
As a simplistic example, consider a social utility function u:S→R that measures the gross world product of a social outcome. The resulting utilitarian principle f↦argmaxX(u∘f) would oblige maximizing gross world product. In practice, utilitarians typically endorse more nuanced utility functions that account for factors like individual well-being, fairness, and existential risk.
3.2. Aggregative principles
Aggregativism offers an alternative strategy to specifying social choice principle. Like utilitarianism, it defines the principle Ψ:(X→S)→P(X) using two components:
The function Π:(X→P)→P(X) should model a self-interested human in the following sense: for each personal context g:X→P the subset Π(f)⊆X should contain the options that the hypothetical human might choose in that context. A personal context g:X→P is an assignment of a personal outcome to each of the options, analogously to a social context. For example, if g:X→P maps some options to finding a dollar and the remaining options to drowning in a swamp, then presumably Π(g) contains only the former options.
Given the social zeta function ζ:S→P and a model of self-interested human Π:(X→P)→P(X), the aggregationist principle is defined by Ψ(f):=Π(ζ∘f). Note that if f:X→S is the social context, then the composition ζ∘f:X→P calculates the hypothetical personal outcome resulting from each option, thereby providing a personal context g:X→P. The aggregative principle f↦Π(ζ∘f) says that the social planner should choose an option a self-interested human might choose in this personal context.
For example, consider a social zeta function ζ:S→P that maps each social outcome s to the personal outcome of living every individual's life in sequence, starting with the earliest-born humans. The resulting aggregative principle f↦Π(ζ∘f) obliges affecting society such that living the concatenated lives is personally desirable.
3.3. Structural similarity between the two strategies
This comparison reveals the structural similarity between utilitarianism and aggregativism. Both strategies specify the principle Ψ using a two components:
Both Π, the model of a self-interested human, and the argmaxX operator are choice principles: Π is a personal choice principle, it 'chooses' one the options based on their associated personal outcomes, and argmaxX is a real choice principle, it 'chooses' one of the options based on their associated real value. (Of course, argmaxX doesn’t literally choose anything, it’s simply a mathematical operator, but so too is Π.)
In general, for any space R, let's say an R-context is any function with type-signature X→R, and an R-choice principle is any function with type-signature (X→R)→P(X). That is, an R-choice principle Φ, when provided with an R-context r:X→R, identifies some subset Φ(f)⊆X of the options which are 'acceptable'.
How might one use an R-choice principle Φ to specify a social choice principle Ψ? Well, what's needed is some function σ:S→R from social outcomes to elements of R. This function σ will extends any social context f:X→S to an R-context σ∘f:X→R, which can then be provided to the R-choice principle to identify the acceptable options. Formally, Ψ:f↦Φ(σ∘f). This is how utilitarianism and aggregativism succeed in defining social choice principles. The key difference is that utilitarianism uses real numbers while aggregativism uses personal outcomes.
4. Equivalence between aggregativism and utilitarianism
4.1. Three conditions for equivalence
Despite their differences, there are natural conditions under which the utilitarian and aggregative principles are equivalent, in the sense that a social planner is permitted to choice an option, under the utilitarian principle, if and only if they are permitted to choice the same option under the aggregative principle.
Formally, let Ψu denote the utilitarian principle Ψu:f↦argmaxX(u∘f) and let Ψa denote the aggregative principle Ψa:f↦Π(ζ∘f); under what conditions does Ψu(f)=Ψa(f) for all social contexts f:X→S?
In the previous article, we showed that LELO, HL, and ROI each employ social zeta functions which aggregates the personal outcomes across all individuals in the population. Formally, ζ(s):=α(γ(−,s)M(π)) where I is a fixed set of individuals; γ:I×S→P is a fixed function mapping a social outcome s∈S and an individual i∈I to the personal outcome γ(i,s)∈P that i faces when s obtains; M is the monad capturing a notion of 'collection'; π∈M(I) be a fixed collection of individuals impartially representing the population; and α:M(P)→P is an M-algebra specifying how to aggregate collections of personal outcomes into a single personal outcome.
Supposing ζ has the general form above, and the three conditions below are satisfied, then the utilitarian principle Ψu and the aggregative principle Ψa are mathematically equivalent:
Formally, the first condition states that the function Π:(X→P)→P(X) has the form Π(f)=argmaxX(v∘f) for some personal utility function v:P→R which assigns a real-valued utility v(p)∈R to each personal outcome p∈P. Even by itself, this condition is quite strong. It implies that if, for some personal context f:X→P, two options x1 and x2 result in the same personal outcome, i.e. f(x1)=f(x2), then the human might choose x1 if and only if they might choose x2. Hence, this condition precludes nonconsequential considerations.
Let's call this condition "Humans Maximise Personal Utility" (HMPU).
Let α:M(P)→P denote an M-algebra on personal outcomes, describing how to aggregate a collection of personal outcomes into a single personal outcome. Let β:M(R)→R denote an M-algebra on real numbers, describing how to aggregate a collection of real numbers into a single real number. The second condition states that v∘α=β∘vM. Informally, this condition means that the personal utility of an aggregate of personal outcomes is the aggregate of the personal utilities of each personal outcome being aggregated. In mathematical jargon, the personal utility function v:P→R must be a homomorphism between the M-algebras (P,α) and (R,β), which means it preserves the algebraic structure on P and R.
Let's call this condition "Rationality of Personal Utility" (RUP).
Formally, the third condition states that u(s)=β((v∘γ)(−,s)M(π)), where v:P→R is the personal utility function introduced in HMPU, γ:I×S→P is the function assigning personal outcomes to each individual in each social outcome, π∈M(I) is the distinguished collection of individuals representing the population, and β:M(R)→R is the M-algebra describing how to aggregate a collection of real numbers into a single real number. Informally, this condition states that the social utility of a social outcome is the aggregate of the personal utilities of the personal outcomes faced by all individuals in the population.
Let's call this condition "Social Utility Aggregates Personal Utilities" (SUAPU).
The aggregative principle (when our model of a self-interested human is a rational personal utility maximiser) is equivalent to the utilitarian principle (when social utility is the impartial aggregation of personal utility over each individual). The full proof is elementary and uninsightful.[4]
Now, these three conditions are only approximately true, and they fail in systematic ways. However, the theorem will help elucidate exactly the extent to which the aggregative principle approximates the corresponding utilitarian principle. Namely, the aggregative principle will approximate the utilitarian principle to the degree that these conditions hold.
Because RPU and SUAPU depend on the specific monad M under discussion, I will spell out the details for three paradigm examples: the list monad List (representing finite sequences), the distribution monad Δ (representing probability distributions), and the nonempty finite powerset monad P+f (representing nonempty finite sets).
5. Equivalence between LELO and longtermist total utilitarianism
The previous section proved an equivalence, under certain conditions, between aggregative principles and utilitarian principles. This section will apply that theorem to the monad List, which is used to formalise Live Every Life Once (LELO). We will see that LELO is equivalent to longtermist total utilitarianism.
5.2. Monoidal rationality of personal utility?
The real numbers admit a concatenation operator in the obvious way, i.e., there exists a function sum:List(R)→R defined by sum([r1,…,rk]):=0+r1+⋯+rk. This is simply the well-known summation operator, which sends a list of real values to their sum.
Let's unpack RPU, which formally states that v∘conc=sum∘vList. In other words, for any list of personal outcomes [p1,…,pn], we have equality between v∘conc([p1,…,pn]) and 0+v(p1)+⋯+v(pn). Informally, the personal utility of a concatenated outcome equals the sum of the personal utilities of the outcomes being concatenated. This 'monoidal' rationality condition constrains how humans must value the concatenation of different personal outcomes.
In the previous article, we saw that the concatenation operator conc:List(P)→P can be equivalently presented by a binary operator ▹ and a constant ϵ∈P, with the intended interpretation p▹p′:=conc([p1,p2]) and ϵ:=conc([]). We can restate the RPU condition in terms of ▹ and ϵ with two equations: v(ϵ)=0 and v(p▹p′)=v(p)+v(p′) for all p,p′∈P.
How realistic is the RPU condition? That is, supposing humans do maximise a personal utility function, how monoidally rational is it? I think this condition is approximately true, but unrealistic in several ways. I'll assume here that p▹p′ is interpreted as facing p and then facing p′ in sequence, rather than some exotic notion of concatenation.
Firstly, RPU rules out permutation-dependent values. It precludes a personal utility function v:P→R such that v(p1▹p2)≠v(p2▹p1). Informally, RPU assumes human values must be invariant to the ordering of experiences: they cannot value saving the best till last, nor saving the worse till last. In particular, RPU assumes that humans values are time-symmetric, which seems unrealistic, as illustrated by the following examples. Compare the process of learning, i.e. ending with better beliefs than one started with, with the process of unlearning, i.e. ending with worse beliefs than one started with. Humans seem to value learning above unlearning, but such time-asymmetric values are precluded by RPU. Similarly, humans seem to value a history of improvement over a history of degradation, even if both histories are different permutations of the same list of moments, but such values are precluded by RPU.
Secondly, RPU rules out time-discounted values. Under exponential time-discounting, a common assumption in economics, the personal utility function v:P→R obeys the equation v(p1▹p2)=v(p1)+(1+δ)−duration(p1)⋅v(p2). Here duration:P→R≥0 gives the duration of each outcome and δ>0 is the discount rate. This discounting formula weights the first outcome p1 more than the second outcome p2, with the difference growing exponentially with the duration of p1. For instance, let p1 and p′1 be equally valuable experiences lasting different durations, like a minute of ecstasy and a week of contentment respectively. Time-discounting implies that v(p1▹p2) depends more on v(p2) than v(p′1▹p2) does. However, RPU precludes this possibility, as it requires that δ=0, i.e. that humans are equally concerned with all life stages, not discounting future rewards relative to present ones
Thirdly, RPU rules out path-dependent values. Informally, whether I value a future p more than a future q must be independent of my past experiences. But this is an unrealistic assumption about human values, as illustrated in the following examples. If p denotes reading Moby Dick and q denotes reading Oliver Twist, then humans seem to value p▹p less than p▹q but value q▹p more than q▹q. This is because humans value reading a book higher if they haven't already read it, due to an inherent value for novelty in reading material. Alternatively, if p and q denote being married to two different people, then humans seem to value p▹p more than p▹q but value q▹p less than q▹q. This is because humans value being married to someone for a decade higher if they've already been married to them, due to an inherent value for consistency in relationships.[5] But RPU would precludes such path-dependent values.
5.2. Social utility sums personal utility?
Now let's unpack SUAPU, which formally states that u(s)=sum((v∘γ(−,s))List(l)). In other words, the social utility function is the sum of personal utilities over the individuals in the distinguished list representing the population. That is, if l=[i1,…,in]∈List(I) is a list of individuals representing the entire population impartially, then for any social outcome s∈S, its social utility u(s) is given by v(γ(i1,s))+⋯+v(γ(in,s)).
How realistic is the SUAPU condition? The answer depends on one's axiological theory. Indeed, SUAPU is a statement of longtermist total utilitarianism. This is a strong assumption that precludes a social utility function from exhibiting certain properties, analogous to how RPU constrains the personal utility function. Specifically, SUAPU precludes social utility functions with the following features:
Nonetheless, I think that HMPU, RPU, and SUAPU are useful approximations, even if they aren't perfectly true. To the extent that these assumptions do hold, Live Every Life Once (LELO) and longtermist total utilitarianism will be roughly equivalent. This explains why MacAskill appeals to LELO to argue for longtermist utilitarianism in his book "What We Owe The Future" (2022). Indeed, MacAskill's implicit argument can be summarized as follows:
6. Equivalence between HL and average utilitarianism
We've seen how to apply the general equivalence, under certain conditions, between aggregative principles and utilitarian principles, e.g. between LELO and longtermist total utilitarianism. This section will apply that theorem to the monad Δ, which is used to formalise Harsanyi's Lottery (HL). We will see that HL is equivalent to average utilitarianism.
6.1. Convex rationality of personal utility?
The real numbers admit an interpolation operator in the obvious way, i.e., there exists a function mean:Δ(R)→R defined by mean(⟨r1:λ1∣…∣rn:λn⟩):=λ1⋅r1+⋯+λn⋅rn. This is simply the well-known mean-value operator, which sends a distribution of real values to their weighted average.
Let's unpack RPU, which formally states that v∘E=mean∘vΔ. In other words, for any distribution of personal outcomes ⟨p1:λ1∣⋯∣pn:λn⟩, we have equality between v(E[⟨p1:λ1∣⋯∣pn:λn⟩]) and λ1⋅v(p1)+⋯+λn⋅v(pn). Informally, the personal utility of an interpolated outcome is the average of the personal utilities outcomes being interpolated. This 'convex' rationality condition constrains how humans must value the interpolation of different personal outcomes.
In the previous article, we saw that the interpolation operator E:Δ(P)→P can be equivalently presented by a family of binary operators {+λ:λ∈(0,1)}, with the intended interpretation p+λp′=E(⟨p:λ∣p′:1−λ⟩). We can restate the RPU condition in terms of +λ with the family of equations:v(p+λp′)=λ⋅v(p)+(1−λ)⋅v(p′).
How realistic is the RPU condition? That is, supposing humans do maximise a personal utility function, how convexly rational is it? I think this condition is approximately true, but unrealistic in several ways. I'll assume here that p+λp′ is interpreted as a lottery between p with likelihood λ and p′ with likelihood 1−λ.
Firstly, RPU rules out valuing determinacy. Informally, a lottery can't be valued below each determinate outcomes. But perhaps this is an unrealistic assumption, as illustrated in the following examples. If p1 denotes dying on Monday and p2 denotes dying on Tuesday, then humans might value both determinate outcomes over the lottery between them, e.g. v(p1)≈v(p2) but v(p1+0.6p2)<v(p1),v(p2). This is because humans may inherently value determinacy about the day of their death. But RPU precludes valuing determinacy.
Secondly, RPU rules out valuing randomness. Informally, a lottery can't be valued above each determinate outcomes. But perhaps this is an unrealistic assumption, as illustrated in the following examples. If p1 and p2 denotes marrying two different people, then humans might value both determinate outcomes less than the lottery between them, e.g. v(p1)≈v(p2) but v(p1+0.6p2)>v(p1),v(p2). This is because humans may inherently value randomness about whom they marry. Again, RPU precludes valuing randomness.
Thirdly, RPU rules out rules out values discontinuous in the underlying likelihoods. Formally, if λi→λ∞ is a convergent sequence in (0,1), then RPU implies v(p+λip′)→v(p+λp′). Moreover, if λi→0 then v(p+λp′)→v(p′) and if λi→1 then v(p+λp′)→v(p). But perhaps this is an unrealistic assumption, as illustrated in the following examples. If pokay denotes an okay outcome and pcata denotes a catastrophic outcome, then humans might value the lottery pcata+λpokay substantially less than pokay for all λ∈(0,1), i.e. limλ→0v(pcata+λpokay)<v(pokay). Informally, the human values the zeroness catastrophe's likelihood. Analogously, if pdefeat denotes a terrible defeat and pvictory denotes a great victory, then humans might value the lottery pvictory+λpdefeat substantially more than pdefeat for all λ∈(0,1), i.e. limλ→0v(pvictory+λpdefeat)>v(pdefeat). Informally, the human values the nonzeroness victory's likelihood. But RPU precludes valuing either zeroness or nonzeroness of likelihoods, because this value would be discontinuous in the underlying likelihoods.
That being said, I think human values approximate convex rationality far better than they approximate monoidal rational. In fact, while mainstream economics does not assume monoidal rationality (e.g. via time-discounting) it does assume convex rationality. Convex rationality is a straightforward application of von Neumann-Morgenstern (VNM) expected utility theory. Hence, I accept convex rationality of human values, at least when interpolation p+λp′ is interpreted as a lottery between p and p′.[6]
6.2. Social utility averages personal utility?
Now let's unpack SUAPU, which formally states that u(s)=mean((v∘γ(−,s))Δ(π)). In other words, the social utility function is the weighted average of personal utility over the individuals in distinguished distribution representing the population. That is, if π=⟨i1:λ1∣⋯∣in:λn⟩∈Δ(I) is a distribution of individuals representing the entire population impartially, then for any social outcome s∈S, its social utility u(s) is given by λ1⋅v(γ(i1,s))+⋯+λn⋅v(γ(in,s)).
How realistic is the SUAPU condition? The answer depends on one's axiological theory. Indeed, SUAPU is a statement of average utilitarianism. This is a strong assumption that precludes a social utility function from exhibiting certain properties, analogous to how RPU constrains the personal utility function. Specifically, SUAPU precludes social utility functions with the following features:
Nonetheless, I think that HMPU, RPU, and SUAPU are reasonable approximations, even if not perfectly true. To the extent that these assumptions hold, Harsanyi's Lottery (HL) and average utilitarianism will be roughly equivalent. This explains why Harsanyi appeals to HL to argue for average utilitarianism in his paper 'Cardinal Welfare, Individualistic Ethics, and Interpersonal Comparisons of Utility' (1955).
Harsanyi's implicit argument can be summarized as follows:
7. Equivalence between ROI and difference principle
We've seen how to apply the general equivalence, under certain conditions, between aggregative principles and utilitarian principles, e.g. between LELO and longtermist total utilitarianism, or between HL and average utilitarianism. This section will apply that theorem to the monad P+f, which is used to formalise Rawls' Original Position (ROI). We will see that ROI is equivalent to the difference principle.
7.1. Semilatticial rationality of personal utility?
The real numbers admit a fusion operator in the obvious way, i.e., there exists a function min:P+f(R)→R where min({r1,…,rn}) is the largest r∗∈R satisfying r∗≤ri for each i=1,…,n. This is simply the well-known minimisation operator, which sends a nonempty finite subset of the real values to their minimum.
Let's unpack RPU, which formally states that v∘⨁=min∘vP+f. In other words, for any nonempty finite subset of personal outcome {p1,…,pn}, we have equality between v∘⨁({p1,…,pn}) and min{v(p1),…,v(pn)}. Informally, the personal utility of a fused outcome is the minimum of the personal utilities of the outcomes being fused. This 'semilatticial' rationality condition constrains how humans must value the fusion of different personal outcomes.
In the previous article, we saw that the fusion operator ⨁:P+f(P)→P can be equivalently presented by a single binary operator ⊕, with the intended interpretation p⊕p′=⨁({p,p′}). We can restate the RPU condition in terms of ⊕ with the single equation v(p⊕p′)=min{v(p),v(p′)}.
How realistic is the RPU condition? That is, supposing humans do maximise a personal utility function, how semilatticially rational is it? I think this condition is approximately true, but unrealistic in several ways. I'll assume here that p⊕p′ is interpreted as a Knightian uncertainty between facing p and facing p′. Then semilatticial rationality requires that humans are pessimistic, i.e. they value the disjunction of different outcomes no greater than the worst, as if the alternative will be selected by an adversary.
Firstly, RPU rules out valuing determinacy or indeterminacy. Informally, a disjunction can't be valued below each determinate outcome, nor higher. But perhaps this is an unrealistic assumption: humans might value the disjunction p1⊕p2 lower than either p1 or p2 because humans inherently value determinacy in this case. Or humans might value the disjunction p1⊕p2 lower than either p1 or p2 because humans inherently value indeterminacy in this case. But RPU precludes such values.
Second, RPU rules out non-pessimistic considerations. Informally, adding additional possibilities can never increase the value of a disjunction. But this is an unrealistic assumption about human values, as illustrated in the following examples. If pokay is a typical comfortable life and pbad is a life of horrific torture, then humans may value the outcome pokay higher than pbad and value the outcome pbad⊕pokay higher than the outcome pbad. In particular, the outcome pbad⊕pokay gives a possibility of being fine, while pbad is certain to result in torture. However, RPU precludes such values.
Overall, I think human values approximate semilatticial rationality. Indeed, suppose you face genuine Knightian uncertainty between a set of possibilities p1,…,pn, with personal utilities v(p1),…,v(pn) respectively. What's the ex-ante value? There's not much to done to construct the ex-ante personal value for p1⊕⋯⊕pn other than minimising over the possibilities, i.e. v(p1⊕⋯⊕pn) equals min({v(p1),…,v(pn)}). The only alternative is to deny that the ex-ante value of p1⊕⋯⊕pn depends solely on the ex-post value of the constituent possibilities, or else employ a different semilattice on R.[7] Moreover, Wald's maximin model, and robust optimisation more generally, are popular principles of decision-making. These principles involve maximising a semilatticially rational utility function. Hence, I accept semilattical rationality of human values, at least when the fusion p1⊕p2 is interpreted as Knightian uncertainty between p1 and p2, a mode of ignorance which is rarely encountered.[8]
7.2. Social utility minimises personal utility?
Now let's unpack SUAPU, which formally states that u(s)=min((v∘γ(−,s))P+f(A)). In other words, the social utility function to the minimum of personal utilities over the individuals in the distinguished nonempty subset representing the population. That is, if A={i1,…,in}∈P+f(I) is a nonempty subset of individuals representing the entire population impartially, then for any social outcome s∈S, the social utility u(s) is given by min({v(γ(i1,s)),…,v(γ(in,s)}).
How realistic is the SUAPU condition? The answer depends on one's axiological theory. Indeed, SUAPU is a statement of Rawls' difference principle. This is a strong assumption that precludes a social utility function from exhibiting certain properties, analogous to how RPU constrains the personal utility function. Specifically, SUAPU precludes social utility functions with the following features:
Nonetheless, I think that HMPU, RPU, and SUAPU are somewhat reasonable approximations, though probably the least plausible in the ROI context than the LELO context or HL context. To the extent that these assumptions hold, Rawls' Original Position (ROI) and his difference principle will be roughly equivalent. This explains why Rawls appeals to ROI to argue for difference utilitarianism in his book "The Theory of Justice" (1973).
Rawls' implicit argument can be summarized as follows:
8. Conclusion
To summarise, I first formalised social choice principles using functions of type-signature (X→S)→P(X). This allowed me to define the utilitarian principle corresponding to a given social utility function, and the aggregative principles corresponding to a given social zeta function. As discussed in my previous article, this social zeta function maps a social outcome to the aggregated personal outcomes of each individual. Using the formalism, I proved that, under three natural conditions, the aggregative principle is mathematically equivalent a corresponding utilitarian principle. Because these conditions are approximately true, aggregativism approximates utilitarianism. even though aggregativism avoids the theoretical pitfalls of utilitarianism, we should nonetheless expect aggregativism to generate roughly-utilitarian recommendations in practical social contexts, and thereby retain the most appealing insights from utilitarianism. Moreover, this explains why MacAskill, Harsanyi, and Rawls each appeal to aggregative principles to defend their respective utilitarian principles.
In this next article, I will enumerate the theoretical pitfalls that face utilitarianism, and how aggregativism overcomes them.
See Appraising aggregativism and utilitarianism for a thorough defence.
In fact, the function mapping each option x∈X to the principle Ψ:f↦{x} is a canonical embedding of the space of options into the space of social choice principle.
See Game Theory without Argmax.
The aggregative principle is f↦Π(ζ∘f), where f:X→S is a social context, Π is the human model, and ζ:S→P is the social zeta function. This means a social planner should choose an option if a self-interested human would choose the associated personal outcome. By HMPU, Π has the form f↦argmaxX(v∘f), where v:P→R is the personal utility function. This means a self-interested human will choose an option that maximizes personal utility. Hence, aggregativism is the principle f↦argmaxX(v∘ζ∘f). Intuitively, this means a social planner should choose an option which maximizes the personal utility of the associated personal outcome.
The social zeta function ζ:S→P is defined by ζ(s):=α(γ(−,s)M(π)), where α:M(P)→P is the aggregation function for personal outcomes, γ:I×S→P assigns personal outcomes to individuals, and π∈M(I) is the distinguished collection representing the population. Intuitively, this means the personal outcome associated to a social outcome is the aggregate of the personal outcomes across all individuals in society.
Now, RPU asserts that v∘α=β∘vM, i.e. that the personal utility of the aggregate of personal outcomes is the aggregate of personal utilities of each outcome. Given ζ(s):=α(γ(−,s)M(π)), we obtain (v∘ζ)(s)=(β∘vM)(γ(−,s)M(π)). Intuitively, this means the personal utility of the personal outcome associated to a social outcome is the aggregate of the personal utilities of the personal outcomes faced by each individual in society.
Now, SUAPU asserts that u(s)=β((v∘γ)(−,s)M(π)), where u:S→R is the social utility function, β:M(R)→R is the aggregation function for real numbers, v:P→R is the personal utility function, γ:I×S→P assigns personal outcomes to individuals, and π∈M(I) is the distinguished collection representing the population. Intuitively, this means the social utility of a social outcome is the aggregate of the personal utilities of the personal outcomes faced by each individual in society.
This entails that u=v∘ζ. To see this, note that the right-hand-sides of the equations (v∘ζ)(s)=(β∘vM)(γ(−,s)M(π)) and u(s)=β((v∘γ)(−,s)M(π)) are identical: (β∘vM)(γ(−,s)M(π))=β((v∘γ)(−,s)M(π)). Indeed, this follows from the functorality of the lifting operator. Therefore, v∘ζ(s)=u(s) for all s∈S. Intuitively, this means the social utility of a social outcome is the personal utility of its associated personal outcome.
Hence, the aggregative principle is f↦argmaxX(u∘f). To see this, note thatargmaxX(v∘ζ∘f)=argmaxX(u∘f) because v∘ζ=u. Intuitively, this means a social planner following the aggregative principle should choose an option which maximizes the social utility of the resulting social outcome. The utilitarian principle is f↦argmaxX(u∘f). Hence the aggregative principle is equivalent to the utilitarian principle conditional on HMPU, RPU, and SUAPU. □
Of course, whether these particular cases violate RPU depends on which function Π:(X→P)→P+f(X) models the self-interested human, and which personal utility function v:P→R is used to characterise Π. Nonetheless, I think that any reasonable Π or v:P→R will contain both examples of novelty value and of consistency value.
We might also ask: are human values convexly rational with respect to other convex algebras on personal outcomes? Recall that, in my previous article, we examined a novel interpretation of p+λp′ as the direct interpolation in some high-dimensional vector space Rd. To obtain semantically meaningful vector representations of personal outcomes, we might leverage the activation space of a large language model like GPT-3. The interpolation p+λp′ of two vector representations p,p′∈Rd is simplyλ⋅p+(1−λ)⋅p′. Under this interpretation of +λ, the RPU condition says that personal utility is a linear probe. Formally, RPU requires the personal utility function v:Rd→R to satisfy the equation v(λ⋅p+(1−λ)⋅p′)=λ⋅v(p)+(1−λ)⋅v(p′) for all vectors p,p∈Rd and interpolation weights λ∈(0,1). Whether RPU holds in this setting depends on the specific vector representation of outcomes.
The real numbers admit another fusion operator, max:P+f(R)→R, which we could consider. But the semilattice (R,max) will generate a condition of semilatticial rational which is even less plausible than that generated by the semilattice (R,min). Namely, it requires v(p1⊕p2)=max{v(p1),v(p2)}, e.g. humans would value Knightian uncertainty between horrific torture and a comfortable life no higher than certainty in a comfortable life.
In my previous article, we examined a conjunctive interpretation of the fusion of personal outcomes, in contrast to Rawls' disjunctive interpretation. In particular, if p1 and p2 are personal outcomes then p1⊕p2 is the personal outcome of facing p1 and p2 simultaneously. How should we understand semilatticial rationality, which formally states that for any nonempty finite subset of personal outcome {p1,…,pn}, we have equality between v∘⨁({p1,…,pn}) and min{v(p1),…,v(pn)}? Under this fusion operator, semilatticial rationality requires that humans are "glass half-empty". Informally, the value of facing outcomes p1,…,pn simultaneously is no greater than the value of the worst constituent outcome. That is, v(p1⊕⋯⊕pn)=min{v(p1),…,v(pn)}.
Here's how this rationality condition might arise naturally: Imagine a set of "catastrophes", such as being bored, being cold, being dead. Each catastrophe is represented with a personal outcome p and a value v(p)∈R. For example, v(bored)=−5, v(cold)=−10, and v(dead)=−1000. Moreover, the utility of a complex personal outcome, such as being bored and cold simultaneously, is determined by the worst catastrophe. That is, v(bored⊕cold)=−10. It implies that facing multiple catastrophes, which are equally disastrous, is no worse than facing only one such catastrophe, i.e. if v(hungry)=−10 then v(cold⊕hungry)=−10.