In the previous post, I presented the concept of a maximal lottery-lottery as follows:
We start with a finite set of candidates and a distribution on utility functions on . Given lottery-lotteries , we say that dominates if . A maximal lottery-lottery is an such that for all other , dominates .
Conjecture: For any and , there always exists a maximal lottery-lottery.
We discussed how we can construct a powerful and elegant new voting system, if only we can guarantee that a maximal lottery-lottery exists. However it is an open problem whether or not they exist. In this post, I will discuss a few different angles of attacking this conjecture.
An Infinite Game
First, consider the following partial result:
Theorem: Let be any nonempty finite subset of . Then, there exists an such that for all other , dominates .
Proof: Consider the following symmetric two-player zero-sum game between Alice and Bob: Alice chooses an , and Bob chooses a . Alice gets utility equal to , and Bob gets the negative of this utility. This game must have a Nash equilibrium, and in this Nash equilibrium, Alice must play a mixed strategy . Since this is a Nash equilibrium in a symmetric zero-sum game, Alice must get at least 0 utility in expectation, no matter what Bob plays. Alice getting at least 0 utility in expectation is exactly saying that dominates .
This proof is valid (you can check the details), so why doesn't it generalize to infinite subsets , and in particular ? The problem is that games with infinitely many options need not have Nash equilibria. For example consider the game where Alice and Bob each choose a natural number, and whoever chooses the larger number wins.
Fortunately, often in cases like this, you can get past finiteness assumptions by instead using compactness assumptions, and is compact. Indeed, games with a compact option space and continuous utility functions always have mixed strategy equilibria.
Unfortunately, the game we constructed does not have a continuous utility function! This is easy to see, because if assigns probability 1 to a single utility function , if , then Alice wins and gets utility 1, but as we continuously move around , we can discontinuously jump to the game being a tie, and then to Bob winning.
Fortunately, there are a bunch of papers analyzing situations like this. Discontinuities like this show up a bunch in auctions, which also have an infinite compact option space and discontinuous utilities, resulting from discontinuities in who wins the auction.
Unfortunately, Sam Eisenstat and I spent a while looking through a bunch of papers like this, and failed to find anything that we could apply. Several things looked like they might be able to work, but everything we tried ultimately turned out to be a dead end.
A Limit of Finite Games
The above partial result is actually pretty practically useful on its own. Even if maximal lottery-lotteries do not exist, you can always just take to be the set of all lotteries that assign each candidate a probability that is a multiple of for some large natural number . Because of this, practically speaking, we can use the maximal lottery-lotteries voting system today, without ever proving the conjecture!
Further, we can consider what happens as we send to infinity. If there is an odd number of voters that each have a utility function in general position, we will always get a unique distribution on that dominates all other distributions on , and we can look at what happens to as goes to infinity. Hopefully this converges, but even if it doesn't, we can take any limit point, and ask whether that limit point is a maximal lottery-lottery. My guess is that the sequence does in fact converge, and that the (unique) limit point is a maximal lottery-lottery, but I have not had any luck yet proving either of these.
However, this enables us to get some empirical results! We can just compute some of these finite approximations. Even with three candidates, we have to solve a game with different options, and I have not yet optimized it beyond using an off the shelf linear programming solver in Haskell, running for a few minutes, so the highest I got was . (I am sure we can do much better.)
I took three candidates and three voters, and gave each voter a utility function in general position. Here is a picture of . The blank space means that 0 (or near 0) utility is put there. The numbers show the negative of the natural logarithm of the probability that that point is chosen. (Remember, this is a distribution on distributions on a three element set, or equivalently a distribution on the triangle.)
______________________4.4.5.5.5.3.4.4.5.5.5.5.4.5.5.__________
__________________________7.____________4.6.5.8.5.4.________
__________________4.5.5.______6.5.__5.__5.9.5.4.4.4.______
________________4.5.5.4.______5.____5.__5.__4.4.5.5.____
______________5.5.5.5.7.____5.4.5.7.5.108.6.5.4.4.4.5.
______________5.________________________5.4.5.____7.
____________7.5.4.5.____________6.5.6.__4.__5.6.__
______5.4.5.8.__5.8.____6.5.__6.__7.__6.4.6.4.5.
____6.7.5.5.5.5.__5.6.8.6.5.7.5.____6.5.5.__4.
6.6.5.4.5.4.4.5.6.__6.__6.5.__6.____5.6.7.4.
__5.6.5.________5.__5.6.________5.6.8.5.5.
4.5.5.__4.6.__5.6.7.5.6.6.____6.4.5.7.4.
5.4.5.5.5.__5.5.5.7.5.________5.6.4.4.
6.5.5.5.6.__5.__5.5.6.____7.8.4.7.7.
5.5.5.5.5.____5.5.5.5.__5.6.7.__4.
7.4.5.6.__6.5.5.4.6.____4.4.5.5.
4.__8.5.__5.7.__5.5.5.__6.____
5.6.6.6.5.6.5.______________
4.________________________
________________________
______________________
____________________
__________________
________________
______________
____________
__________
________
______
____
__
for the same candidates and voters looks similar, but with less detail, and some minor changes. This is a lot of data, but since we are going to use it to sample and then sample from the sample, for the purpose of maximal lottery-lotteries, we only really care about the probability each candidate is elected. For the above example, we get (0.262, 0.415, 0.323) for and (0.263,0.417,0.320) for , which seems like some slight empirical evidence towards converging. Obviously there is a lot of room for a much more careful and larger experimental analysis.
Fractals!?!
I believe that maximal lottery-lotteries exist, and are the limit of the above finite approximations, and (when they don't put all probability mass on the distribution that puts all probability mass on a single candidate) usually have some fractal structure. I think the complexity of the above picture is evidence of this, but I also have some mathematical intuition for why this makes sense (which feels hard to put into words):
Being a Nash equilibrium involves equalizing a bunch of constraints. These constraints are combined with the fact that the distribution is supported entirely within the triangle, and so where probability mass would naturally want to be placed outside the triangle, it gets cut off on the boundary. When it is cut off like this, something has to change inside the triangle to fix new constraints that become broken, and this causes the constraints to sort of reflect off the boundary of the triangle, somewhat like ripples. (Some of this intuition is coming from seeing similar things happen when studying nim.) Sorry that I don't have a more articulate description, but the point is, I think we are getting fractals.
Unfortunately, I think this is going to make it harder to show that maximal lottery-lotteries exist, since we will likely not have simple closed forms of what the maximal lottery-lotteries are.
A Generalization of Colonel Blotto
Another way to look at maximal lottery-lotteries is as a generalization of the Colonel Blotto game.
In the the continuous Colonel Blotto game, there are two opposing officers that each have some resources, that they can distribute between some battlefields. We will assume that they each have 1 total unit of resources. Whoever sends the most resources to any given battlefield wins that battlefield. (If they send the same amount, they each win half the battlefield.) Each officer gets utility equal to the number of battlefields that they win. This game and its variants have been analyzed quite a bit over the last century.
Finding maximal lotteries is like solving a variant of the Colonel Blotto game. In the original game, you choose how to distribute resources from the simplex where each battlefield gets non-negative resources, and the total of all the resources is 1, In this variant, you instead have some other convex polytope representing the allowable ways to distribute resources.
To reduce finding maximal lottery-lotteries to a Colonel Blotto game, we consider each voter to be a battlefield. (We will only discuss the case where there is a finite set of equal weight voters here.) Each candidate can be thought of as assigning a utility to each voter, which we can also interpret as sending a number of resources equal to that utility. Thus, the polytope of allowable ways to send resources to battlefields is just the convex hull of these points that come from the candidates. Each point in this convex hull corresponds to (at least) one distribution on candidates, and a solution to this Colonel Blotto game corresponds to a maximal lottery-lottery.
This old RAND paper analyzing Colonel Blotto, seems like some further evidence that we might be looking at fractals.
A Call for Help
I think that proving the existence of maximal lottery-lotteries would actually be a substantial advance for voting theory. (Although it might take some work to get people interested in something that challenges this many of the assumptions of the field.) However, I am not intending to work on this anymore. If you want to take a shot at proving they exist, and have questions, get in touch with me.
Without a proof of existence, I am probably not going to bother writing this up as more than these blogposts, but if anyone manages to prove it, I would be happy to write a paper together. Also if anyone feels very excited about working on this and getting it published in spite of not having a proof, I could maybe be up for collaborating on something like that, especially if someone collects a bunch of experimental evidence and better fractal pictures and is willing to do a lot of the writing.
I have a reduction of this problem to a (hopefully) simpler problem. First up, establish the notation used.
[n] refers to the set {1...n}. n is the number of candidates. Use C as an abbreviation for the space Δ[n], it's the space of probability distributions over the candidates. View C as embedded in Rn−1, and set the origin at the center of C.
At this point, we can note that we can biject the following:
1: Functions of type [n]→[0,1]
2: Affine functions of type C→[0,1]
3: Functions of the form λx.⟨a,x⟩+c, where x,a∈Rn−1, and and c∈R, and everything's suitably set so that these functions are bounded in [0,1] over C. (basically, we extend our affine function to the entire space with the Hahn-Banach theorem, and use that every affine function can be written as a linear function plus a constant) We can reexpress our distribution V over utility functions as a distribution over these normal vectors a.
Now, we can reexpress the conjecture as follows. Is it the case that there exists a μ:ΔC s.t. for all ν:ΔC, we have
Ex,y,a∼μ×ν×V[sgn(⟨a,x−y⟩)]≥0Where sgn is the function that's -1 if the quantity is negative, 0 if 0, and 1 if the quantity is positive. To see the equivalence to the original formulation, we can rewrite things as
Ex,y,a∼μ×ν×V[1⟨a,x⟩>⟨a,y⟩−1⟨a,y⟩>⟨a,x⟩]≥0Where the bold 1 is an indicator function. And split up the expectation and realize that this is a probability, so we get
Px,y,a∼μ×ν×V[⟨a,x⟩>⟨a,y⟩]−Px,y,a∼μ×ν×V[⟨a,y⟩>⟨a,x⟩]≥0And this then rephrases as
Px,y,U∼μ×ν×V[U(x)>U(y)]≥Px,y,U∼μ×ν×V[U(y)>U(x)]Which was the original formulation of the problem.
Abbreviating the function Ex,y,a∼μ×ν×V[sgn(⟨a,x−y⟩)] as f(μ,ν), then a necessary condition to have a μ:ΔC that dominates everything is that
supμ∈ΔCinfν∈ΔCf(μ,ν)≥0If you have this property, then you might not necessarily have an optimal μ that dominates everything, but there are μ that get a worst-case expectation arbitrarily close to 0. Namely, even if the worst possible ν is selected, then the violation of the defining domination inequality happens with arbitrarily small magnitude. There might not be an optimal lottery-lottery, but there are lottery-lotteries arbitrarily close to optimal where this closeness-to-optimality is uniform over every foe. Which seems good enough to me. So I'll be focused on proving this slightly easier statement and glossing over the subtle distinction between that, and the existence of truly optimal lottery-lotteries.
As it turns out, this slightly easier statement (that sup inf is 0 or higher) can be outright proven assuming the following conjecture.
Stably-Good-Response Conjecture: For every ν:ΔC, and ϵ>0, there exists a μ:ΔC and a δ>0 s.t.
infν′:d(ν,ν′)<δf(μ,ν′)>−ϵPretty much, for any desired level of suckage and any foe ν, there's a probability distribution μ you can pick which isn't just a good response (this always exists, just pick ν itself), but a stably good response, in the sense that there's some nonzero level of perturbation to the foe where μ remains a good response no matter how the foe is perturbed.
Theorem 1 Assuming the Stably-Good-Response Conjecture, supμinfνf(μ,ν)≥0.
I'll derive a contradiction from the assumption that 0>supμinfνf(μ,ν). Accordingly, assume the strict inequality.
In such a case, there is some ϵ s.t. 0>−ϵ>supμinfνf(μ,ν). Let the set Aμ:={ν|f(μ,ν)>−ϵ}. Now, every ν lies in the interior of Aμ for some μ, by the Stably-Good-Response Conjecture. Since ΔC is a compact set, we can isolate a finite subcover and get some finite set M of probability distributions μ s.t. ∀ν∃μ∈M:f(μ,ν)>−ϵ.
Now, let the set Bν:={μ∈c.h(M)|f(μ,ν)<−ϵ}. Since −ϵ>supμinfνf(μ,ν), this family of sets manages to cover all of c.h(M) (convex hull of our finite set.) Further, for any fixed ν, f(μ,ν) is a continuous function c.h(M)→R (a bit nonobvious, but true nontheless because there's only finitely many vertices to worry about). Due to continuity, all the sets Bν will be open. Since we have an open cover of c.h(M), which is a finite simplex (and thus compact), we can isolate a finite subcover, to get a finite set N of ν s.t. ∀μ∈c.h(M)∃ν∈N:f(μ,ν)<−ϵ. And now we can go
−ϵ>maxμ∈c.h(M)minν∈Nf(μ,ν)≥maxμ∈c.h(M)minν∈c.h(N)f(μ,ν)=minν∈c.h(N)maxμ∈c.h(M)f(μ,ν)≥minν∈c.h(N)maxμ∈Mf(μ,ν)>−ϵThe first strict inequality was from how all μ∈c.h(M) had some ν∈N which made f(μ,ν) get a bad score. The ≥ was from expanding the set of options. The = was from how f is a continuous linear function when restricted to c.h(M)×c.h(N), both of which are compact convex sets, so the minimax theorem can be applied. Then the next ≥ was from restricting the set of options, and the > was from how every ν∈ΔC had some μ∈M that'd make f(μ,ν) get a good score, by construction of M (and compactness to make the inequality a strict one).
But wait, we just showed −ϵ>−ϵ, that's a contradiction. Therefore, our original assumption must have been wrong. Said original assumption was that 0>supμinfνf(μ,ν), so negating it, we've proved that
supμinfνf(μ,ν)≥0As desired.