I guess I don't see what's so wrong with "you can't do all the actions in your action space," which just seems like another way of saying "this action space isn't the right formalization of the problem."
I disagree. Sometimes your entire payoffs also change when you change your action space (in the informal description of the problem). That is the point of the last example, where precommitment changes the possible payoffs, not only restricts the action space.
I remain confused about Newcomb-style problems in general, and counterfactual mugging in particular. They seem to typically be fairly trivially self-contradictory when you look at them from the lens of computability theory.
Omega, a perfect predictor, flips a coin. If it comes up tails Omega asks you for $100. If it comes up heads, Omega pays you $10,000 if it predicts that you would have paid if it had come up tails.
Either the agent must be running a decidable algorithm, or there is no such Omega satisfying the requirements[1].
In the former case, there is no contradiction[2].
In the latter case, the problem is a moot point.
Suppose you had an algorithm[3] for Omega. Then I feed your Omega an agent that does this:
Unfortunately, it's fairly simple to show that Omega cannot avoid a contraction here. (About the best that Omega can do is loop forever / fail to halt... which still doesn't fit the problem.)
This is a contradiction, so Omega cannot exist.
Or Omega is super-Turing, in which case all bets are off.
Jensen's inequality as used in this context only applies if the agent can do unbounded computation, for one. (Rough sketch of an example with finite computation: suppose I have an agent that does at most C operations. If I ask it for the best outcome over [1] (that is, one outcome with a value of 1), the answer is 1, fairly straightforwardly. If I instead ask it for the best outcome over C outcomes each with a value of and one outcome with a value of 1, the agent has an expected value of no more than , as it doesn't have enough operations to scan all possible outcomes looking for the good one.)
Strictly speaking, Omega is non-deterministic as stated, but you can recover determinism by passing the coinflip result into Omega as an argument.
This is non-trivial, but doable. It requires a vaguely quine-like construct.
The usual formulations typically have Omega being "bigger" than your agent, such that Omega can predict your agent, but your agent cannot predict Omega. This is generally taken to be unproblematic because physical agents are always bounded by time + memory constraints, which makes them analogous to finite-state machines rather than full Turing machines. (If unbounded computation were permitted, on the other hand, then you would indeed need a super-Turing version of Omega to pull off the trick.)
...in which case much of game theory does not apply and the paradoxes, by and large, aren't actually paradoxical, no?
Paradoxical decision problems are paradoxical in the colloquial sense (such as Hilbert's hotel or Bertrand's paradox), not the literal sense (such as "this sentence is false"). Paradoxicality is in the eye of the beholder. Some people think Newcomb's problem is paradoxical, some don't. I agree with you and don't find it paradoxical.
In that case why are people spending so much effort on it[1]?
Ditto, why is there so much argumentation based around applying game theory to Newcomb's problem[2] (or variants) when much of game theory does not apply to it?
*****
Paradoxical decision problems are paradoxical in the colloquial sense (such as Hilbert's hotel or Bertrand's paradox), not the literal sense (such as "this sentence is false").
I think that there is a lot of effort being wasted due to not clearly distinguishing counterintuative results and self-contradictory results in general, and this is a prime example.
The LessWrong "Newcomb's Problem" tag has 41 entries with a combined total of over 1,000 comments, for instance.
See also any argument[3] that says that X is invalid because it provides 'the wrong solution' to Newcomb's Problem.
Such as "Evidential decision theories are inadequate rational decision theories. For either they provide wrong solutions to Newcomb's problem [...]", which is a direct quote from the overview of Newcomb's problem linked in the description of the tag 'Newcomb's Problem' on this site.
There are a couple of paradoxes in decision theory, typically involving beings making perfect predictions, copies of yourself, and similar shenanigans. If you're unfamiliar with all of this, take a look at Newcomb's problem. Attempted resolutions of these paradoxes usually involve either the construction of complex decision theories (such as FDT ) or copious hand-waving and informal arguments (the philosophy literature on the subject).
This post has two messages.
I have three guiding principles.
Ordinary decision theory
We're dealing with a rational actor, presumably you, an action space A, and a set of random utilities Ua,a∈A. I take the action set seriously – it is possible to choose any of the actions in the set. We know the distribution of Ua for every a∈A. Let's say that an acta∗ is rational if it maximizes the expected utility EUa,i.e.,a∗∈argmaxaEUa.We'll assume that argmaxaEUa exists. (Defining a rational act is hard when a maximum does not exist. For instance when the act a gives deterministic utility Ua=∑ai=1i−2 when A=N, and even harder when the utility function is discontinuous at infinity. See "Satan's apple".)
You may want to pause for a second or two to figure out if you agree with this notion of a rational act. Is it the usual definition? Is it strange in any way?
Let's take a look at a simple example. Suppose you're given the option to flip either (A) a biased coin with probability 0.7 for heads, (B) a biased coin with probability 0.6 for heads. In either case, you can bet $1 and on either tails or heads, winning $2 if you're right and losing your $1 if you're wrong. Your action space is
A={(coin A,heads),(coin B,heads),don't play.(coin A,tails),(coin B,tails),}The expected utilities are
EUa={0.4,0.2,0.−0.4,−0.2,}.Unsurprisingly, choosing coin A and heads is the rational act. So far, so good.
Ordinary decision theory wants you to choose between different random variables Ua, not just between different acts a and a fixed source of randomness U. Usually, you would be restricted to acting within the context of a single random variable, as in the von Neumann–Morgenstern utility theorem, where we maximize Eu(a,X). This distinction usually doesn't matter mathematically, as it's mathematically trivial to transform a EUa problem to a Eu(a,X) problem. It does matter for interpretation though!
Now take note of the following:
Not every problem can be formulated as an ordinary decision problem, either because they are inherently game-theoretical (counterfactual mugging, Psykosh's non-anthropic problem) or because they encode more than a single ordinary decision problem (opaque Newcomb).
Newcomb's problem and incompatibility between descriptions
Consider the following decision problem. You have two acts, a=1 and a=2. The utilities Ua are deterministic and equal to U1=$1,000,000 and U2=$10,000. Which act do you choose? Using the definition of a rational act above, you choose a=1. This is utterly uncontroversial, isn't it? It turns out it isn't, as adding mathematically irrelevant information to the problem makes the majority of philosophers choose a=2!
Now, let's add Newcomb's description to the problem so that U1 corresponds to one-boxing and U2 to two-boxing. The mathematical formulation of Newcomb's problem coincides with the setup above: “You have two acts, a=1 and a=2. The utilities Ua are deterministic, and are equal to U1=$1,000,000 and U2=$10,000.” This information was sufficient for U1 to be rational – adding more information can't change that.
Newcomb's problem feels like a significant problem because it involves a clash between intuitions; on one hand, you wish to reason causally, which leads to the incorrect application of dominance analysis. On the other hand, you want to do the rational act.
I can see two resolutions to these
My impression is that (b) is the favored resolution to Newcomb's problem on LessWrong. While that's an OK resolution, I dislike it. It (informally!) implies that you cannot choose any action from your action set. As I said, I take action sets seriously (and so should you). The resolution cannot be generalized to situations with proper randomness involved, as omega cannot predict your choice if you have access to a source of randomness that he doesn't, but it's mathematically 100% OK to formulate Newcomb's problem with mixed strategies instead of pure ones. Resolution (a) is better. No one is claiming that Newcomb's problem is possible. It's not possible if there's a true source of randomness that you can observe but Omega can't. The trick is to see through your intuitions, write the problem down, and solve it.
(Newcomb's problem is sometimes used as a short morality tale, as in "LOL! Don't you know the human brain is a machine and can be predicted?!?" Some people have not gotten that message, but I don't think the confusion around the problem is primarily about the possibility of powerful predictors.)
Is causal decision theory useful? Yes, as a device to construct decision-theoretic problems, but not as a solution method. Writing down small causal graphs helps you to think clearer, but can't ever change the fundamental idea of choosing the act that maximizes expected utility. Causal decision theory is ordinary decision theory – but only if the causal description matches the mathematical description of the problem. In these cases, causal decision theory will add irrelevant, but not misleading, information.
The same cannot be said for evidential decision theory, whose added information is actively misleading (think about, e.g., the smoking lesion).
I insist on using the notation a∗∈argmaxaEUa. instead of Eu(a,X) as it clarifies what we are dealing with. You could Newcomb's problem using the notation Eu(a,X), with a complicated enough utility function u and outcome variable X. To write down Newcomb's problem using u(a,X), we could, for instance, define X=0 and u(a,X)=u(a) equal to $1,000,000 if a=1 and $10,000 otherwise. That's pretty easy, but notice that X is not equal to the amount of money left in the boxes, which we would expect. Now, take a look at a slightly modified Newcomb's problem, where Omega predicts your action with 90% accuracy. Then
p(u1=$1,000,000)=0.9,p(u1=$0)=0.1,p(u2=$10,000)=0.9,p(u2=$1,010,000)=0.9.To write down u(a,X), we can let X be the random variable that equals 1 (Omega predicted correctly) with probability 0.9 and 0 with probability 0.1. Then we define
u(a,1)={$1,000,000a=1$10,000a=2,u(a,2)={$0a=1,$1,010,000a=2.This formulation is unsatisfying, as X isn't the natural outcome (did Omega put money in the opaque box?). That's why I advocate for Ua.
Counterfactual game-theory
It's a consequence of Jensen's inequality that, for any random variable Z
maxa∈AE[Ua]≤E(maxa∈AE[Ua∣Z]).In other words, you cannot expect to lose expected utility by gaining information. A couple of decision-theoretic paradoxes ostensibly violate Jensen's inequality. That is a big deal since it implies that we are either (i) dealing with an inconsistent decision problem, or (ii) the problem cannot be formulated as an ordinary decision problem. After all, Jensen's inequality is a theorem.
Recall counterfactual mugging
Here's an intuitive analysis of this problem. If you could force yourself to the strategy of paying when X=heads, then
EUpay=12E[Upay∣X=heads)]+12E[Upay∣X=tails],=12[10000−100]=4950.On the other hand, refusing to pay is the optimal action, conditioned on X=tails, which yields a utility of 0. But using this strategy makes
4950=maxa∈AEUa>E(maxa∈AE[Ua∣X])=0,which appears to imply that you lose money, in expectation, by gaining information. As Jensen's inequality is violated, counterfactual mugging isn't an ordinary decision-theoretic problem. What is it then?
The most natural solution (one-shot) solution to this problem uses game theory, by letting you and your counterfactual self be the two players. The Nash equilibrium is “not pay”, so it does not optimize expected utility across both players (i.e., it's not Pareto optimal), but that's not exactly uncommon in game theory. Just as defection is rational in the Prisoner's dilemma, I see no problem with not paying in counterfactual mugging; your counterfactual self isn't you. (The counterfactual player does not make an action in counterfactual mugging, but he does in the counterfactual Prisoner's dilemma.)
Psy-kosh's non-anthropic problem can also be analyzed with a game-theoretic lens to explain why coordinating has the highest payoff and is thus rational, provided it is an option. There is simply no reason to believe game-theoretic reasoning to be Pareto-optimal, so it should be avoided if possible.
I haven't seen any detailed analyses of Psy-kosh's problem and similar "counterfactual" game-theoretic problems yet, and I don't believe they are fully solved. But I do believe counterfactual game-theory forms an interesting subset of unordinary decision problems. Developing a theory for counterfactual game theory could be worthwhile. Please tell me if you want to work on it (and like the approach of this post).
When the size of the action space affects Ua
Finally, we have paradoxical decision problems about precommitment. In transparent Omega, both boxes are transparent, the $10,000 box and the $1,000,000 box. Omega perfectly predicts your choice and puts $1,000,000 in the transparent box if he predicts one-boxing.
If you could remove the strategy a2 from the action set, then Omega would correctly predict a2, and put $1,000,000 in the box, so that maxa∈AEUa=Ua1=$1,000,000. Since A={a2}, the decision problem is trivial.
Asmall={a1},Usmalla1=$1,000,000,A={a1,a2},Ua1=$0,Ua2=$10,000.Now consider the case when the action space isn't reduced, i.e., A={a1,a2}. Since Omega predicts perfectly, he can't possibly place $1,000,000 in the box, as it would make it possible for you to two-box. (Mere possibility is enough. If you're given a choice, you're given a choice.). Thus we're left with two separate decision problems:
You can combine both problems into one:
Here you're asked to choose between the small decision problem, yielding a guaranteed $1,000,000, and the big decision problem, yielding a guaranteed $10,000. In conclusion, you should make your decision straight away if the fee is smaller $990,000.
A standard precommitment version of Viewcomb won't work. If you write a contract pledging to pay $11,000 when two-boxing, it's always rational to one-box. But the (irrational) possibility of one-boxing is still there, and Omega, being a perfect predictor, cannot allow you to make that choice. Thus you're stuck in the large decision problem.