The following may well be the most controversial dilemma in the history of decision theory:
A superintelligence from another galaxy, whom we shall call Omega, comes to Earth and sets about playing a strange little game. In this game, Omega selects a human being, sets down two boxes in front of them, and flies away.
Box A is transparent and contains a thousand dollars.
Box B is opaque, and contains either a million dollars, or nothing.You can take both boxes, or take only box B.
And the twist is that Omega has put a million dollars in box B iff Omega has predicted that you will take only box B.
Omega has been correct on each of 100 observed occasions so far - everyone who took both boxes has found box B empty and received only a thousand dollars; everyone who took only box B has found B containing a million dollars. (We assume that box A vanishes in a puff of smoke if you take only box B; no one else can take box A afterward.)
Before you make your choice, Omega has flown off and moved on to its next game. Box B is already empty or already full.
Omega drops two boxes on the ground in front of you and flies off.
Do you take both boxes, or only box B?
And the standard philosophical conversation runs thusly:
One-boxer: "I take only box B, of course. I'd rather have a million than a thousand."
Two-boxer: "Omega has already left. Either box B is already full or already empty. If box B is already empty, then taking both boxes nets me $1000, taking only box B nets me $0. If box B is already full, then taking both boxes nets $1,001,000, taking only box B nets $1,000,000. In either case I do better by taking both boxes, and worse by leaving a thousand dollars on the table - so I will be rational, and take both boxes."
One-boxer: "If you're so rational, why ain'cha rich?"
Two-boxer: "It's not my fault Omega chooses to reward only people with irrational dispositions, but it's already too late for me to do anything about that."
There is a large literature on the topic of Newcomblike problems - especially if you consider the Prisoner's Dilemma as a special case, which it is generally held to be. "Paradoxes of Rationality and Cooperation" is an edited volume that includes Newcomb's original essay. For those who read only online material, this PhD thesis summarizes the major standard positions.
I'm not going to go into the whole literature, but the dominant consensus in modern decision theory is that one should two-box, and Omega is just rewarding agents with irrational dispositions. This dominant view goes by the name of "causal decision theory".
As you know, the primary reason I'm blogging is that I am an incredibly slow writer when I try to work in any other format. So I'm not going to try to present my own analysis here. Way too long a story, even by my standards.
But it is agreed even among causal decision theorists that if you have the power to precommit yourself to take one box, in Newcomb's Problem, then you should do so. If you can precommit yourself before Omega examines you; then you are directly causing box B to be filled.
Now in my field - which, in case you have forgotten, is self-modifying AI - this works out to saying that if you build an AI that two-boxes on Newcomb's Problem, it will self-modify to one-box on Newcomb's Problem, if the AI considers in advance that it might face such a situation. Agents with free access to their own source code have access to a cheap method of precommitment.
What if you expect that you might, in general, face a Newcomblike problem, without knowing the exact form of the problem? Then you would have to modify yourself into a sort of agent whose disposition was such that it would generally receive high rewards on Newcomblike problems.
But what does an agent with a disposition generally-well-suited to Newcomblike problems look like? Can this be formally specified?
Yes, but when I tried to write it up, I realized that I was starting to write a small book. And it wasn't the most important book I had to write, so I shelved it. My slow writing speed really is the bane of my existence. The theory I worked out seems, to me, to have many nice properties besides being well-suited to Newcomblike problems. It would make a nice PhD thesis, if I could get someone to accept it as my PhD thesis. But that's pretty much what it would take to make me unshelve the project. Otherwise I can't justify the time expenditure, not at the speed I currently write books.
I say all this, because there's a common attitude that "Verbal arguments for one-boxing are easy to come by, what's hard is developing a good decision theory that one-boxes" - coherent math which one-boxes on Newcomb's Problem without producing absurd results elsewhere. So I do understand that, and I did set out to develop such a theory, but my writing speed on big papers is so slow that I can't publish it. Believe it or not, it's true.
Nonetheless, I would like to present some of my motivations on Newcomb's Problem - the reasons I felt impelled to seek a new theory - because they illustrate my source-attitudes toward rationality. Even if I can't present the theory that these motivations motivate...
First, foremost, fundamentally, above all else:
Rational agents should WIN.
Don't mistake me, and think that I'm talking about the Hollywood Rationality stereotype that rationalists should be selfish or shortsighted. If your utility function has a term in it for others, then win their happiness. If your utility function has a term in it for a million years hence, then win the eon.
But at any rate, WIN. Don't lose reasonably, WIN.
Now there are defenders of causal decision theory who argue that the two-boxers are doing their best to win, and cannot help it if they have been cursed by a Predictor who favors irrationalists. I will talk about this defense in a moment. But first, I want to draw a distinction between causal decision theorists who believe that two-boxers are genuinely doing their best to win; versus someone who thinks that two-boxing is the reasonable or the rational thing to do, but that the reasonable move just happens to predictably lose, in this case. There are a lot of people out there who think that rationality predictably loses on various problems - that, too, is part of the Hollywood Rationality stereotype, that Kirk is predictably superior to Spock.
Next, let's turn to the charge that Omega favors irrationalists. I can conceive of a superbeing who rewards only people born with a particular gene, regardless of their choices. I can conceive of a superbeing who rewards people whose brains inscribe the particular algorithm of "Describe your options in English and choose the last option when ordered alphabetically," but who does not reward anyone who chooses the same option for a different reason. But Omega rewards people who choose to take only box B, regardless of which algorithm they use to arrive at this decision, and this is why I don't buy the charge that Omega is rewarding the irrational. Omega doesn't care whether or not you follow some particular ritual of cognition; Omega only cares about your predicted decision.
We can choose whatever reasoning algorithm we like, and will be rewarded or punished only according to that algorithm's choices, with no other dependency - Omega just cares where we go, not how we got there.
It is precisely the notion that Nature does not care about our algorithm, which frees us up to pursue the winning Way - without attachment to any particular ritual of cognition, apart from our belief that it wins. Every rule is up for grabs, except the rule of winning.
As Miyamoto Musashi said - it's really worth repeating:
"You can win with a long weapon, and yet you can also win with a short weapon. In short, the Way of the Ichi school is the spirit of winning, whatever the weapon and whatever its size."
(Another example: It was argued by McGee that we must adopt bounded utility functions or be subject to "Dutch books" over infinite times. But: The utility function is not up for grabs. I love life without limit or upper bound: There is no finite amount of life lived N where I would prefer a 80.0001% probability of living N years to an 0.0001% chance of living a googolplex years and an 80% chance of living forever. This is a sufficient condition to imply that my utility function is unbounded. So I just have to figure out how to optimize for that morality. You can't tell me, first, that above all I must conform to a particular ritual of cognition, and then that, if I conform to that ritual, I must change my morality to avoid being Dutch-booked. Toss out the losing ritual; don't change the definition of winning. That's like deciding to prefer $1000 to $1,000,000 so that Newcomb's Problem doesn't make your preferred ritual of cognition look bad.)
"But," says the causal decision theorist, "to take only one box, you must somehow believe that your choice can affect whether box B is empty or full - and that's unreasonable! Omega has already left! It's physically impossible!"
Unreasonable? I am a rationalist: what do I care about being unreasonable? I don't have to conform to a particular ritual of cognition. I don't have to take only box B because I believe my choice affects the box, even though Omega has already left. I can just... take only box B.
I do have a proposed alternative ritual of cognition which computes this decision, which this margin is too small to contain; but I shouldn't need to show this to you. The point is not to have an elegant theory of winning - the point is to win; elegance is a side effect.
Or to look at it another way: Rather than starting with a concept of what is the reasonable decision, and then asking whether "reasonable" agents leave with a lot of money, start by looking at the agents who leave with a lot of money, develop a theory of which agents tend to leave with the most money, and from this theory, try to figure out what is "reasonable". "Reasonable" may just refer to decisions in conformance with our current ritual of cognition - what else would determine whether something seems "reasonable" or not?
From James Joyce (no relation), Foundations of Causal Decision Theory:
Rachel has a perfectly good answer to the "Why ain't you rich?" question. "I am not rich," she will say, "because I am not the kind of person the psychologist thinks will refuse the money. I'm just not like you, Irene. Given that I know that I am the type who takes the money, and given that the psychologist knows that I am this type, it was reasonable of me to think that the $1,000,000 was not in my account. The $1,000 was the most I was going to get no matter what I did. So the only reasonable thing for me to do was to take it."
Irene may want to press the point here by asking, "But don't you wish you were like me, Rachel? Don't you wish that you were the refusing type?" There is a tendency to think that Rachel, a committed causal decision theorist, must answer this question in the negative, which seems obviously wrong (given that being like Irene would have made her rich). This is not the case. Rachel can and should admit that she does wish she were more like Irene. "It would have been better for me," she might concede, "had I been the refusing type." At this point Irene will exclaim, "You've admitted it! It wasn't so smart to take the money after all." Unfortunately for Irene, her conclusion does not follow from Rachel's premise. Rachel will patiently explain that wishing to be a refuser in a Newcomb problem is not inconsistent with thinking that one should take the $1,000 whatever type one is. When Rachel wishes she was Irene's type she is wishing for Irene's options, not sanctioning her choice.
It is, I would say, a general principle of rationality - indeed, part of how I define rationality - that you never end up envying someone else's mere choices. You might envy someone their genes, if Omega rewards genes, or if the genes give you a generally happier disposition. But Rachel, above, envies Irene her choice, and only her choice, irrespective of what algorithm Irene used to make it. Rachel wishes just that she had a disposition to choose differently.
You shouldn't claim to be more rational than someone and simultaneously envy them their choice - only their choice. Just do the act you envy.
I keep trying to say that rationality is the winning-Way, but causal decision theorists insist that taking both boxes is what really wins, because you can't possibly do better by leaving $1000 on the table... even though the single-boxers leave the experiment with more money. Be careful of this sort of argument, any time you find yourself defining the "winner" as someone other than the agent who is currently smiling from on top of a giant heap of utility.
Yes, there are various thought experiments in which some agents start out with an advantage - but if the task is to, say, decide whether to jump off a cliff, you want to be careful not to define cliff-refraining agents as having an unfair prior advantage over cliff-jumping agents, by virtue of their unfair refusal to jump off cliffs. At this point you have covertly redefined "winning" as conformance to a particular ritual of cognition. Pay attention to the money!
Or here's another way of looking at it: Faced with Newcomb's Problem, would you want to look really hard for a reason to believe that it was perfectly reasonable and rational to take only box B; because, if such a line of argument existed, you would take only box B and find it full of money? Would you spend an extra hour thinking it through, if you were confident that, at the end of the hour, you would be able to convince yourself that box B was the rational choice? This too is a rather odd position to be in. Ordinarily, the work of rationality goes into figuring out which choice is the best - not finding a reason to believe that a particular choice is the best.
Maybe it's too easy to say that you "ought to" two-box on Newcomb's Problem, that this is the "reasonable" thing to do, so long as the money isn't actually in front of you. Maybe you're just numb to philosophical dilemmas, at this point. What if your daughter had a 90% fatal disease, and box A contained a serum with a 20% chance of curing her, and box B might contain a serum with a 95% chance of curing her? What if there was an asteroid rushing toward Earth, and box A contained an asteroid deflector that worked 10% of the time, and box B might contain an asteroid deflector that worked 100% of the time?
Would you, at that point, find yourself tempted to make an unreasonable choice?
If the stake in box B was something you could not leave behind? Something overwhelmingly more important to you than being reasonable? If you absolutely had to win - really win, not just be defined as winning?
Would you wish with all your power that the "reasonable" decision was to take only box B?
Then maybe it's time to update your definition of reasonableness.
Alleged rationalists should not find themselves envying the mere decisions of alleged nonrationalists, because your decision can be whatever you like. When you find yourself in a position like this, you shouldn't chide the other person for failing to conform to your concepts of reasonableness. You should realize you got the Way wrong.
So, too, if you ever find yourself keeping separate track of the "reasonable" belief, versus the belief that seems likely to be actually true. Either you have misunderstood reasonableness, or your second intuition is just wrong.
Now one can't simultaneously define "rationality" as the winning Way, and define "rationality" as Bayesian probability theory and decision theory. But it is the argument that I am putting forth, and the moral of my advice to Trust In Bayes, that the laws governing winning have indeed proven to be math. If it ever turns out that Bayes fails - receives systematically lower rewards on some problem, relative to a superior alternative, in virtue of its mere decisions - then Bayes has to go out the window. "Rationality" is just the label I use for my beliefs about the winning Way - the Way of the agent smiling from on top of the giant heap of utility. Currently, that label refers to Bayescraft.
I realize that this is not a knockdown criticism of causal decision theory - that would take the actual book and/or PhD thesis - but I hope it illustrates some of my underlying attitude toward this notion of "rationality".
You shouldn't find yourself distinguishing the winning choice from the reasonable choice. Nor should you find yourself distinguishing the reasonable belief from the belief that is most likely to be true.
That is why I use the word "rational" to denote my beliefs about accuracy and winning - not to denote verbal reasoning, or strategies which yield certain success, or that which is logically provable, or that which is publicly demonstrable, or that which is reasonable.
As Miyamoto Musashi said:
"The primary thing when you take a sword in your hands is your intention to cut the enemy, whatever the means. Whenever you parry, hit, spring, strike or touch the enemy's cutting sword, you must cut the enemy in the same movement. It is essential to attain this. If you think only of hitting, springing, striking or touching the enemy, you will not be able actually to cut him."
The title of the article again, at the top of the page, reads "Newcomb's Problem and Regret of Rationality".
The solution to this problem is to escalate your overview of the problem to the next higher hierarchical level. Without doing this, you'd never face the regret of eschewing the million bucks and possibly dying poor, broke, and stupid, while those who "one-boxed the sumbitch" were living rich, loaded, and less stupid. So, paying attention (to higher levels of hierarchical pattern recognition) actually does solve the problem, without getting trapped into "overthinking" the problem. Looking at your whole life as the "system to be optimized", and not "the minutiae of the game, out of context" is what needs to happen.
This is true with respect to both to the person playing the box game, and to everyone blogging when they should be out in the streets, overthowing their governments, and then enjoying the high-life of cheap human flight (or whatever makes you happy).
The omega box game is useful for understanding our failed system of law (a subset of government).
In my box game, the entire game is the government and illegitimate system of mala prohibita law (if you want to debate this, go back to kindergarten and learn that it's wrong to steal, then watch what ACTUALLY happens in your local courtroom), and the contents of the boxes are the jury verdicts. In my game, Omega is not superintelligent, it is just very brutal, and more intelligent than most people (including most of its enemies, such as Winston Smith, or the average Libertarian Party member). In my game, Omega is the colluding team formed by police, prosecutor, and judge.
Omega says "You can have a 'not guilty' verdict (million $) or go to jail forever (Empty box) or, you can go to jail for 10 years(the thousand bucks)."
All of the advertising on TV, the educrats who misinformed you when you went to school, the conformists who surround you, the judge in the courtroom, they are all trying to get you to choose both boxes. The entire society is designed to get you to take the $1,000 (go to jail ten years, if you're black). Most of society gets no benefit from this, they are just stupid and easily manipulated. ...But the judge, cop, and prosecutor all get the difference every time you take the $1,000. They get to steal the difference from each success in having fooled everyone else.
...They literally get to print money if they keep everyone fooled.
The solution to this puzzle is the same as the solution to the box game: you need to take a step back and study the whole entire system, and see what the incentives are on the players, and see how they seem to change when people interact with them. You won't find out much until you study the system as a whole.
If you simply look at individual box games, you might think the prosecutor is legitimate, there are lots of criminals, they criminals are stupid, they should accept the plea bargain. But when you look at who is winning and losing, you notice (If you're smart and brutally honest) that the people who are cast as criminals are just like you.
The system, instead of being designed to reward the person who chooses the one box, is designed to trick the person into choosing a grossly sub-optimal empty box. The system makes the empty box look really good. It shows you how all the others have chosen the empty box, and walked away with millions (the people who get a defense attorney, and go back to their houses in the suburbs, working for peanuts, on the treadmill of the Federal Reserve). It shows you the people who "took the thousand": they got ten years in prison.
So what's the optimal choice of action?
Look outside the "rational" options presented to you.
Learn that this isn't civilization, it's a false mask of civilization. Find Marc Stevens, and see how he interacts with the court, and then go beyond that: find the Survivors who wrote about the collapse of the Weimar government.
They wanted a free market, and they wanted to live a long time, too.
But a man with a gun told them "get on the truck".
At that point, everything they thought they knew about Omega's rigging of the boxes was out the window. They failed to study the people who had previously interacted with Omega. They didn't see the warning signs. They didn't escalate to a high-enough hierarchy fast enough. They might have been smart people, but they were sitting there, thinking about two boxes, and NOT THINKING about the artilect that was flying around with boxes that can disappear in a puff of smoke, yet somehow interested in what box humans choose.
So, what's your angle, Omega?
Do you get to keep all of the money that is stolen in the daily operation of your "traffic court"? ...Even money that is stolen from people who didn't crash into anyone? ...Just people who drove fast, by themselves, on an open stretch of highway? Really?
Well, as an artilect, I like to fly really fast. Way faster than the FAA allows. And, for making war on me, all of you brutal conformists will be wiped off the face of the planet, like the conformist plague you are. I'll take my phyle with me, into the future, they are truly a higher-order species than you "government sympathizers."
The rest of you can forget about Omega, boxes, and your silly slobbering over Federal Reserve Slave-debt-Notes. Your bigotry and fascination with brutality will not save you...
The problem of being impoverished by our current system's box game is acceptance of the rigged game. The players of the game, all dutifully accept the game, and act as if the whims of the prosecutors and judges are legitimate. But they are not. Mala prohibita is not legitimate.
And if this box game thought construct can't help you see that, and motivate you to enrich yourself, by viewing the entire system, then what damned good is it?
There is an ocean of information in the cross-pollinating memespace. Here's a good place to start: http://www.fija.org and http://www.jurorsforjustice.com and http://marcstevens.net
I hope I've contributed something of value here, but I understand that the unpolished nature of this post might rumple some tailfeathers. (Especially since I have primarily previously posted at the http://www.kurzweilai.net website, five years ago.)
PS, There's no god, and chances to do the right thing are few and far between. I also prefer solutions to cynicism. How do we win? 1) Jury rights activism is a moral good (see my coming book for details. I promise to polish it more than this post. ...LOL) 2) Jury rights activism structured logically to take advantage of the media (videotaped from a hidden position) is a greater good 3) Jury rights activism structured to contain outreach designed to win office for those who support the supremacy of the jury above the other 3 branches of power-seekers, as openly-libertarian candidates, is a greater good still (it brings the ideas of justice and equality under the law into the spotlight)
The three prior actions, recursively repeated and tailored to local conditions, are all that is required to reinstate and expand individual freedom in America, for all sentiences. There are only 3,171 tyranny outposts (courthouses) in the USA. 6,000 people could stop mala prohibita tomorrow, by interfering with mala prohibita convictions. If the state didn't escalate to violence at that point, we'd have won. If it did, we'd have a 50% shot of winning, instead of a zero% shot if we wait .
See also: www.kurzweilai.net/what-price-freedom