cousin_it comments on Thomas C. Schelling's "Strategy of Conflict" - Less Wrong

81 Post author: cousin_it 28 July 2009 04:08PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (148)

You are viewing a single comment's thread. Show more comments above.

Comment author: Wei_Dai 28 July 2009 09:37:06PM 4 points [-]

In the radio example, there is no way for me to convince you that the receive capability is truly broken. Given that, there is no reason for me to actually break the receive ability, and you should distrust any claim on my part that the receive ability has been broken.

But Schelling must have been able to follow this reasoning, so what point was he trying to illustrate with the radio example?

Comment author: cousin_it 28 July 2009 09:50:23PM *  4 points [-]

Actually the book tells about a psychological experiment conducted similarly to the situation I described, but there the subjects were told outright by the trustworthy experimenter that their partner couldn't learn their whereabouts. But I still think that average-rational humans would often fall for the radio trick. Expected utility suggests you don't have to believe your partner 100%; a small initial doubt reinforced over a day could suffice.

And yep, the problem of making your precommitments trustworthy is also discussed in much detail in the book.

Comment author: Wei_Dai 29 July 2009 02:11:45AM 7 points [-]

There may be an interesting connection between this example and AIs knowing each other's source code. The idea is, if one AI can unilaterally prove its source code to another without the receiver being able to credibly deny receipt of the proof, then it should change its source code to commit to an unfair agreement that favors itself, then prove this. If it succeeds in being the first to do so, the other side then has no choice but to accept. So, Freaky Fairness seems to depend on the details of the proof process in some way.

Comment author: orthonormal 29 July 2009 06:52:53PM *  7 points [-]

If it succeeds in being the first to do so, the other side then has no choice but to accept.

This presumes that the other side obeys standard causal decision theory; in fact, it's an illustration of why causal decision theory is vulnerable to exploitation if precommitment is available, and suggests that two selfish rational CDT agents who each have precommitment options will generally wind up sabotaging each other.

This is a reason to reject CDT as the basis for instrumental rationality, even if you're not worried that Omega is lurking around the corner.

Comment author: Wei_Dai 29 July 2009 07:59:38PM 4 points [-]

You can reject CDT but what are you going to replace it with? Until Eliezer publishes his decision theory and I have a chance to review it, I'm sticking with CDT.

I thought cousin_it's result was really interesting because it seems to show that agents using standard CDT can nevertheless convert any game into a cooperative game, as long as they have some way to prove their source code to each other. My comment was made in that context, pointing out that the mechanism for proving source code needs to have a subtle property, which I termed "consensual".

Comment author: Vladimir_Nesov 29 July 2009 08:55:08PM *  7 points [-]

One obvious "upgrade" to any decision theory that has such problems is to discard all of your knowledge (data, observations) before making any decisions (save for some structural knowledge to leave the decision algorithm nontrivial). For each decision that you make (using given decision algorithm) while knowing X, you can make a conditional decision (using the same decision algorithm) that says "If X, then A else B", and then recall whether X is actually true. This, for example, mends the particular failure of not being able to precommit (you remember that you are on the losing branch only after you've made the decision to do a certain disadvantageous action if you are on the losing branch).

Comment author: Wei_Dai 30 July 2009 12:37:57AM 3 points [-]

You can claim that you are using such a decision theory and hence that I should find your precommitments credible, but if you have no way of proving this, then I shouldn't believe you, since it is to your advantage to have me believe you are using such a decision theory without actually using it.

From your earlier writings I think you might be assuming that AIs would be intelligent enough to just know what decision algorithms others are using, without any explicit proof procedure. I think that's an interesting possibility to consider, but not a very likely one. But maybe I'm missing something. If you wrote down any arguments in favor of this assumption, I'd be interested to see them.

Comment author: Vladimir_Nesov 30 July 2009 10:16:31AM 3 points [-]

That was an answer for your question about what should you replace CDT with. If you won't be able to convince other agents that you now run on timeless CDT, you gain a little smaller advantage than otherwise, but that's a separate problem. If you know that your claims of precommitment won't be believed, you don't precommit, it's that easy. But sometimes, you'll find a better solution than if you only lived in a moment.

Also note that even if you do convince other agents about the abstract fact that your decision theory is now timeless, it won't help you very much, since it doesn't prove that you'll precommit in a specific situation. You only precommit in a given situation if you know that this action makes the situation better for you, which in case of cooperation means that the other side will be able to tell whether you actually precommited, and this is not at all the same as being able to tell what decision theory you use.

Since using a decision theory with precommitment is almost always an advantage, it's easy to assume that a sufficiently intelligent agent always uses something of the sort, but that doesn't allow you to know more about their actions -- in fact, you know less, since such agent has more options now.

Comment author: Wei_Dai 30 July 2009 08:17:02PM *  4 points [-]

But sometimes, you'll find a better solution than if you only lived in a moment.

Yes, I see that your decision theory (is it the same as Eliezer's?) gives better solutions in the following circumstances:

  • dealing with Omega
  • dealing with copies of oneself
  • cooperating with a counterpart in another possible world

Do you think it gives better solutions in the case of AIs (who don't initially think they're copies of each other) trying to cooperate? If so, can you give a specific scenario and show how the solution is derived?

Comment author: Eliezer_Yudkowsky 29 July 2009 04:13:44AM 4 points [-]

Unless, of course, you already know that most AIs will go ahead and "suicidally" deny the unfair agreement.

Comment author: cousin_it 29 July 2009 07:25:00AM *  1 point [-]

Yes. In the original setting of FF the tournament setup enforces that everyone's true source code is common knowledge. Most likely the problem is hard to solve without at least a little common knowledge.

Comment author: Wei_Dai 29 July 2009 12:14:26PM *  1 point [-]

Yes. In the original setting of FF the tournament setup enforces that everyone's true source code is common knowledge. Most likely the problem is hard to solve without at least a little common knowledge.

Hmm, I'm not seeing what common knowledge has to do with it. Instead, what seems necessary is that the source code proving process must be consensual rather than unilateral. (The former has to exist, and the latter cannot, in order for FF to work.)

A model for a unilateral proof process would be a trustworthy device that accepts a string from the prover and then sends that string along with the message "1" to the receiver if the string is the prover's source code, and "0" otherwise.

A model for a consensual proof process would be a trustworthy device that accepts from the prover and verifier each a string, and sends a message "1" to both parties if the two strings are identical and represent the prover's source code, and "0" otherwise.

Comment author: cousin_it 29 July 2009 06:04:17PM *  0 points [-]

In your second case one party can still cheat by being out of town when the "1" message arrives. It seems to me that the whole endeavor hinges on the success of the exchange being common knowledge.

Comment author: Wei_Dai 29 July 2009 06:15:55PM 0 points [-]

In your second case one party can still cheat by being out of town when the "1" message arrives.

I'm not getting you. Can you elaborate on which party can cheat, and how. And by "second case" do you mean the "unilateral" one or the "consensual" one?

Comment author: cousin_it 29 July 2009 09:17:15PM *  0 points [-]

The "consensual" one.

For a rigorous demonstration, imagine this: while preparing to play the Freaky Fairness game, I managed to install a subtle bug into the tournament code that will slightly and randomly distort all source code inputs passed to my algorithm. Then I submit some nice regular quining-cooperative program. In the actual game your program will assume I will cooperate, while mine will see you as a defector and play to win. When the game gives players an incentive to misunderstand, even a slight violation of "you know that I know that you know..." can wreak havoc, hence my emphasis on common knowledge.

Comment author: Wei_Dai 29 July 2009 11:29:14PM *  0 points [-]

In the actual game your program will assume I will cooperate, while mine will see you as a defector and play to win.

I see what you're saying now, but this seems easy to prevent. Since you have changed your source code to FF, and I know you have, I can simply ask you whether you believe I am a defector, and treat you as a defector if you say "yes". I know your source code so I know you can't lie (specify Freaky Fairness to include this honesty). Doesn't that solve the problem?

ETA: There is still a chance of accidental miscommunication, but you no longer have an incentive to deliberately cheat.

Comment author: cousin_it 30 July 2009 09:24:56AM *  1 point [-]

In this solution you have an incentive to similarly be outa town when I say "no". Think through it recursively. Related topics: two generals problem, two-phase commit.