Perplexed comments on The Aliens have Landed! - Less Wrong

33 Post author: TimFreeman 19 May 2011 05:09PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (156)

You are viewing a single comment's thread. Show more comments above.

Comment author: Perplexed 19 May 2011 06:49:02PM 1 point [-]

How is offering to supply ice cream characterized as "extortion"?

In any case, I was not using the scenario as a reductio against universal unreciprocated altruism. That notion fails under its own weight, due to complete absence of support.

Comment author: Wei_Dai 19 May 2011 07:06:16PM *  5 points [-]

Sorry, I misread your comment and thought it was an extortion scenario similar to the OP. Now that I've read it more carefully, it's not clear to me that we shouldn't give up the Niobium in order to provide those humans workers with ice cream. (ETA: why did you characterize those humans as indentured workers? It would have worked as well if they were just ordinary salaried workers.)

That notion fails under its own weight, due to complete absence of support.

Altruists certainly claim to have support for their stated preferences. Or one could argue that preferences don't need to have support. What kind of support do you have for liking ice cream, for example?

Comment author: wedrifid 19 May 2011 07:12:09PM 3 points [-]

Sorry, I misread your comment and thought it was an extortion scenario similar to the OP.

Your reading wasn't far off: "in all of these thought experiments" makes your reply remain relevant.

Comment author: Perplexed 19 May 2011 07:25:37PM 5 points [-]

True enough. My main objection to calling my ice cream negotiating tactic 'extortion' is that I really don't like the "just say 'No' to extortion" heuristic. I see no way of definitionally distinguishing extortion from other, less objectionable negotiating stances. Nash's 1953 cooperative game theory model suggests that it is rational to yield to credible threats. I.e. saying 'no' to extortion doesn't win! An AI that begins with the "just say no" heuristic will self-modify to one that dispenses with that heuristic.

Comment author: Wei_Dai 19 May 2011 11:11:26PM *  6 points [-]

I don't think anybody is suggesting building an explicit "just say 'No' to extortion" heuristic into an AI. (I agree we do not have a good definition of "extortion" so when I use the word I use it in an intuitive sense.) We're trying to find a general decision theory that naturally ends up saying no to extortion (when it makes sense to).

Here's an argument that "saying 'no' to extortion doesn't win" can't be the full picture. Some people are more credibly resistant to extortion than others and as a result are less likely to be extorted. We want an AI that is credibly resistant to extortion, if such credibility is possible. Now if other players in the picture are intelligent enough, to the extent of being able to deduce our AI's decision algorithm, then isn't being "credibly resistant to extortion" the same as having a decision algorithm that actually says no to extortion?

ETA: Of course the concept of "credibility" breaks down a bit when all agents are reasoning this way. Which is why the problem is still unsolved!

Comment author: timtyler 20 May 2011 05:15:27PM 1 point [-]

Of course the concept of "credibility" breaks down a bit when all agents are reasoning this way.

It does what? How so?

Comment author: Perplexed 20 May 2011 05:06:16AM *  0 points [-]

I don't think anybody is suggesting building an explicit "just say 'No' to extortion" heuristic into an AI. (I agree we do not have a good definition of "extortion" so when I use the word I use it in an intuitive sense.) We're trying to find a general decision theory that naturally ends up saying no to extortion (when it makes sense to).

That is pretty incoherent. If you are trying to come up with a general decision theory that wins and also says no to extortion, then you have overdetermined the problem (or will overdetermine it once you supply your definition). If you are predicting that a decision theory that wins will say no to extortion, then it is a rather pointless claim until you supply a definition. Perhaps what you really intend to do is to define 'extortion' as 'that which a winning decision theory says no to'. In which case, Nash has defined 'extortion' for you - as a threat which is not credible, in his technical sense.

ETA: Of course the [informal] concept of "credibility" breaks down a bit when all agents are reasoning this way. Which is why the problem is still unsolved!

Why do you say the problem is still unsolved? What issues do you feel were not addressed by Nash in 1953? Where is the flaw in his argument?

Part of the difficulty of discussing this here is that you have now started to use the word "credible" informally, when it also has a technical meaning in this context.

Comment author: lessdazed 19 May 2011 11:34:49PM 0 points [-]

"Commit to just saying 'no' and proving that when just committing to just saying 'no' and proving that wins."

Perhaps something like that.

Comment author: timtyler 19 May 2011 07:48:02PM *  5 points [-]

I really don't like the "just say 'No' to extortion" heuristic.

Well you don't want to signal that you give in to extortion. That would just increase the chances of people attempting extortion against you. Better to signal that you are on a vendetta to stamp out extortion - at your personal expense!!!

Comment author: Perplexed 19 May 2011 10:03:45PM 0 points [-]

There is an idea, surprisingly prevalent on a rationality website, that costless signaling is an effective way to influence the behavior of rational agents. Or in other words, that it is rational to take signalling at face value. I personally doubt that this idea is correct. In any case, I reiterate that I suggest yielding only to credible threats. My own announcements do not change the credibility of any threats available to agents seeking to exploit me.

Comment author: lessdazed 19 May 2011 11:38:00PM 0 points [-]

Perhaps what is really being expressed is the belief that social costs are real, and that mere pseudonymous posting has costs.

Comment author: Perplexed 20 May 2011 05:10:08AM 2 points [-]

huh?????

Comment author: lessdazed 20 May 2011 08:31:52AM 1 point [-]

Maybe if you provided examples of people seeming to say that "costless signaling is an effective way to influence the behavior of rational agents," we could ask them what they meant, and they might say something like "no signaling is actually costless".

Statements like "Someone going on record as having opinion X has given decent reason to suppose that person (believes he or she) actually holds opinion X," are interpretable of having either of the two meanings above. Since you didn't provide examples, I wasn't persuaded that you are describing people's ideas, and I suspect ambiguous statements like that are behind our clash of intuitions about what people think.

Comment author: Perplexed 20 May 2011 01:57:52PM 0 points [-]

Maybe if you provided examples of people seeming to say that "costless signaling is an effective way to influence the behavior of rational agents," we could ask them what they meant, and they might say something like "no signaling is actually costless".

Ok. That makes some sense. Though I still don't have a clue as to why you mention "social costs" or "pseudonymous posting".

So, for the example of people seeming to say that costless signaling is an effective way to influence the behavior of rational agents, I would direct you to the comment to which I was replying. Tim wrote:

Well you don't want to signal that you give in to extortion. That would just increase the chances of people attempting extortion against you. Better to signal that you are on a vendetta to stamp out extortion - at your personal expense!!!

I interpreted that as advocating costless signaling as a way of influencing the behavior of would-be extortionists. My response to that advocacy: Announcing that I am on a vendetta is cheap talk, and influences no one. No rational agent will believe such self-serving puffery unless I actually experience a level of personal expense commensurate with what I hope to gain by convincing them. Which makes the signaling not just costly, but irrational.

Comment author: timtyler 19 May 2011 10:14:08PM *  0 points [-]

My own announcements do not change the credibility of any threats available to agents seeking to exploit me.

They inflluence the liklihood of them being made in the first place - by influencing the attacker's expected payoffs. Especially if it appears as though you were being sincere. Your comment didn't look much like signalling. I mean, it doesn't seem terribly likely that someone would deliberately publicly signal that they are more likely than unnamed others to capitulate if threatened with an attempt at extortion.

Credibly signalling resistance to extortion is non-trivial. Most compelling would be some kind of authenticated public track record of active resistance.

Comment author: timtyler 20 May 2011 05:21:32PM 2 points [-]

I see no way of definitionally distinguishing extortion from other, less objectionable negotiating stances.

Well, a simple way would be to use the legal definition of extortion. That should at least help prevent house fires, kidnapping, broken windows and violence.

...but a better definition should not be too difficult - for instance: the set of "offers" which you would rather not be presented with.

Comment author: wedrifid 19 May 2011 07:43:59PM 3 points [-]

My main objection to calling my ice cream negotiating tactic 'extortion

My objection to calling the ice cream negotiation tactic 'extortion' is it just totally isn't. It's an offer of a trade.

Nash's 1953 cooperative game theory model suggests that it is rational to yield to credible threats. I.e. saying 'no' to extortion doesn't win!

Then it's a good thing we've made developments in our models in the last six decades!

Comment author: Perplexed 19 May 2011 10:08:49PM 2 points [-]

Then it's a good thing we've made developments in our models in the last six decades!

Cute. But perhaps you should provide a link to what you think is the relevant development.

Comment author: timtyler 20 May 2011 05:25:59PM *  3 points [-]

Well, the key concept underlying strong resistance to extortion is reputation management. Once you understand the long-term costs of becoming identified as a vulnerable "mark" by those in the criminal underground, giving in to extortion can start to look a lot less attractive.

Comment author: Perplexed 20 May 2011 06:47:26PM 0 points [-]

Tim, we are completely talking past each other here. To restate my position:

Nash in 1953 characterized rational 2 party bargaining with threats. Part of the solution was to make the quantitative distinction between 'non-credible' threats (which should be ignored because they cost the threatener so much to carry out that he would be irrational to do so), and 'credible' threats - threats which a threatener might rationally commit to carry out.

Since Nash is modeling the rationality of both parties here, it is irrational to resist a credible threat - in fact, to promise to do so is to make a non-credible threat yourself.

Hence, in Nash's model, cost-less signaling is pointless if both players are assumed to be rational. Such signaling does not change the dividing line between threats that are credible, and rationally should succeed, and those which are non-rational and should fail.

As for the 'costly signalling' that takes place when non-credible threats are resisted - that is already built into the model. And a consequence of the model is that it is a net loss to attempt to resist threats that are credible.

All of this is made very clear in any good textbook on game theory. It would save us all a great deal of time if you keep your amateur political theorizing to yourself until you read those textbooks.

Comment author: timtyler 20 May 2011 09:06:11PM *  2 points [-]

I am kinda surprised that you are in such a muddle about this - and are willing to patronise me over the issue!

"Don't negotiate with terrorists" and "don't give into extortion" are well-known maxims. As this thread illustrates, you don't seem to understand why they exist. I do understand. It isn't terribly complicated. I expect I can explain it to you.

If a government gives in to terrorist demands during a hijacking, it sends a signal to all the other terrorists in the world that the government is vulnerable to extortion. Subsequently the government is likely to face more hijackings.

So... in addition to the obvious cost associated with the immediate demands of the terrorists, there is a hidden cost associated with gaining a reputation for giving in to terrorists. That hidden cost is often huge. Thus the strategy of not giving in to terrorist demands - even if doing so looks attractive on the basis of a naive cost-benefit analysis.

Other forms of extortion exhibit similar dynamics...

Comment author: TimFreeman 20 May 2011 09:14:45PM *  2 points [-]

So, in addition to the obvious cost associated with the immediate demands of the terrorists, there is a hidden cost associated with getting a reputation for giving in to terrorists. That hidden cost is often huge. Thus the strategy of not giving in to terrorists.

So if Thud cooperated with some less drastic version of Fred's plan that left a future to care about, he would be causing humans to get a reputation for giving in to extortion, even if the particular extortion he was faced with would not have been prevented by the aliens knowing he probably would not have given in. This is a different argument from the backward causality UDT seems to use in this situation, and AIXI could get it right by simulating the behavior of the next extortionist.

Good idea. Thanks for posting.

Comment author: atucker 19 May 2011 09:53:31PM -1 points [-]

My objection to calling the ice cream negotiation tactic 'extortion' is it just totally isn't. It's an offer of a trade.

To elaborate a bit:

I'll give you utility if you give me utility is a trade.

I won't cause you disutility if you give me utility is extortion.

Comment author: ArisKatsaris 20 May 2011 06:45:08PM *  7 points [-]

I'll give you utility if you give me utility is a trade.

I won't cause you disutility if you give me utility is extortion.

I don't think that's exactly the right distinction. Let's say you go to your neighbour because he's being noisy.

Scenario A: He says "I didn't mean to disturb you, I just love my music loud. But give me 10 dollars, and sure, I'll turn the volume down." I'd call that a trade, though it's still about him not giving you disutility.

Scenario B: He says "Yeah, I do that on purpose, so that I can make people pay me to turn the volume down. It'll be 10 bucks. " I'd call that an extortion.

The difference isn't between the results of the offer if you accept or reject -- the outcomes and their utility for you is the same in each (loud music, silence - 10 dollars).

The difference is that in Scenario B, you wish the other person had never decided to make this offer. It's not the utility of your options that are to be compared with each other, but the utility of the timeline where the trade can be made vs the utility of the timeline where the trade can't be made...

In the Trade scenarios, if you can't make a trade with the person, he's still being noisy, and utility minimizes. In the Extortion scenarios, if you can't make a trade with the person, he has no reason to be noisy, and utility maximizes.

I'll probably let someone else to transform the above description into equations containing utility functions.

Comment author: atucker 20 May 2011 07:02:53PM -1 points [-]

Yeah, I was being sloppy.

The more important part for extortion is that they threaten to go out of their way to cause you harm. Schelling points and default states are probably relevant for the distinction.

You can't read a payoff table and declare it extortion or trade.

Comment author: ArisKatsaris 21 May 2011 06:36:25AM *  1 point [-]

Schelling points and default states are probably relevant for the distinction.

Meh. I hope we can define extortion much simpler than that.

How about "Extortion: Any offer of trade (t) by A to B, where A knows that the likely utility of B would be maximized if A had in advance treated (t) as certainly rejected."

In short extortion is any offer to you in which you could rationally wish you had clearly precommitted to reject it (and signalled such precommitment effectively), and A knows that.

Comment author: [deleted] 21 May 2011 07:46:20AM 2 points [-]

How about "Extortion: Any offer of trade (t) by A to B, where A knows that the likely utility of B would be maximized if A had in advance treated (t) as certainly rejected."

In short extortion is any offer to you in which you could rationally wish you had clearly precommitted to reject it (and signalled such precommitment effectively), and A knows that.

Another example.

A and B share an apartment, and so far A has been doing all that household chores even though both A and B care almost equally about a clean house. (Maybe A cares slightly more, so that A's cleanliness threshold is always reached slightly before B's threshold, so that A ends up doing the chore every time.)

So one day A gives B an ultimatum: if they do not share household chores equally, A will simply go on strike.

B realizes, too late, that B should have effectively and convincingly pre-committed earlier to never doing household chores, since this way A would never be tempted to offer the ultimatum.

A is aware of all this and breathes a sigh of relief that he made his ultimatum before B made that pre-commitment.

By the above definition, A is an extortionist.

Comment author: [deleted] 21 May 2011 07:14:22AM *  2 points [-]

How about "Extortion: Any offer of trade (t) by A to B, where A knows that the likely utility of B would be maximized if A had in advance treated (t) as certainly rejected."

In short extortion is any offer to you in which you could rationally wish you had clearly precommitted to reject it (and signalled such precommitment effectively), and A knows that.

B is threatening to kill his hostage unless a million dollars is deposited in B's offshore account and B safely arrives outside of legal jurisdiction.

A tells B that if B kills the hostage then A will kill B, but if B lets the hostage go then, in trade, A will not kill B.

B realizes, too late, that B should have set things up so that the hostage would automatically be killed if B didn't get what he wanted even if B got cold feet late in the game (this could be done by employing a third party whose professional reputation rests on doing as he is initially instructed regardless of later instructions). This would have greatly strengthened B's bargaining position.

A is aware of all this and breathes a sigh of relief that B did not have sufficient foresight.

Is A an extortionist? He is by the above definition.

Comment author: Perplexed 19 May 2011 10:06:37PM 1 point [-]

And what is the distinction between giving utility and not giving disutility? As consequentialists, I thought we were committed to the understanding that they are the same thing.

Comment author: atucker 19 May 2011 11:06:45PM -1 points [-]

The distinction is that I can commit to not giving into extortion, and not also turn down possibly beneficial trades.

Comment author: Perplexed 20 May 2011 04:40:33AM 0 points [-]

You seem to be assuming that committing to 'not giving in to extortion' will be effective in preventing rational threats from being made and carried out. Why do you assume that? Or, if you are not making that assumption, then how can you claim that you are not also turning down possibly beneficial trades?

Comment author: timtyler 20 May 2011 05:17:18PM 0 points [-]

You seem to be assuming that committing to 'not giving in to extortion' will be effective in preventing rational threats from being made and carried out. Why do you assume that?

Because then you don't get a reputation in the criminal underground for being vulnerable to extortion - and so don't face a circlling pack of extortionists, each eager for a piece of you.

Comment author: Perplexed 19 May 2011 07:15:12PM 0 points [-]

What kind of support do you have for liking ice cream, for example?

None at all. But then I don't claim that it is a universal moral imperative that will be revealed to be 'my own imperative' once my brain is scanned, the results of the scan are extrapolated, and the results are weighted in accordance with how "muddled" my preferences are judged to be.

Comment author: Wei_Dai 19 May 2011 07:30:36PM 3 points [-]

I see, so you're saying that universal unreciprocated altruism fails as a universal moral imperative, not necessarily as a morality that some people might have. Given that you used the word "crazy" earlier I thought you were claiming that nobody should have that morality.

Comment author: timtyler 19 May 2011 07:42:51PM 3 points [-]

I think it is easily possible to imagine naturalists describing some kinds of maladaptive behaviour as being "crazy". The implication would be that the behaviour was being caused by some kind of psychological problem interfering with their brain's normal operation.

Comment author: Perplexed 20 May 2011 04:39:33PM 0 points [-]

I thought you were claiming that nobody should have that morality.

I do claim that. In two flavors.

  1. Someone operating under that moral maxim will tend to dispense with that maxim as they approach reflective equilibrium.

  2. Someone operating under that 'moral' maxim is acting immorally - this operationally means that good people should (i.e. are under a moral obligation to) shun such a moral idiot and make no agreements with him (since he proclaims that he cannot be trusted to keep his commitments).

Part of the confusion between us is that you seem to want the word 'morality' to encompass all preferences - whether a preference for chocolate over vanilla, or a preference for telling the truth over lying, or a preference for altruism over selfishness. It is the primary business of metaethics to make the distinction between moral opinions (i.e. opinions about moral issues) and mere personal preferences.

Comment author: Wei_Dai 20 May 2011 07:18:05PM *  2 points [-]

Part of the confusion between us is that you seem to want the word 'morality' to encompass all preferences - whether a preference for chocolate over vanilla, or a preference for telling the truth over lying, or a preference for altruism over selfishness.

No, I don't want that. In fact I do not currently have a metaethical position beyond finding all existing metaethical theories (that I'm aware of) to be inadequate. In my earlier comment I offered two possible lines of defense for altruism, because I didn't know which metaethics you prefer:

Altruists certainly claim to have support for their stated preferences. Or one could argue that preferences don't need to have support.

In your reply to that comment you chose to respond to only the second sentence, hence the "confusion".

Anyway, why don't you make a post detailing your metaethics, as well as your arguments against "universal unreciprocated altruism"? It's not clear to me what you're trying to accomplish by calling people who believe such things (many of whom are very smart and have already seriously reflected on these issues) "crazy" without backing up your claims.

Comment author: Perplexed 20 May 2011 09:17:24PM *  0 points [-]

It's not clear to me what you're trying to accomplish by calling people who believe such things (many of whom are very smart and have already seriously reflected on these issues) "crazy" without backing up your claims.

I'm not sure why you think I have called anyone crazy. What I said above is that a particular moral notion is crazy.

Perhaps you instead meant to complain that (in the grandparent) I had referred to the persons in question as "moral idiots". I'm afraid I must plead guilty to that bit of hyperbole.

Anyway, why don't you make a post detailing your metaethics, as well as your arguments against "universal unreciprocated altruism"?

I am gradually coming to think that there is little agreement here as to what the word metaethics even means. My current understanding is that metaethics is what you do to prepare the linguistic ground so that people operating under different ethical theories and doctrines can talk to each other. Meta-ethics strives to be neutral and non-normative. There are no meta-ethical facts about the world - only definitions that permit discourse and disputation about the facts.

Given this interpretation of "meta-ethics", it would seem that what you mean to suggest is that I make a post detailing my normative ethics, which would include an argument against "universal unreciprocated altruism" (which I take to be a competing theory of normative ethics).

Luke and/or Eliezer and/or any trained philosopher here: I would appreciate feedback as to whether I finally have the correct understanding of the scope and purpose of meta-ethics.

Comment author: Wei_Dai 20 May 2011 11:02:02PM 0 points [-]

Given this interpretation of "meta-ethics", it would seem that what you mean to suggest is that I make a post detailing my normative ethics, which would include an argument against "universal unreciprocated altruism" (which I take to be a competing theory of normative ethics).

I thought you might have certain metaethical views, which might be important for understanding your normative ethics. But yes, I'm mainly interested in hearing about your normative ethics.