Comment author: PeerGynt 30 October 2014 04:51:46AM *  1 point [-]

Less Wrong

Less Wrong (German: Weniger Falsch) was an association of philosophers gathered on the internet in 2007, chaired by Eliezer Yudkowsky. Among its members were Yvain, Lukeprog, Michael Vassar, Will Newsome and Gwern. PeerGynt was an eminent student at the time. He was allowed to participate in meetings, but was not a member of Less Wrong.

Members of Less Wrong had a common attitude towards philosophy, consisting of an applied rationalism drawn from Eliezer Yudkowsky, whose Sequences formed the basis for the group's philosophy. Less Wrong's influence on 21st century philosophy was immense, and much later work was in response to the group's thoughts.

The pre-history of Less Wrong began with blog posts on the philosophy of science and epistemology from 2006, promoted by Robin Hanson on Overcoming Bias.

(This is only half joking. If you want the rest of the future history of Less Wrong, it is available at http://en.wikipedia.org/wiki/Vienna_Circle . )

(Edited to fix Google Translate's German grammar)

Comment author: fubarobfusco 08 October 2014 04:55:43PM *  3 points [-]

I've seen Chesterton's quote used or misused in ways that assume that an extant fence must have some use that is both ① still existent, and ② beneficial; and that it can only be cleared away if that use is overbalanced by some greater purpose.

But some fences were created to serve interests that no longer exist: Hadrian's Wall, for one. The fact that someone centuries ago built a fence to keep the northern barbarians out of Roman Britain does not mean that it presently serves that purpose. Someone who observed Hadrian's Wall without knowledge of the Roman Empire, and thus the wall's original purpose, might correctly conclude that it serves no current military purpose to England.

For that matter, some fences exist to serve invidious purposes. To say "I don't see the use of this" is often a euphemism for "I see the harm this does, and it does not appear to achieve any counterbalancing benefit. Indeed, its purpose appears to have always been to cause harm, and so it should be cleared away expeditiously."

Comment author: PeerGynt 08 October 2014 11:47:01PM 2 points [-]

But some fences were created to serve interests that no longer exist: Hadrian's Wall, for one. The fact that someone >centuries ago built a fence to keep the northern barbarians out of Roman Britain does not mean that it presently >serves that purpose. Someone who observed Hadrian's Wall without knowledge of the Roman Empire, and thus the >wall's original purpose, might correctly conclude that it serves no current military purpose to England.

At the risk of generalizing from fictional evidence: This line of reasoning falls apart when it turns out that the true reason for the wall is to keep Ice Zombies out of your kingdom. Chesterton would surely have seen the need be damn sure that the true purpose is to keep the wildlings out, before agreeing to reduce the defense at the wall.

Comment author: ChristianKl 29 July 2014 08:54:52PM -1 points [-]

The point I was trying to make is that, while I see females as agents in real life, in this analogy I am discussing the ethics of a choice that is only made by men.

You don't succeed in avoiding getting mind killed yourself. You switch for no reason towards real life.

For any of those things, if you give me a specific reason why it is relevant to the choice made by the Green Martians, then it certainly should have been part of the analogy.

General ethical consideration suggest that you only inflict pain on other humans if they consent. A doctor will only operate on a patient if the patient consents, even if the doctor believes that a decision to not consent is bad for the patient given the stated preferences of the patient. Respecting that decision means respecting the agentship of the patient.

That's even true for decisions such as whether to get vaccinated where herd immunity is a concern. No single person if forced to feel pain by getting vaccinate for the good of the group.

Comment author: PeerGynt 29 July 2014 09:15:52PM *  1 point [-]

You don't succeed in avoiding getting mind killed yourself. You switch for no reason towards real life.

Discussing the issue in terms of real life does not itself imply that I've been mindkilled (though it may increase the chance that the discussion ends up being subject to mindkill). If you think I have been mindkilled, please show me a specific instance where I used arguments as soldiers, or where I failed to update in response to a properly made argument.

General ethical consideration suggest that you only inflict pain on other humans if they consent.

That is a totally acceptable ethical view that is fully consistent with my parable. At no stage did I assert "Since we only care about Martians, it is acceptable for them to do anything they want to the Earthlings". Instead, I invited you to have discussion about what actions are ethical and which actions are not ethical. In such a discussion, one of the possible sides you can take is that the Martians should never tickle anyone without consent.

However, the real world implication of this assertion of it is that no man should attempt to interact with women unless they are certain that they are sufficiently high status to avoid seeming creepy.

(Note that I probably shouldn't have used "stinging pain" as an analogy for creepiness and social awkwardness. This was an overcompensation in order to avoid seeming biased in favor of men).

Comment author: ChristianKl 29 July 2014 04:56:34PM 1 point [-]

I definitely see the humans as agents, whose preferences are morally relevant.

Agents make decisions. The moment you ignore decision making and only think in terms of preferences agentship is gone.

Comment author: PeerGynt 29 July 2014 05:03:28PM 0 points [-]

Sure. The point I was trying to make is that, while I see females as agents in real life, in this analogy I am discussing the ethics of a choice that is only made by men. The analogy therefore did not require a fully specified model of females as agents.

There are many true things in the world that I chose not to specify in the analogy. For any of those things, if you give me a specific reason why it is relevant to the choice made by the Green Martians, then it certainly should have been part of the analogy. However, there is no law of nature that says "females should always be fully specified as agents in any analogy"

In response to comment by [deleted] on Open thread, July 28 - August 3, 2014
Comment author: Lumifer 29 July 2014 03:00:10PM 4 points [-]

Okay, women have a preference along a single axis which they do nothing about and do not express at all. The framework as described is all about what active agenty men could or should do to entirely passive npc women. I'm very far from being a feminist, but come on -- this is objectification and "don't worry your pretty head about it".

Comment author: PeerGynt 29 July 2014 03:33:39PM *  2 points [-]

It is true that some participants in the analogy are "non-player characters". That is because some ethical questions only have implications for the choices of a subset of the agents. It should be permissible to discuss these ethical questions. Doing this properly will require adding information about all stakeholders whenever it is relevant, but it does not necessarily require all stakeholders to be "playable" in the sense that they actively make ethical decisions.

It is also true that the women in my story have a preference on a single axis, and that in real life, they also have preferences on other axes. I did not specify those preferences in the analogy, because I did not see the point in adding complications that do not have relevance to the resolution of the ethical question, which is a choice faced only by Martians.

If you feel that there is an additional axis which has important implications for the ethical choice that the Green Martians are facing, please specify what that axis is and why it is important. This would be an important contribution to the discussion. Otherwise, this comes across as saying "you should have added additional complications that were not relevant, in order to sufficiently signal that women are important ethical agents and not objects".

The fact that women are important ethical agents is so obvious that it is not even worth debating. However, I shouldn't have to signal this at every opportunity as a precondition for taking part in the discussion, especially not when this would require me to add unnecessary information to the story.

As for why the women don't express their preference not to be tickled by green martians, this is simply because I took this preference to be obvious and common knowledge to all participants in the analogy.

Comment author: Punoxysm 28 July 2014 08:48:43PM 7 points [-]

I feel like parables here on LW, especially the longer and more tortured ones, are pretty much fallacy and bias breeding grounds. A couple egregious offenders, to my mind, include

Blue and Green Martians; about pick-up artistry

and

The Fable of the Dragon Tyrant; about death

Why do I take issue with them? Because while using analogies, including fanciful ones, can help us take the outside view on a problem where we are irrationally biased, these sorts of parables can also be a selective re-telling of the facts, and conclusions drawn from them simply don't transfer to the real world because of the way those facts are distorted, elided or transformed. Any argument with the conclusion then has to take us back to whatever the real-world analogue is, and explain why the parable is flawed.

In other words, a parable (particularly the long-winded, over-constructed sort people like to post on LW) can give you an outside view but it often just pulls you away from the only way to actually solve a problem: to engage with the problem, down the the gritty details that will be erased or distorted by a parable.

Comment author: PeerGynt 28 July 2014 09:29:21PM 9 points [-]

I'm fairly sure this comment was not exactly intended as a compliment, but I can think of worse insults than having my writing put in the same category as Nick Bostrom. As the author of the first of these parables, even I recognize that these two stories differ very significantly in quality

The Blue and Green Martians parable was an attempt to discuss a question of ethics that is important to many members of this community, and which it is almost impossible to discuss elsewhere. The decision to use an analogy was an attempt to minimize mindkill. This did not succeed. However, I am fairly sure that if I had chosen not to use an analogy, the resulting flamewar would have been immense. This probably means that there are certain topics we just can't discuss, which feels distinctly suboptimal, but I'm not sure I have a better solution.

Comment author: PeerGynt 27 July 2014 12:51:38AM *  6 points [-]

It's been a day since this discussion peaked, and I've had a chance to think a little bit more about this on a meta-level:

First of all, having a community built around epistemic hygiene is extremely valuable. Discussions about topics that involve mindkill are incredibly unpleasant, and may make it impossible for such a community to be successful. I therefore fully understand people who want to keep these discussions away from Less Wrong, and I won't post again on this topic or any other mindkilling topic.

That said, I think the inability to discuss this rationally and dispassionately is a major problem for society in general, which may contribute to some individuals with abnormal psychology reacting in unpredictable ways. My only view on the object-level question is that low status men get a raw deal, that there is no good solution to the problem, and that PUA is probably very bad ethically. I have natural sympathy for low-status men, but I recognize that I may be biased because I have no experience seeing the situation from a female perspective. The post was an attempt to invite people to help me update my moral beliefs, by hearing from people who do not have those biases. Importantly, this could not have been done on any other website, because they would have been unable to convince me that their moral framework was coherent, that they are being honest about their ethical beliefs, or that they have made a good-faith effort to model low status males as relevant participants in the moral calculus.

There was some really good discussion in this thread, particularly Skeptical Lurker's comment, which may have made me aware of a contradiction in my worldview. I will keep thinking about what updates may be necessary. This is exactly the kind of feedback that I was hoping for - changing your mind is a good thing, and it is only possible if we can gave a genuine discussion as rationalists.

That said, for the Less Wrong community, the cost of these discussions probably outweighs the benefits.

Comment author: Lumifer 26 July 2014 01:37:06AM 1 point [-]

From the reactions, it is tempting to conclude that most people object to PUA partially on epistemic grounds.

It seems that a common objection is that your analogy does not match well the PUA worldview.

Comment author: PeerGynt 26 July 2014 01:44:40AM 1 point [-]

This is possible. Could someone please explain the important aspects of the PUA worldview that are being misrepresented? Particularly if they are relevant to the ethical question I am interested in? This would certainly help me clear up some confusion.

Comment author: Lumifer 26 July 2014 01:22:08AM 3 points [-]

I believe there exists an objective moral standard which is part of the territory. ... Obviously, we are unable to know whether our ethical maps correspond to the ethical territory.

Hold on. You're saying that there's objective morality but it's unknowable in principle? Then on what basis do you believe it exists and why would its existence even matter?

If we don't believe there is such a thing as an objective ethical standard ... then I fail to see the point in even discussing ethics.

Even if you think that ethics are a semi-arbitrary social construct, they are very useful for human societies and so worth discussing.

Comment author: PeerGynt 26 July 2014 01:33:56AM *  0 points [-]

Fair point - I should have phrased that differently. I think I intended that both in the weak sense "Our prior on moral statements should never be 0 or 1" and also in the slightly stronger sense "Ethics is difficult, so our priors should have high variance"

Comment author: PeerGynt 26 July 2014 12:53:25AM 7 points [-]

I want to point out that I, perhaps incorrectly, assumed this thought experiment could be interesting even to the anti-PUA crowd, because it would help them distill their thinking about whether they object to PUA on epistemic grounds ("they have incorrect beliefs about female psychology") or if they object on moral grounds ("they draw incorrect / evil conclusions about the ethical implications of the theory")

From the reactions, it is tempting to conclude that most people object to PUA partially on epistemic grounds. However, it is hard for me to understand why a disagreement about facts would lead to such heated debate in a community based around the Litany of Tarski.

View more: Prev | Next