Comment author: Z_M_Davis 05 June 2009 06:19:38AM 2 points [-]

Re violence, do see "Bayesans vs. Barbarians."

Comment author: StanR 05 June 2009 09:16:47AM *  1 point [-]

I hope to post on that post shortly, after giving it some thought.

Gris, just as bias against violence may be the reason it's hardly ever considered, alternatively, it may not only be a rational position, but a strategically sensible one. Please consider looking at the literature concerning strategic nonviolence. The substantial literature at the Albert Einstein Institute is good for understanding nonviolent strategy and tactics against regimes, and the insights provided translate into courses of action in other conflicts, as well.

Comment author: taw 04 June 2009 06:40:41AM 0 points [-]

A third possibility: Humans aren't in general capable of accurately reflecting on their preferences.

Three is pretty much like one. If utility functions work, there must be some way of figuring them out, I hoped someone figured it out already.

If utility functions are a bad match for human preferences, that would seem to imply that humans simply tend not to have very consistent preferences. What major premise does this invalidate?

Utilitarian model being wrong doesn't necessarily mean that a different model based on different assumptions doesn't exist. I don't know which assumptions need to be broken.

Comment author: StanR 04 June 2009 07:31:55AM *  2 points [-]

The general premise in the mind sciences is that there are different selves, somehow coordinated through the cortical midline structures. Plenty of different terms have been used, and hypotheses suggested, but the two "selves" I use for shorthand come from Daniel Gilbert: Socrates and the dog. Socrates is the narrative self, the dog is the experiencing self. If you want something a bit more technical, I suggest the lectures about well-being (lecture 3) here, and to get really technical, this paper on cognitive science exploring the self.

Comment author: jimrandomh 31 May 2009 11:23:39PM 0 points [-]

I feel I should jump in here, as you appear to be talking past each other. There is no confusion in the system 1/system 2 distinction; you're both using the same definition, but the bit about decoys and shields was actually the core of PJ's post, and of the difference between your positions. PJ holds that to change someone's mind you must focus on their S1 response, because if they engage S2, it will just rationalize and confabulate to defend whatever position their S1 holds. Now, I have no idea how one would go about altering the S1 response of someone who didn't want their response altered, but I do know that many people respond very badly to rational arguments that go against their intuition, increasing their own irrationality as much as necessary to avoid admitting their mistake.

Comment author: StanR 01 June 2009 01:39:44AM *  3 points [-]

I don't believe we are, because I know of no evidence of the following:

evolutionarily speaking, a big function of system 2 is to function as a decoy/shield mechanism for keeping ideas out of a person. And increasing a person's skill at system 2 reasoning just increases their resistance to ideas.

Perhaps one or both of us misunderstands the model. Here is a better description of the two.

Originally, I was making a case that attempting to reason was the wrong strategy. Given your interpretation, it looks like pjeby didn't understand I was suggesting that, and then suggested essentially the same thing.

My experience, across various believers (Christian, Jehovah's Witness, New Age woo-de-doo) is that system 2 is never engaged on the defensive, and the sort of rationalization we're talking about never uses it. Instead, they construct and explain rationalizations that are narratives. I claim this largely because I observed how "disruptable" they were during explanations--not very.

How to approach changing belief: avoid resistance by avoiding the issue and finding something at the periphery of belief. Assist in developing rational thinking where the person has no resistance, and empower them. Strategically, them admitting their mistake is not the goal. It's not even in the same ballpark. The goal is rational empowerment.

Part of the problem, which I know has been mentioned here before, is unfamiliarity with fallacies and what they imply. When we recognize fallacies, most of the time it's intuitive. We recognize a pattern likely to be a fallacy, and respond. We've built up that skill in our toolbox, but it's still intuitive, like a chess master who can walk by a board and say "white mates in three."

Comment author: pjeby 31 May 2009 04:15:31PM 0 points [-]

It's not just nontrivial, it's incredibly hard. Engaging "system 2" reasoning takes a lot of effort, lowering sensitivity to, and acute awareness of, social cues and signals.

Engaging system 2 is precisely what you don't want to do, since evolutionarily speaking, a big function of system 2 is to function as a decoy/shield mechanism for keeping ideas out of a person. And increasing a person's skill at system 2 reasoning just increases their resistance to ideas.

To actually change attitudes and beliefs requires the engagement of system 1. Otherwise, even if you convince someone that something is logical, they'll stick with their emotional belief and just avoid you so they don't have to deal with the cognitive dissonance.

(Note that this principle also applies to changing your own beliefs and attitudes - it's not your logical mind that needs convincing. See Eliezer's story about overcoming a fear of lurking serial killers for an example of mapping System 2 thinking to System 1 thinking to change an emotional-level belief.)

Comment author: StanR 31 May 2009 10:43:27PM *  1 point [-]

pjeby, sorry I wasn't clear, I should have given some context. I am referencing system 1 and 2 as simplified categories of thinking as used by cognitive science, particularly in behavioral economics. Here's Daniel Kahneman discussing them. I'm not sure what you're referring to with decoys and shields, which I'll just leave at that.

To add to my quoted statement, workarounds are incredibly hard, and focusing on reasoning (system 2) about an issue or belief leaves few cycles for receiving and sending social cues and signals. While reasoning, we can pick up those cues and signals, but they'll break our concentration, so we tend to ignore them while reasoning carefully. The automatic, intuitive processing of the face interferes with the reasoning task; e.g. we usually look somewhere else when reasoning during a conversation. To execute a workaround strategy, however, we need to be attuned to the other person.

When I refer to belief, I'm not referring to fear of the dark or serial killers, or phobias. Those tend to be conditioned responses--the person knows the belief is irrational--and they can be treated easily enough with systematic desensitization and a little CBT thrown in for good measure. Calling them beliefs isn't wrong, but since the person usually knows they're irrational, they're outside my intended scope of discussion: beliefs that are perceived by the believer to be rational.

People are automatically resistant to being asked to question their beliefs. Usually it's perceived as unfair, if not an actual attack on them as a person: those beliefs are associated with their identity, which they won't abandon outright. We shouldn't expect them to. It's unrealistic.

What should we do, then? Play at the periphery of belief. To reformulate the interaction as a parable: We'll always lose if we act like the wind, trying to blow the cloak off the traveller. If we act like the sun, the traveller might remove his cloak on his own. I'll think about putting a post together on this.

Comment author: AdeleneDawner 29 May 2009 08:15:41PM 0 points [-]

I was going to say "there are more workarounds than you think", but that's probably my selection bias talking again. That said, there are workarounds, in some situations. It's still not a trivial thing to learn, though.

Comment author: StanR 31 May 2009 09:27:51AM *  3 points [-]

It's not just nontrivial, it's incredibly hard. Engaging "system 2" reasoning takes a lot of effort, lowering sensitivity to, and acute awareness of, social cues and signals.

The mindset of "let's analyze arguments to find weaknesses," aka Annoynance's "rational paths," is a completely different ballgame than most people are willing to play. Rationalists may opt for that game, but they can't win, and may be reinforcing illogical behavior. Such a rationalist is focused on whether arguments about a particular topic are valid and sound, not the other person's rational development. If the topic is a belief, attempting to reason it out with the person is counterproductive. Making no ground when engaging with people on a topic should be a red flag: "maybe I'm doing the wrong thing."

Does anyone care enough for me to make a post about workarounds? Maybe we can collaborate somehow Adelene, I have a little experience in this area.

Comment author: stcredzero 29 May 2009 04:31:33PM 0 points [-]

I suspect that meetup.com's continued existence is fueled by single people trying to actually meet interesting others.

Comment author: StanR 31 May 2009 07:32:24AM *  6 points [-]

I was part of a meetup on "alternative energy" (to see if actual engineers went to the things--I didn't want to date a solar cell) when I got an all-group email from the group founder about an "event" concerning The Secret* and a great opportunity to make money. Turned out it was a "green" multi-level marketing scam he was deep in, and they were combining it with the The Secret. Being naive, at first I said I didn't think the event was appropriate, assuming it might lead to some discussion. He immediately slandered me to the group, but I managed to send out an email detailing his connections to the scam before I was banned from the group. I did get a thank you from one of the members, at least.

I looked through meetup and found many others connected to him. Their basic routine involves paying the meetup group startup cost, having a few semi-legit meetings, and then using their meetup group as a captive audience.

I admit, I was surprised. I know it's not big news, but the new social web has plenty of new social scammers, and they're running interference. It's hard to get a strong, clear message out when opportunists know how to capitalize on easy money: people wanting to feel and signal like they're doing something. I honestly don't think seasteading can even touch that audience, but then again, I'm not sure you'd want to.

Comment author: PhilGoetz 15 April 2009 02:20:12PM *  0 points [-]

Depending on how polarized the sides are, the audience is either mostly, or completely going there to watch a fight and root for their team.

Like I said, entertainment.

If they just want to have pride in being right, and rallying their base, they shouldn't keep up a pretense of spreading rational atheism. They want both, but they can't have it.

Theory 2 proposes a way that rallying the base spreads atheism.

Comment author: StanR 16 April 2009 04:14:32AM *  1 point [-]

And I said rational atheism, not atheism.

Granted, I didn't express my thoughts on that clearly. I think there is a fundamental difference between attempting to get someone to agree with your opinions and helping them develop rationality--likely through both similar and different opinions. I think the latter is a better moral play, and it's genuine.

What is the higher priority result for a rational atheist targeting a theist: - a more rational theist - a not-any-more-rational-than-before atheist - an irrational agnostic

I think the biggest "win" of the results is the first. But groups claiming to favor rationality most still get stuck on opinions all the time (cryogenics comes to mind). Groups tend to congregate around opinions, and forcing that opinion becomes the group agenda, even though they believe it's the right and rational opinion to have. It's hard, because a group that shares few opinions is hardly a group at all, but the sharing of opinions and exclusion of those with different opinions works against an agenda of rationality. And shouldn't that be the agenda of a rational atheist?

I think the "people who don't care" of your 1st theory are either 1) unimportant or 2) don't exist, depending on the meaning of the phrase.

I think theory 2 makes a fatal mistake, it emphasizes the cart (opinion) rather than the horse (rational thinking). I'm willing to grant they're not so separate and cut-and-dry, but I wanted to illustrate the distinction I see.

Comment author: PhilGoetz 14 April 2009 09:48:11PM *  0 points [-]

It sounds like in your first paragraph about debates, you're saying a lot of people target their arguments ineffectively. Any theories on why that is, or whether there are specific biases involved?

I think that the main purpose of debates, in the minds of their sponsors, is to entertain rather than to inform. The news network that sets up a debate chooses people who will make colorful accusations against each other. They would be disappointed if the debaters reached agreement halfway through the debate.

A secondary purpose of debates may be for each side to educate and motivate its base.

Comment author: StanR 15 April 2009 06:13:42AM 0 points [-]

I think your secondary purpose is actually the primary purpose, excluding sponsors, who I agree, usually set up the debate for entertainment.

Even if both sides claim that changing minds is the purpose, the actions show otherwise. The "change minds" or "reveal the truth" is a convenient lie, and one that's actually believed. Plus, it would be tacky and uncivilized to state the real reason for the debate, best to claim a more noble imperative--and believe it.

Depending on how polarized the sides are, the audience is either mostly, or completely going there to watch a fight and root for their team. Although the audience may respect the other side if they play well, they're rooting for their side getting in some choice jabs, resulting in a KO. I don't think, leading up to the event, any side of a debate actually says "this will really change some minds!" No, it's usually "we're going to show them why we're right and they're wrong," or some more sophisticated equivalent of "it's beat-down time."

I will grant that if one side comes off as incredibly foolish, some may abandon it, but how often does that happen? Betting on a side already makes a person more confident of that choice being the right one.

Dawkins and other aggressive-in-that-way atheists irritate me: they're being very irrational about either their purpose or their approach. If they want more rational atheists, their chosen methods are very poor. If they just want to have pride in being right, and rallying their base, they shouldn't keep up a pretense of spreading rational atheism. They want both, but they can't have it.

I want to make a longer and more focused reply on what I see as the core questions: how can we change minds, and how should we? I've been wanting to tackle it here for a while, but I have trouble keeping up with this site.

Comment author: Cameron_Taylor 23 March 2009 12:31:17PM 7 points [-]

One of the things that I found most useful when I came across OvercomingBias was the frequent linking though most of the posts. In fact, in the beginning I spent a couple of hours on each new post, clicking through the links and absorbing the required background knowledge.

Especially now that posting has become somewhat more eclectic, I can see the value of providing a reference to some of the key topics that we assume for so much of our discussions.

In the mean time, some may find this list of Eleizer's OB posts useful. The dependency graphs there are handy too.

Comment author: StanR 24 March 2009 07:52:34AM 2 points [-]

Absolutely, linking really improves the resource.

A link for each major claim or background topic would be much appreciated. Sometimes I wonder if there shouldn't be an original post layer, also containing the comments, and a wiki-ish layer, that could provide more links and notations. That way, an entrant could dig deeper, and regulars could participate in bridging those gaps, but could also continue in the original post layer without the wiki-ish clutter.

Learning through participation is a problem when a post generates 30+ comments, some of which are asking for "beginner" clarifications of known-by-regulars concepts. I think it's better to include the beginners in grappling with the concepts and attempting to build the course for themselves and others. Isn't it a bit odd that so many learn by following their interest, filling-in gaps as needed, and yet later on those very same people will attempt to teach others using a more linear method?

Looking back on something I know, I see the map of knowledge, and think "Ah, I might have been better off had I learned these foundational basics first." The click-and-pursue nature of the web seems stacked against that method, and really, would have I been better off? Maybe I wouldn't have pursued as much as I did if I couldn't choose my own path.

If a wiki-ish layer is too crazy, maybe fundamental concepts should be more present in the tags, or be a separate little part? Related links and fundamental concepts relating to the post could be voted up or down. I could vote up a "hindsight bias" tag while downvoting a smartass "jedi" tag, and propose or vote up a link to another post that explored something (say, hindsight bias) in more detail.

View more: Next