We've all had arguments that seemed like a complete waste of time in retrospect. But at the same time, arguments (between scientists, policy analysts, and others) play a critical part in moving society forward. You can imagine how lousy things would be if no one ever engaged those who disagreed with them.
This is a list of tips for having "productive" arguments. For the purposes of this list, "productive" means improving the accuracy of at least one person's views on some important topic. By this definition, arguments where no one changes their mind are unproductive. So are arguments about unimportant topics like which Pink Floyd album is the best.
Why do we want productive arguments? Same reason we want Wikipedia: so people are more knowledgeable. And just like the case of Wikipedia, there is a strong selfish imperative here: arguing can make you more knowledgeable, if you're willing to change your mind when another arguer has better points.
Arguments can also be negatively productive if everyone moves further from the truth on net. This could happen if, for example, the truth was somewhere in between two arguers, but they both left the argument even more sure of themselves.
These tips are derived from my personal experience arguing.
Keep it Friendly
Probably the biggest barrier to productive arguments is the desire of arguers to save face and avoid publicly admitting they were wrong. Obviously, it's hard for anyone's views to get more accurate if no one's views ever change.
- Keep things warm and collegial. Just because your ideas are in violent disagreement doesn't mean you have to disagree violently as people. Stay classy.
- To the greatest extent possible, uphold the social norm that no one will lose face for publicly changing their mind.
- If you're on a community-moderated forum like Less Wrong, don't downvote something unless you think the person who wrote it is being a bad forum citizen (ex: spam or unprovoked insults). Upvotes already provide plenty of information about how comments and submissions should be sorted. (It's probably safe to assume that a new Less Wrong user who sees their first comment modded below zero will decide we're all jerks and never come back. And if new users aren't coming back, we'll have a hard time raising the sanity waterline much.)
- Err on the side of understating your disagreement, e.g. "I'm not persuaded that..." or "I agree that x is true; I'm not as sure that..." or "It seems to me..."
- If you notice some hypocrisy, bias, or general deficiency on the part of another arguer, think extremely carefully before bringing it up while the argument is still in progress.
Inquire about Implausible-Sounding Assertions Before Expressing an Opinion
If someone suggests something you find implausible, start asking friendly questions to get them to clarify and justify their statement. If their reasoning seems genuinely bad, you can refute it then.
As a bonus, doing nothing but ask questions can be a good way to save face if the implausible assertion-maker turns out to be right.
Be careful about rejecting highly implausible ideas out of hand. Ideally, you want your rationality to be a level where even if you started out with a crazy belief like Scientology, you'd still be able to get rid of it. But for a Scientologist to berid themselves of Scientology, they have to consider ideas that initially seen extremely unlikely.
It's been argued that many mainstream skeptics aren't really that good at critically evaluating ideas, just dismissing ones that seem implausible.
Isolate Specific Points of Disagreement
Stick to one topic at a time, until someone changes their mind or the topic is declared not worth pursuing. If your discussion constantly jumps from one point of disagreement to another, reaching consensus on anything will be difficult.
You can use hypothetical-oriented thinking like conditional probabilities and the least convenient possible world to figure out exactly what it is you disagree on with regard to a given topic. Once you've creatively helped yourself or another arguer clarify beliefs, sharing intuitions on specific "irreducible" assertions or anticipated outcomes that aren't easily decomposed can improve both of your probability estimates.
Don't Straw Man Fellow Arguers, Steel Man Them Instead
You might think that a productive argument is one where the smartest person wins, but that's not always the case. Smart people can be wrong too. And a smart person successfully convincing less intelligent folks of their delusion counts as a negatively productive argument (see definition above).
Play for all sides, in case you're the smartest person in the argument.
Rewrite fellow arguers' arguments so they're even stronger, and think of new ones. Arguments for new positions, even—they don't have anyone playing for them. And if you end up convincing yourself of something you didn't previously believe, so much the better.
If You See an Opportunity To Improve the Accuracy of Your Knowledge, Take It!
This is often called losing an argument, but you're actually the winner: you and your arguing partner both invested time to argue, but you were the only one who received significantly improved knowledge.
If you're worried about losing face or seeing your coalition (research group, political party, etc.) diminish in importance from you admitting that you were wrong, here are some ideas:
- Say "I'll think about it". Most people will quiet down at this point without any gloating.
- Just keep arguing, making a mental note that your mind has changed.
- Redirect the conversation, pretend to lose interest, pretend you have no time to continue arguing, etc.
Some of these techniques may seem dodgy, and honestly I think you'll usually do better by explaining what actually changed your mind. But they're a small price to pay for more accurate knowledge. Better to tell unimportant false statements to others than important false statements to yourself.
Have Low "Belief Inertia"
It's actually pretty rare that the evidence that you're wrong comes suddenly—usually you can see things turning against you. As an advanced move, cultivate the ability to update your degree of certainty in real time to new arguments, and tell fellow arguers if you find an argument of theirs persuasive. This can actually be a good way to make friends. It also encourages other arguers to share additional arguments with you, which could be valuable data.
One psychologist I agree with suggested that people ask
- "Does the evidence allow me to believe?" when evaluating what they already believe, but
- "Does the evidence compel me to believe?" when evaluating a claim incompatible with their current beliefs.
If folks don't have to drag you around like this for you to change your mind, you don't actually lose much face. It's only long-overdue capitulations that result in significant face loss. And the longer you put your capitulation off, the worse things get. Quickly updating in response to new evidence seems to preserve face in my experience.
If your belief inertia is low and you steel-man everything, you'll reach the super chill state of not having a "side" in any given argument. You'll play for all sides and you won't care who wins. You'll have achieved equanimity, content with the world as it actually is, not how you wish it was.
When you can equally well argue that X is true, and that X is false, it means that your arguing is quite entirely decoupled from truth, and as such, both of those arguments are piles of manure that shouldn't affect anyone's beliefs. It is only worth making one of such arguments to counter the other for the sake of keeping the undecided audience undecided. Ideally, you should instead present a very strong meta-ish argument that the argumentation for both sides which you would be able to make, is complete nonsense.
(Unfortunately that gets both sides of argument pissed off at you.)
That's not actually true.
In most real-world cases, both true statements and false statements have evidence in favor of them, and the process of assembling and presenting that evidence can be a perfectly functional argument. And it absolutely should affect your belief: if I present you with novel evidence in favor of A, your confidence in A should decrease.
Ideally, I should weigh my evidence in favor of A against my evide... (read more)