Followup to: Illusion of Transparency: Why No One Understands You, Expecting Short Inferential Distances
A few years ago, an eminent scientist once told me how he'd written an explanation of his field aimed at a much lower technical level than usual. He had thought it would be useful to academics outside the field, or even reporters. This ended up being one of his most popular papers within his field, cited more often than anything else he'd written.
The lesson was not that his fellow scientists were stupid, but that we tend to enormously underestimate the effort required to properly explain things.
He told me this, because I'd just told him about my experience publishing "An Intuitive Explanation of Bayesian Reasoning". This is still one of my most popular, most blogged, and most appreciated works today. I regularly get fan mail from formerly confused undergraduates taking statistics classes, and journalists, and professors from outside fields. In short, I successfully hit the audience the eminent scientist had thought he was aiming for.
I'd thought I was aiming for elementary school.
Today, when I look back at the Intuitive Explanation, it seems pretty silly as an attempt on grade school:
- It's assumed that the reader knows what a "probability" is.
- No single idea requires more than a single example.
- No homework problems! I've gotten several complaints about this.
(Then again, I get a roughly equal number of complaints that the Intuitive Explanation is too long and drawn-out, as that it is too short. The current version does seem to be "just right" for a fair number of people.)
Explainers shoot way, way higher than they think they're aiming, thanks to the illusion of transparency and self-anchoring. We miss the mark by several major grades of expertise. Aiming for outside academics gets you an article that will be popular among specialists in your field. Aiming at grade school (admittedly, naively so) will hit undergraduates. This is not because your audience is more stupid than you think, but because your words are far less helpful than you think. You're way way overshooting the target. Aim several major gradations lower, and you may hit your mark.
PS: I know and do confess that I need to work on taking my own advice.
Addendum: With his gracious permission: The eminent scientist was Ralph Merkle.
Douglas writes: Suppose I want to discuss a particular phenomena or idea with a Bayesian. Suppose this Bayesian has set the prior probability of this phenomena or idea at zero. What would be the proper gradient to approach the subject in such a case?
I would ask them for their records or proof. If one is a consistent Bayesian who expects to model reality with any accuracy, the only probabilities it makes sense to set as zero or one are empirical facts specificied at a particular point in space-time (such as: "I made X observation of Y on Z equipment at W time") or statements within a formal logical system (which are dependent on assumptions and can be proved from those assumptions).
Even those kinds of statements are probably not legitimate candidates for zero/one probability, since there is always some probability, however minuscule that we have misremembered, misconstrued the evidence or missed a flaw in our proof. But I believe these are the only kinds of statements which can, even in principle have probabilities of zero or 1.
All other statements run up against possibilities for error that seem (at least to my understanding) to be embedded in the very nature of reality.