I've been reading through the sequences, and am currently working through the Intro to Bayes' Theorem (by the fact that I'm reading the Intro to Bayes (finally), you can tell that I'm pretty early in the process). It's been quite thought provoking. I'm finally getting questions right more reliably, and wanted to share one of the visualization tools that helped me, at least. There are many "applets" strewn about, written in Java, that help one to visualize what the various probability components are doing. In the mammography example, at least, an the idea of a sieve popped into my head as a neat way to think about what the test is doing.
I'm planning to take fairly extensive notes (more about that in a soon-to-come post), but thought I'd share a little "re-write" of that problem with a graphic in case it's of any use, and also in case I've blundered in my understanding. Re-writing things in my own words helps make them my own -- I realize that this is probably going to come across as really, really, incredibly, simplistic, but it's where I'm at!
In case it's not intuitive... it's supposed to show 100% of women broken into their measured partitions of 1% with cancer and 99% without. Those respective groups are then "sifted," and the known reliability of the sieve for each of those groups is used to determine p(cancer|test+).
I'm open to aesthetic critiques as well -- I enjoy making things like this and knowing how intuitive it is to look at is helpful. It didn't turn out how my mind visualized it, but I figured it was decent enough for a start.
This was made using emacs org-mode, LaTeX, and TikZ.
Update: per some comments, I tried to make things more clear in a redo. The original picture shown is HERE.
----- Click for bigger picture or download -----
As far as the take-home practical message goes, on my reading it was never about how well doctors could "diagnose cancer" per se based on mammogram results--rather, the reason we ask about P(cancer | positive) is because it ought to inform our decision about whether a biopsy is really warranted. If a healthy young woman from a population with an exceedingly low base rate for breast cancer has a positive mammogram, the prior probability of her having cancer may still be low enough that there might actually be negative expected value in following up with a biopsy; after all, let's not forgot that a biopsy is not a trivial procedure and things do sometimes go wrong.
So I think this actually does have some implication for real-world clinical care: we ought to question whether it is wise to automatically follow up all positive mammograms with biopsies. Maybe it is, and maybe it isn't, but I don't think we should take the question for granted as appears to be the case.
If a biopsy is the next step in diagnosing breast cancer after a positive mammogram, then we shouldn't perform mammograms on anyone it still wouldn't be worth biopsying should their mammogram turn up positive.