An information cascade is a problem in group rationality. Wikipedia has excellent introductions and links about the phenomenon, but here is a meta-ish example using likelihood ratios.
Suppose in some future version of this site, there are several well-known facts:
- All posts come in two kinds, high quality (insightful and relevant) and low quality (old ideas rehashed, long hypotheticals).
- There is a well-known prior 60% chance of anything being high quality, rather than low quality. (We're doing well!)
- Readers get a private signal, either "high" or "low", their personal judgement of quality, which is wrong 20% of the time.
- The number of up and down votes is displayed next to each post. (Note the difference from the present system, which only displays up minus down. This hypothesis makes the math easier.)
- Readers are competent in Bayesian statistics and strive to vote the true quality of the post.
Let's talk about how the very first reader would vote. If they judged the post high quality, then they would multiply the prior likelihood ratio (6:4) times the bayes factor for a high private signal (4:1), get (6*4:4*1) = (6:1) and vote the post up. If they judged the post low quality then they would instead multiply by the bayes factor for a low private signal (1:4), get (6*1:4*4) = (3:8) and vote the post down.
There were two scenarios for the first reader (private information high or low). If we speculate that the first reader did in fact vote up, then there are two scenarios for the second scenario: There are two scenarios for the second reader:
- Personal judgement high: (6:4)*(4:1)*(4:1) = (24:1), vote up.
- Personal judgement low: (6:4)*(1:4)*(4:1) = (6:4), vote up against personal judgement.
Note that now there are two explanations for ending up two votes up. It could be that the second reader actually agreed, or it could be that the second reader was following the first reader and the prior against their personal judgement. That means that the third reader gets zero information from the second reader's personal judgement! The two scenarios for the third reader, and every future reader, are exactly analogous to the two scenarios for the second reader.
- Personal judgement high: (6:4)*(4:1)*(4:1) = (24:1), vote up.
- Personal judgement low: (6:4)*(1:4)*(4:1) = (6:4), vote up against personal judgement.
This has been a nightmare scenario of groupthink afflicting even diligent bayesians. Possible conclusions:
- Don't strive to vote the true quality of the post, strive to vote your personal judgement.
- Try to avoid even noticing the score. (Maybe scores could even be occluded, like spoiler-text?)
- Information cascades are dangerous and interesting. We should develop good cognitive citizenship techniques.
- Broadcast novel evidence, not conclusions.
Note: Olle found an error that necessitated a rewrite. I apologize.
I think the explanation for this on LessWrong is the same as the explanation on Reddit (which, from what I understand, served as the code base for LW).
People don't have unlimited time, and they are willing to spend time on LW reading good posts, but unwilling to waste time reading bad posts. Thus many people will somehow filter lower posts (I do so by simply sorting the posts from highest rated to lowest rated, and read until I run out of time or get bored).
If many people do this, then posts which are "generally agreed as good" will then to shoot up as everyone votes them up. Posts which are bad will generally drop down to the filter point (usually around 0), and then stay there rather than go further below.
If everyone voted on every post, you would expected bad posts to continue dropping down, and a post which was "generally agreed as bad" would have just as large a magnitude as posts which are "generally agreed as good", except in the negative direction.
But we don't see posts with a score of -23. They seem to either be neutral, or good. So my theory that people filter to only see (and thus only vote on) neutral or better article seems to be able to predict what actually happens with the scores.