Claim: memeticity in a scientific field is mostly determined, not by the most competent researchers in the field, but instead by roughly-median researchers. We’ll call this the “median researcher problem”.
Prototypical example: imagine a scientific field in which the large majority of practitioners have a very poor understanding of statistics, p-hacking, etc. Then lots of work in that field will be highly memetic despite trash statistics, blatant p-hacking, etc. Sure, the most competent people in the field may recognize the problems, but the median researchers don’t, and in aggregate it’s mostly the median researchers who spread the memes.
(Defending that claim isn’t really the main focus of this post, but a couple pieces of legible evidence which are weakly in favor:
- People did in fact try to sound the alarm about poor statistical practices well before the replication crisis, and yet practices did not change, so clearly at least some people did in fact see the problem and were in fact not memetically successful at the time. The claim is more general than just statistics-competence and replication, but at least in the case of the replication crisis it seems like the model must be at least somewhat true.
- Again using the replication crisis as an example, you may have noticed the very wide (like, 1 sd or more) average IQ gap between students in most fields which turned out to have terrible replication rates and most fields which turned out to have fine replication rates.
… mostly, though, the reason I believe the claim is from seeing how people in fact interact with research and decide to spread it.)
Two interesting implications of the median researcher problem:
- A small research community of unusually smart/competent/well-informed people can relatively-easily outperform a whole field, by having better internal memetic selection pressures.
- … and even when that does happen, the broader field will mostly not recognize it; the higher-quality memes within the small community are still not very fit in the broader field.
In particular, LessWrong sure seems like such a community. We have a user base with probably-unusually-high intelligence, community norms which require basically everyone to be familiar with statistics and economics, we have fuzzier community norms explicitly intended to avoid various forms of predictable stupidity, and we definitely have our own internal meme population. It’s exactly the sort of community which can potentially outperform whole large fields, because of the median researcher problem. On the other hand, that does not mean that those fields are going to recognize LessWrong as a thought-leader or whatever.
We argue that memeticity—the survival and spread of ideas—is far more complex than the influence of average researchers or the appeal of articulate theories. Instead, the persistence of ideas in any field depends on a nuanced interplay of feedback mechanisms, boundary constraints, and the conditions within the field itself.
In fields like engineering, where feedback is often immediate and tied directly to empirical results, ideas face constant scrutiny through testing and validation. Failed practices here are swiftly corrected, creating a natural selection process that fosters robustness. Theories in engineering and similar disciplines rely on mathematical modeling to bridge concepts with real-world outcomes. This alignment between model and outcome isn’t instantaneous, but the structural setup of the field encourages what we might call “antifragility”: ideas that survive these feedback loops emerge stronger and more reliable, not solely because of the competence of individual researchers but because of the field’s built-in corrective pressures.
In contrast, fields like social sciences or linguistics often lack such direct empirical anchors. Theories in these areas can persist on the basis of articulation, cultural resonance, or ideological alignment, sometimes for decades. The classic linguistic theories of the 1970s, for instance, endured largely because they fit well within the intellectual climate of the time, with little empirical scrutiny available to challenge their assumptions. Without rigorous feedback, such theories may linger, shaping academic thought without the resilience-testing that empirical pressure imposes.
The emergence of large language models (LLMs) introduces a new dimension of feedback in these traditionally insulated fields. LLMs can analyze extensive linguistic and behavioral data, revealing patterns that either align with or contradict established theories. This new capacity acts as an initial “stress test” for long-standing ideas, challenging assumptions that may have previously gone unexamined. However, while LLMs provide valuable insights, they are not infallible arbiters of truth. Their analysis depends on training data that can inherit biases from past frameworks, so they function as a starting point rather than a comprehensive solution. The true rigor of empirical validation—akin to engineering’s feedback loops—remains essential for developing resilient theories.
In summary, the memetic success of ideas depends not just on the competency or articulacy of individual researchers but on how effectively feedback mechanisms, field boundaries, and empirical standards shape those ideas. Fields with strong, built-in corrective feedback—often mathematically modeled—are inherently more resilient to the persistence of weak ideas. Fields without such constraints are vulnerable to influence by articulation alone, creating environments where ideas can thrive without robust validation. The introduction of LLMs offers a valuable corrective force, but one that must be used with awareness of its limitations. By integrating empirical rigor and maintaining reflective practices, disciplines across the spectrum can ensure that memeticity aligns more closely with resilience, rather than rhetorical appeal alone.