A “softer” version of this problem arises when people agree that they don’t know if their reasoning is rational, either because they didn’t seek any evidence or didn’t try to come up with any arguments refuting their reasoning or models, or because they don’t know how to know if their theory is true or false
Can you demonstrate that you don't have the same problem?
The problem is always relative (depends on the amount of energy you are willing to invest in improving your models, and thus the expected information gain and pragmatic value increases from these improvements), and, anyway, demonstrating that I don't have a problem is not Popperian. Per Popper's epistemology, only the existence of a problem could be demonstrated, not the absence of it.
The particular passage that you have cited concerns more with the cases when people conjecture some theories or simply make up their minds on some topic (though, per Deutsch, the latter is a special case of the former), without any explicit epistemology in mind, or with an explicit epistemic methodology which nevertheless resembles a cargo cult rather than a rational, criticised stance, as captured in the title of Feynman's essay. Thus, the problem arises when people dismiss pointers towards the fact that their epistemology is "intuitive" or severely out of date, and suggested to base their further research/conjecturing/opinion-making on better epistemology, even if the theory in question (opinions) happens to be uncriticisable yet.
In this context, I try to follow Popperianism and Bayesianism, i. e., some explicit epistemology. I know there are some formal problems in connecting these two theories, but they don't seem important (valuable) to me enough to study cutting-edge research on this topic, let alone conduct it myself. I know at least one leading world scientist, Sean Carroll, who advocates for the same combo. Thus, all the available evidence leaves me pretty comfortable with the epistemological framework that I currently use and not actively look for problems with it, however, if these problems (or better theories) reach my attention (e. g., via people pointing me to them) and they seem compelling, I will likely invest some effort into researching these problems or theories.
Problem statement
The problem of epistemic irresponsibility, as I will call it in this post, is that people sometimes agree that their reasoning (and hence models that they hold and use as the basis of their action) is irrational or illogical. For example, people knowingly use theories that have already been refuted, such as astrology. Yet, they fail to act on this knowledge and don’t update their models and their action plan accordingly.
A “softer” version of this problem arises when people agree that they don’t know if their reasoning is rational, either because they didn’t seek any evidence or didn’t try to come up with any arguments refuting their reasoning or models, or because they don’t know how to know if their theory is true or false, i. e., don’t have the necessary epistemological apparatus (though, these people would seldom think in these terms). Then, people fail to act on this knowledge. The examples are people’s uncritical perception of news in the media or claims made by politicians, and, to some degree, the unsound scientific practices, such as those described in Richard Feynman’s “Cargo Cult Science”.
The problem of epistemic irresponsibility, or, rather, its inverse, epistemic responsibility, belongs to the discipline of normative logic, i. e. the “pre-Frege” logic, as John Sowa highlighted here:
Epistemic irresponsibility holds people from learning normative logic and rationality
In the context of rationality education, such as at the Center for Applied Rationality, or other programs, epistemic irresponsibility is an obstacle for people to decide they need rationality education at all and therefore read something about epistemology, logic, or rationality, take some courses, etc.
Here, the problem is somewhat circular, because acting on the fact that one doesn’t know normative logic is like already using normative logic!
The practical implication is that it’s hard to convince most people to read something about normative logic and rationality: at best, these will be very low on one’s list of things to study. Their responses would typically be: “I’m a successful person already, why should I spend time on it?”, or “I don’t have time for it, I would rather take that other course which teaches how to program or earn money on the internet.”
Thus, epistemic irresponsibility presents a challenge for spreading logic, rationality, and good science (such as physics) rather than bad science (such as astrology) more generally in the population.
Recognition in the literature, and related ideas
First, I derive the name of the problem, “epistemic irresponsibility”, from this passage by Richard Hamming in The Art of Doing Science and Engineering:
Second, Mortimer Adler in How to Read a Book:
And then, in another place in the book:
Third, David Deutsch’s idea of “taking theories which haven’t yet been refuted seriously”, is a major strand in his book The Fabric of Reality. This is the very beginning of the book:
Forth, a somewhat related idea is Nassim Taleb’s epistemology of risk-taking. In Antifragile, he wrote “You gotta do so if you have an opinion. Live what you say.” He also says that “true virtue is risk-taking”. The idea is that actors with wrong ideas will “die”, metaphorically or literally, and thus good, robust knowledge will survive through the evolution of actors carrying these ideas: people, companies, and countries.
The source of the problem is the energetic cost of updating mental and embodied models
Here, I propose an explanation of epistemic irresponsibility in terms of Active Inference.
The expected free energy formula includes only the “information gain” associated with the basic discriminative model and the “pragmatic value” associated with the basic generative model of the agent. An Active Inference agent strives to minimise the expected free energy through their model updates and actions.
However, if the expected free energy formula elaborates on model hierarchy and tries to capture model learning (aka meta-learning), it should include extra terms. First, it should include the information gain on the model parameters (or model structure) updates: see chapter 7, section 7.5 “Learning and Novelty” in Active Inference. Second, the costs of updating a model should be taken into account, and they can be large. There is a neurological cost: everyone knows that “unlearning” something can be even harder than learning it in the first place, as well as energetic and economic costs in the external world, due to the fact that a person has arranged something in their life (which can be seen as their extended phenotype) according to their models: e. g., they started a business, believing in the prospects of a certain idea, or moved to a certain place, believing that the climate of that place would be good to their health. Undoing this and rearranging one’s life is costly.
I believe more formal variations on this explanation can be found in “An information-theoretic perspective on the costs of cognition” (Zenon et al. 2019) and “Resource-rational analysis: Understanding human cognition as the optimal use of limited computational resources” (Lieder and Griffiths 2019).
There have also been explanations of epistemic irresponsibility in terms of cognitive biases, or psychology, e. g. that epistemic irresponsibility is a sort of ego-protecting reaction of people to a question like “Did you consider taking a course in rationality?”, which may be perceived as an attack or an insult, because it may imply that the addressed person is irrational. However, I suspect that these explanations are less fundamental than those in terms of the energetic costs of updating one’s mental and embodied models.
Overcoming epistemic irresponsibility takes discipline
Thus, if the explanation is correct, the agent should observe the violation of mind- and embodiment-agnostic normative logic, and that it is sometimes rational to keep acting guided by outdated models, which sometimes yield incorrect predictions, because the energetic cost of updating the model immediately may be higher than the cost entailed by the incorrect predictions of the model, at least, temporarily.
For example, if a person who is an astrology teacher is enlightened that astrology is a pseudoscience, they may decide that it’s rational to at least complete an astrology workshop they are currently conducting. If they announce that the workshop is cancelled immediately, they will need to return all the money already collected from workshop participants, and also bear various extra costs, from dealing with legal issues to suffering the psychological discomfort of explaining their behaviour to the workshop participants. Note that we may consider the decision of the astrology teacher (to keep the workshop going) rational but unethical. Including or excluding ethics in the scope of rationality is a matter of terminological convention.
Knowing the root cause of epistemic irresponsibility and reflecting on the cases when one holds “temporarily irrational” beliefs is not enough to overcome this disease in oneself. One should also have a discipline to regularly revisit these problematic places in one’s models and decide when the energetic cost of holding an inaccurate belief starts to outweigh the cost of updating it. Such a discipline is reminiscent of the “habit of remembering asking questions (including What of it?)”, and the “trained ability to answer them precisely and accurately”, which Adler wrote about in How to Read a Book.
This “recipe” for overcoming epistemic irresponsibility doesn’t help to break the circular nature of this problem and thus solve the “rationality outreach challenge”, because the explanation itself requires some understanding of logic, ontology, epistemology, and cognitive science. Epistemic irresponsibility holds people from learning the requisite bits of these knowledge domains in the first place.
Epistemic irresponsibility of companies, communities, societies, and the civilisation
We can think of epistemic irresponsibility applying not only on the personal level but also at higher system levels. Whether or not companies, communities, societies, and the civilisation should be considered agents, it seems reasonable to at least consider them cognitive entities or “information processing units”, and thus subjects to normative logic and rationality.
I would speculate that the societal and political hurdles to addressing climate change (such as climate change denial), which started 40-50 years ago and continue to this day, are an example of epistemic irresponsibility on the societal and civilisational levels.
How to crack the circular nature of epistemic irresponsibility?
I haven’t researched this topic and don’t have any ideas off the top of my head for how to motivate people to study normative logic and rationality, if they respond “I don’t need it, I’m good with where I am” to polite suggestions or advertisements.
I would be grateful if readers shared their experiences or the strategies used at CFAR and similar programs to deal with this challenge.