Mmmm I'm glad you've written this up. It feels to me like the first half of what needs to be done with the concepts of System 1 and System 2, which is dissolving the original meanings (from Kahneman etc).
The second half is... inasmuch as these concepts have become popular, what experiences are people using them to point at? It seems to me that there are actual multiple things happening here. Among rationalists, "System 2" is used to refer to thinking that is verbal, explicit, logical, procedural, linear. "System 1" is kind of a catch-all for... most of the rest of thinking.
There's a much more obvious mapping for this in the brain, but...
...it's unfashionable to talk about.
It's the left and right hemispheres. The left hemisphere indeed operates in a manner that is verbal, explicit, logical, procedural, linear. The right hemisphere operates in a more intuitive, parallel, context-sensitive, uniqueness-oriented way. (I'm summarizing here, but this is based on extensive reading of an excellent meta-analysis of decades of research, that notes that the classic notion of left/right as reason/emotion isn't accurate but that there's something really important going on.)
If you squint, you can kind of imagine the Right hemisphere as Type 1 and the Left hemisphere as Type 2 Processing, but it's importantly different. Neither hemisphere is necessarily automatic or faster or more abstract than the other. Both are, in different senses, conscious, although it seems possible that many people primarily identify with their left hemisphere consciousness.
To illustrate this difference: many of the examples in this post describe categorization ("count the letter a", "this voice, not that voice", etc) as being a Type 1 activity, but categorization is primarily a feature of the Left hemisphere's world. "How many people are in this room right now?" requires this categorization in order to answer. People are reduced to fungible, countable objects; if one exited and another entered, the total doesn't change. This is useful! "What's the mood of this room?" is much harder to answer with this sort of categorization applied. It's a much better question for a Right hemisphere, which can look at the whole scene and capture it in a metaphor or a few adjectives.
So when a rationalist is saying "My S1 has this intuition, but I can't explain it to my System 2", then one thing they might mean is that their Right hemisphere understands something, but is unable to articulate it in a way that their Left hemisphere could understand verbally & linearly and in terms of the concepts and categories that are familiar to it.
The other thing they might mean, however, is not a left-right distinction but a top-bottom or front-back distinction. This is what would be meant in the context of "My System 2 knows not all dogs are scary, but my System 1 still afraid of them since I got bitten when I was 7." This is much better modelled not with S2/S1, but by talking about the neocortex as opposed to the limbic system.
The process that instantly computes 2+2=4 is very different from the process that releases a bunch of adrenaline at the sound of a bark. Are they both Type 1? Seems so, but I don't know this model well enough.
My impression is that the left neocortex (which many rationalists identify as their "System 2") tends to categorize everything else as well, everything else. And this gets labelled "System 1". It makes sense to me that the left neocortex would treat everything else as a big blob, because it first & foremost makes a categorical distinction between its coherent articulable worldview (which it trusts at a level before awareness) and, well, everything else.
I've been playing with this model for a few months, and so far have found a couple clear distinctions between "right hemisphere" and "emotional brain / limbic system / amygdala etc".
One is about simplicity & complexity:
Another is that a right hemisphere intuition tends to know that it can't articulate itself and have a sort of relaxed confidence about that, whereas if something is stimulating emotion then you can easily generate all sorts of explanations (which may be accurate or confabulations) for why your "intuition" is correct.
I've found the recent reading I've done on the brain hemispheres, of Iain McGilchrist's work, to be one of the most important books I've read in my life. His model of hemispheres has profound implications in relation to everything from rationalization & motivated reasoning, to social dynamics & coordination, to technological development & science, to AI Safety & corrigibility.
I think if everyone here had it as a common framework, we'd be able to talk more fruitfully about many of the major tensions that have shown up on LessWrong on the past few years regarding post-rationality/meta-rationality, kenshō/Looking/etc, and more.
Links to learn more:
(I may edit this into a top-level at some point; for now I'm going to leave it as a comment)
I've seen mentions of McGilchrist's work pop up every now and then, but I'm still unclear on what exactly his model adds. In particular, I'm a little cautious of the illusion of understanding that may pop up whenever people just add neuroscience terms to an explanation which wouldn't really need them.
E.g. "seeing a dog makes you feel afraid" versus "seeing a dog causes your amygdala to sound an alarm, making you feel afraid". The second sentence basically just names a particular part of the brain which is involved in the fear response. This isn't really a piece of knowledge that most people would do anything with (and it should hopefully have been obvious without saying that some part of the brain is involved in the response), but it can still feel like it had substantially more information.
A lot of the summaries of McGilchrist that I've seen so far raise similar alarms for me. You suggest that if everyone had his model as a common framework, then talking about various things would become more productive. But as far as I can tell, your description of his work just associates various mental processes with different parts of the brain. What's the benefit of saying "the verbal and explicit mode of thought, which is associated with the left hemisphere", as opposed to just "the verbal and explicit mode of thought"?
Here's one example of a benefit: the left hemisphere is known to have major blindspots, that aren't implied simply by saying "the verbal and explicit mode of thought." Quoting McGilchrist (not sure about page number, I'm looking at Location 5400 in the 2nd edition on Kindle) describing some tests done by temporarily deactivating one hemisphere then the other, in healthy individuals:
Take the following example of a syllogism with a false premise:
- Major premise: all monkeys climb trees;
- Minor premise: the porcupine is a monkey;
- Implied conclusion: the porcupine climbs trees.
Well — does it? As Deglin and Kinsbourne demonstrated, each hemisphere has its own way of approaching this question. At the outset of their experiment, when the intact individual is asked "Does the porcupine climb trees?", she replies (using, of course, both hemispheres): "It does not climb, the porcupine runs on the ground; it's prickly, it's not a monkey." [...] During experimental temporary hemisphere inactivations, the left hemisphere _of the very same individual (with the right hemisphere inactivated) replies that the conclusion is true: "the porcupine climbs trees since it is a monkey." When the experimenter asks, "But is the porcupine a monkey?", she replies that she knows it is not. When the syllogism is presented again, however, she is a little nonplussed, but replies in the affirmative, since "That's what is written on the card." When the right hemisphere of the same individual (with the left hemisphere inactivated) is asked if the syllogism is true, she replies: "How can it climb trees — it's not a monkey, it's wrong here!" If the experimenter points out that the conclusion must follow from the premises stated, she replies indignantly: "But the porcupine is not a monkey!"
In repeated situations, in subject after subject, when syllogisms with false premises, such as "All trees sink in water; balsa is a tree; balsa wood sinks in water," or "Northern lights are often seen in Africa; Uganda is in Africa; Northern lights are seen in Uganda", are presented, the same pattern emerges. When asked if the conclusion is true, the intact individual displays a common sense reaction: "I agree it seems to suggest so, but I know in fact it's wrong." The right hemisphere dismisses the false premises and deductions as absurd. But the left hemisphere sticks to the false conclusion, replying calmly to the effect that "that's what it says here."
In the left-hemisphere situation, it prioritizes the system, regardless of experience: it stays within the system of signs. Truth, for it, is coherence, because for it there is no world beyond, no Other, nothing outside the mind, to correspond with. "That's what it says here." So it corresponds with itself: in other words, it coheres. The right hemisphere prioritises what it learns from experience: the real state of existing things "out there". For the right hemisphere, truth is not mere coherence, but correspondence with something other than itself. Truth, for it, is understood in the sense of being "true" to something, faithfulness to whatever it is that exists apart from ourselves.
However, it would be wrong to deduce from this that the right hemisphere just goes with what is familiar, adopting a comfortable conformity with experience to date. After all, one's experience to date might be untrue to reality: then paying attention to logic would be an important way of moving away from from false customary assumption. And I have emphasized that it is the right hemisphere that helps us to get beyond the inauthentically familiar. The design of the above experiment specifically tests what happens when one is forced to choose between two paths to the truth in answering a question: using what one knows from experience or following a syllogism where the premises are blatantly false. The question was not whether the syllogism was structurally correct, but what actually was true. But in a different situation, where one is asked the different question "Is this syllogism structurally correct?", even when the conclusion flies in the face of one's experience, it is the right hemisphere which gets the answer correct, and the left hemisphere which is distracted by the familiarity of what it already thinks it knows, and gets the answer wrong. The common thread here is the role of the right hemisphere as "bullshit detector". In the first case (answering the question "What is true here?") detecting the bullshit involves using common sense. In the second case (answering "Is the logic here correct?"), detecting the bullshit involves resisting the obvious, the usual train of thought.
For me personally, working with McGilchrist's model has dramatically improved my own internal bullshit-detection capacity. I've started to be able to sometimes smell the scent of rationalizations, even while the thoughts I'm having continue to feel true. This has been helpful for learning, for noticing when I'm being a jerk in relationships, and for noticing how I'm closing myself off to some line of thinking while debugging my code.
And the bullshit detection thing is just one element of it. The book relates dozens of other case studies on differences in way-of-being-and-perceiving of each hemisphere, and connects them with some core theories about the role of attention in cognition.
If you were surprised in reading this comment to discover that it's not the left hemisphere that is best at syllogisms, then I would like to suggest there are important things that are known about the brain that you could learn by reading this book, that would help you think more effectively. (This also applies if you were not-particularly-surprised because your implicit prior was simply "hemispheres are irrelevant"; I was more in this camp.)
But in a different situation, where one is asked the different question "Is this syllogism structurally correct?", even when the conclusion flies in the face of one's experience, it is the right hemisphere which gets the answer correct, and the left hemisphere which is distracted by the familiarity of what it already thinks it knows, and gets the answer wrong.
Wait what? Surely the left/right hemispheres are accidentally reversed here? Or is the book saying that the left hemisphere always answers incorrectly, no matter what question you ask?
The book is saying that the left hemisphere answers incorrectly, in both cases! As I said, this is surprising.
I haven't looked at the original research and found myself curious what would happen with a syllogism that is both invalid and has a false conclusion. My assumption is that either hemisphere would reject something like this:
The left hemisphere seems to be where most of motivated cognition lives. If you've heard the bizarre stories about patients confabulating after strokes (eg "my limb isn't paralyzed, I just don't want to move it) this is almost unilaterally associated with damage to the right hemisphere. Many people, following Gazzinga's lead, seem to have assumed this was just because someone with a left hemisphere stroke can't talk, but if you leave words aside, it is apparent that people with left hemisphere damage are distressed about their paralyzed right arm, whereas people with right hemisphere damage are often in denial.
Likewise, part of the job of a well-functioning left hemisphere is to have blindspots. It's so zoomed in on whatever it's focused on that the rest of the world might as well not exist. If you've heard of the term "hemispatial neglect", that leads to people shaving only half of their face, eating only half of their plate, or attempting to copy a drawing of an ordinary clock and ending up drawing something like this:
...then that's again something that only happens when the left hemisphere is operating without the right (again, can also be shown in healthy patients by temporarily deactivating one hemisphere). The left hemisphere has a narrow focus of attention and only on the right side of things, and it doesn't manage to even notice that it has omitted the other half, because as far as it's concerned, the other half isn't there. This is not a vision thing—asked to recall a familiar scene, such a patient may describe only the right half of it.
The book is saying that the left hemisphere answers incorrectly, in both cases! As I said, this is surprising.
That's not just surprising, that's absurd. I can absolutely believe the claim that the left hemisphere always takes what's written for granted, and solves the syllogism formally. But the claim here is that the left hemisphere pays careful attention to the questions, solves them correctly, and then reverses the answer. Why would it do that? No mechanism is proposed.
I looked at the one paper that's mentioned in the quote (Deglin and Kinsbourne), and they never ask the subjects whether the syllogisms are 'structurally correct'; they only ask about the truth. And their main conclusion is that the left hemisphere always solves syllogisms formally, not that it's always wrong.
If you've heard the bizarre stories about patients confabulating after strokes (eg "my limb isn't paralyzed, I just don't want to move it) this is almost unilaterally associated with damage to the right hemisphere.
Interesting, I didn't know this only happened with the left hemisphere intact.
the claim here is that the left hemisphere pays careful attention to the questions, solves them correctly, and then reverses the answer.
Fwiw I also think that that is an absurd claim and I also think that nobody is actually claiming that here. The claim is something more like what has been claimed about System 1, "it takes shortcuts", except in this case it's roughly "to the left hemisphere, truth is coherence; logical coherence is preferred before map coherence, but both are preferred to anything that appears incoherent."
I looked up the source for the "However" section and it's not Deglin and Kinsbourne but Goel and Dolan, 2003). I looked it up and found it hard to read but my sense is that what it's saying is:
An general, people are worse at answering the validity of a logical syllogism when it contradicts their beliefs. (This should surprise nobody.)
Different parts of the brain appear to be recruited depending on whether the content of a syllogism is familiar:
A recent fMRI study (Goel, Buchel, Frith & Dolan, 2000) has provided evidence that syllogistic reasoning is implemented in two distinct brain systems whose engagement is primarily a function of the presence or absence of meaningful content. During content-based syllogistic reasoning (e.g. All apples are red fruit; All red fruit are poisonous;[All apples are poisonous) a left hemisphere frontal and temporal lobe system is recruited. By contrast, in a formally identical reasoning task with arbitrary content (e.g. All A are B; All B are C;[All A are C) a bilateral parietal system is recruited.
(Note: this is them analyzing what part of the brain is recruited when the task is completed successfully.)
I think what's being implied by McGilchrist citing this paper (although it's unclear to me if this was tested as directly as the Deglin & Kinsbourne study) is that without access to the right hemisphere, the left hemisphere's process would be even more biased, or something.
I'd be interested in your take if you read the 2000 or 2003 papers.
McGilchrist himself has said that it doesn't matter if the neuroscience is all wrong, it makes a good metaphor. See this review, where McGilchrist's "The Master and His Emissary" is quoted:
“If it could eventually be shown…that the two major ways, not just of thinking, but of being in the world, are not related to the two cerebral hemispheres, I would be surprised, but not unhappy. Ultimately what I have tried to point to is that the apparently separate ‘functions’ in each hemisphere fit together intelligently to form in each case a single coherent entity; that there are, not just currents here and there in the history of ideas, but consistent ways of being that persist across the history of the Western world, that are fundamentally opposed, though complementary, in what they reveal to us; and that the hemispheres of the brain can be seen as, at the very least, a metaphor for these…
What [Goethe’s Faust, Schopenhauer, Bergson, Scheler and Kant] all point to is the fundamentally divided nature of mental experience. When one puts that together with the fact that the brain is divided into two relatively independent chunks which just happen broadly to mirror the very dichotomies that are being pointed to – alienation versus engagement, abstraction versus incarnation, the categorical versus the unique, the general versus the particular, the part versus the whole, and so on – it seems like a metaphor that might have some literal truth. But if it turns out to be ‘just’ a metaphor, I will be content. I have a high regard for metaphor. It is how we come to understand the world.”
In which case, why is he peddling it? He is asserting the neuroscience as true. It matters whether it is true, because without it, he's just another purveyor of intellectual artistic ramblings, like the ones he admires. And isn't that dichotomising, in his terms, a left-brain thing to do?
I think that pretty much cuts the ground from under his whole system. It reduces the neuroscience story to a noble lie.
Brief reply about dog thing & just naming a part of the brain—I agree!
But saying "system 1" is also not useful unless you have a richer map of how system 1 works. In order for the emotional brain model to be useful, you need affordances for working with it. I got mine from the Bio-Emotive Framework, and since learning that model & technique, I've been more able to work with this stuff, whatever you want to call it, and part of working with it involves a level of identifying what's going on at a level of detail beyond "S1". There are also of course methods of working with this stuff that don't require such a framework!
I'm appreciating you pointing this out, since it represents a way in which my comment was unhelpful—I didn't actually give people these richer models, I just said "there are models out there that I've found much better than S1/S2 for talking about the same stuff". I've pitched the models, but not actually sharing much of why I find them so useful. Although I've just added a long comment elaborating on some hemisphere stuff so hopefully that helps.
Curated.
System 1 and 2 have gotten a lot of attention on LessWrong over the years, informing many explicit models as well as vague folk-theories circulating the community. Given that the originator of the concept has deprecated it, it seems particularly important for the LW community to have common knowledge of the state of the literature.
I found this post super useful for clearing out old bad ideas how thought works and building a better model in its place, which is now thoroughly part of my worldview. I talk about this post and link to it constantly. As one concrete example, in my Can You Get AGI From a Transformer, there's a spot where I link to this post. It doesn't take up a lot of space in the post, but it's carrying a lot of weight behind the scenes.
The S1/S2 dichotomy has proven very unhelpful for me.
This post (along with the rest of Kaj's amazing sequence) was an crucial nudge away from the S1/S2 frame and towards a way more gearsy model of the mind.
I always understood bias to mean systematic deviations from the correct response (as in the bias-variance decomposition [1], e.g. a bias to be more overconfident, or the bias of being anchored to arbitrary numbers). I read your and Evans' interpretation of it more like bias meaning incorrect in some areas. As Type 2 processing seems to be very flexible and unconstrained, I thought that it might not necessarily be biased but simply sufficiently unconstrained and high variance to cause plenty of errors in many domains.
[1] https://miro.medium.com/max/2567/1*CgIdnlB6JK8orFKPXpc7Rg.png
PS: Thanks for your writing, I really enjoy it a lot.
Thank you for such a well-structured and concise summary of the research! I really like this sequence.
Pretty interesting to see where all of that could lead from the evolutionary (working memory / type 2 processes in animals) and from the mental disorder perspective.
I finished "Thinking fast and slow" a few days ago, and I have remarks.
1) You say
The terms “System 1 and System 2” suggest just that: two distinct, clearly defined systems with their own distinctive properties and modes of operation.
Kahneman is aware of that and tries to prevent it. A quote from the conclusion of the book :
And of course you also remember that the two systems do not really exist in the brain or anywhere else. “System 1 does X” is a shortcut for “X occurs automatically.” And “System 2 is mobilized to do Y” is a shortcut for “arousal increases, pupils dilate, attention is focused, and activity Y is performed.”
2) Either there is something I don't understand ir the examples taken by Melnikoff & Bargh (2018) seem to lack charity. In the bat-and-ball problem, the task of "computing the price of the ball by substracting the total cost and the difference between the two prices" is unintentional, there is not a moment when it was decided. This is what is meant by "the task is performed unintentionnaly". Similarly, in the availability heuristic the task "estimate whether more words begin with the letter K than have K in the third position by assessing how many words with each I can remember" is not intentional. The substitution, attributed to System 1, is not intentional.
3) I don't know if the quotes from Evans (2012) are directed toward Kahneman, but "why should [type 2] reasoning necessarily be normatively correct? [...] why should type 1 processes that operate automatically and without reflection necessarily be wrong ?" has nothing to do with what the book says, and you explain quite well why System 1 will make more mistakes. I don't see what this adds to the rest
4) Similarly, there are two chapters of Thinking Fast and Slow (21 and 22) dedicated to the good and bad of expert intuitions, so the remark about "there is much evidence that expert decision making can often be well served by intuitive rather than reflective thinking" seems out of place
I found your text interesting and quite informative but I don't want people to have the idea that Thinking Fast and Slow is letting important things like this slip.
Under this model, then, Type 2 processing is a particular way of chaining together the outputs of various Type 1 subagents using working memory. Some of the processes involved in this chaining are themselves implemented by particular kinds of subagents.
Something I have encountered in my own self-experiments and tinkering is Type 2 processes that chain together other Type 2 processes (and often some Type 1 subagents as well), meshing well with persistent Type 2 subagents that get re-used due to their practicality and sometimes end up resembling Type 1 subagents as their decision process becomes reflexive to repeat.
Have you encountered anything similar?
Probably, but this description is abstract enough that I have difficulty generating examples. Do you have a more concrete example?
The most notable example of a Type 2 process that chains other Type 2 processes as well as Type 1 processes is my "path to goal" generator, but as I sit here to analyze it I am surprised to notice that much of what used to be Type 2 processing in its chain has been replaced with fairly solid Type 1 estimators with triggers for when you leave their operating scope. I am noticing that what I thought started as Type 2s that call Type 2s now looks more like Type 2s that set triggers via Type 1s to cause other Type 2s to get a turn on the processor later. It's something of an indirect system, but the intentionality is there.
My visibility into the current intricacies of my pseudo-IFS is currently low due to the energy costs maintaining such visibility produces, and circumstances do not make regaining it feasible for a while. As a result, I find myself having some difficulty identifying any specific processes that are Type 2 that aren't super implementation-specific and vague on the intricacies. I apologize for not having more helpful details on that front.
I have something a bit clearer as an example of what started as Type 2 behavior and transitioned to Type 1 behavior. I noticed at one point that I was calculating gradients in a timeframe that seemed automatic. Later investigation seemed to suggest that I had ended up with a Type 1 estimator that could handle a number of common data forms that I might want gradients of (it seems to resemble Riemann sums), and I have something of a felt sense for whether the form of data I'm looking at will mesh well with the estimator's scope.
At least Type 2 behavior turning into Type 1 behavior is a pretty common thing in skill learning; the classic example I've heard cited is driving a car, which at first is very effortful and requires a lot of conscious thought, but then gradually things get so automated that you might not even remember most of your drive home. But the same thing can happen with pretty much any skill; at first it's difficult and requires Type 2 processing, until it's familiar enough to become effortless.
I haven't read Stanovichs' papers you refer to, but in his book "Rationality and the reflective mind" he proposes a seperation of Type 2 processing into 1) serial associative cognition with a focal bias and 2) fully decoupled simulations for alternative hypothesis. (Just noting it because I found it useful for my own thinking.)
In fact, an exhaustive simulation of alternative worlds would guarantee correct responding in the [Wason selection] task. Instead [...] subjects accept the rule as given, assume it is true, and simply describe how they would go about verifying it. They reason from a single focal model— systematically generating associations from this focal model but never constructing another model of the situation. This is what I would term serial associative cognition with a focal bias. It is how I would begin to operationalize the satisficing bias in Type 2 processing posited in several papers by Evans (2006b; Evans, Over, & Handley, 2003). "
The terms System 1 and System 2 were originally coined by the psychologist Keith Stanovich and then popularized by Daniel Kahneman in his book Thinking, Fast and Slow. Stanovich noted that a number of fields within psychology had been developing various kinds of theories distinguishing between fast/intuitive on the one hand and slow/deliberative thinking on the other. Often these fields were not aware of each other. The S1/S2 model was offered as a general version of these specific theories, highlighting features of the two modes of thought that tended to appear in all the theories.
Since then, academics have continued to discuss the models. Among other developments, Stanovich and other authors have discontinued the use of the System 1/System 2 terminology as misleading, choosing to instead talk about Type 1 and Type 2 processing. In this post, I will build on some of that discussion to argue that Type 2 processing is a particular way of chaining together the outputs of various subagents using working memory. Some of the processes involved in this chaining are themselves implemented by particular kinds of subagents.
This post has three purposes:
Let’s start with looking at some criticism of the S1/S2 model endorsed by the person who coined the terms.
What type 1/type 2 processing is not
The terms “System 1 and System 2” suggest just that: two distinct, clearly defined systems with their own distinctive properties and modes of operation. However, there’s no single “System 1”: rather, a wide variety of different processes and systems are lumped together under this term. It is also unclear whether there is any single System 2, either. As a result, a number of researchers including Stanovich himself have switched to talking about “Type 1” and “Type 2” processing instead (Evans, 2012; Evans & Stanovich, 2013; Pennycook, Neys, Evans, Stanovich, & Thompson, 2018).
What exactly defines Type 1 and Type 2 processing?
A variety of attributes have been commonly attributed to either Type 1 or Type 2 processing. However, one criticism is that there is no empirical or theoretical support for such attributes to only occur with one type of processing. For instance, Melnikoff & Bargh (2018) note that one set of characteristics which has been attributed to Type 1 processing is “efficient, unintentional, uncontrollable, and unconscious”, whereas Type 2 processing has been said to be “inefficient, intentional, controllable and conscious”.
(Before you read on, you might want to take a moment to consider the extent to which this characterization matches your intuition of Type 1 and Type 2 processing. If it does match to some degree, you can try to think of examples which are well-characterized by these types, as well as examples which are not.)
.
.
.
.
.
.
.
.
.
.
They note that this correlation has never been empirically examined, and that there are also various processes in which attributes from both sets co-occur. For example:
Regarding the last example, Melnikoff & Bargh note:
They also give examples of what they consider uncontrollable (T1) but inefficient (T2), unintentional (T1) but inefficient (T2), as well as unintentional (T1) but controllable (T2). Further, they discuss each of the four attributes themselves and point out that they all contain various subdimensions. For example, people whose decisions are influenced by unconscious primes are conscious of their decision but not of the influence from the prime, meaning that the process has both conscious and unconscious aspects.
Type 1/Type 2 processing as working memory use
Rather than following the “list of necessary attributes” definition, Evans & Stanovich (2013) distinguish between defining features and typical correlates. In previous papers, Evans has generally defined Type 2 processing in terms of requiring working memory resources and being able to think hypothetically. On the other hand, Stanovich has focused on what he calls cognitive decoupling, which his work shows is highly correlated with fluid intelligence as the defining feature.
Cognitive decoupling can be defined as the ability to create copies of our mental representations of things, so that the copies can be used in simulations without affecting the original representations. For example, if I see an apple in a tree, my mind has a representation of the apple. If I then imagine various strategies of getting the apple - such as throwing a stone at the tree to knock the apple down - I can mentally simulate what would happen to the apple as a result of my actions. But even as I imagine the apple falling down from the tree, I never end up thinking that I can get the real apple down simply by an act of imagination. This because the mental object representing the real apple is decoupled from the apple in my hypothetical scenario. I can manipulate the apple in the hypothetical without those manipulations being passed on to the mental object representing the original apple.
In their joint paper, Evans & Stanovich propose to combine their models and define Type 2 processes as those which use working memory resources (closely connected with fluid intelligence) in order to carry out hypothetical reasoning and cognitive decoupling. In contrast, Type 1 reasoning is anything which does not do that. Various features of thought - such as being automatic and the other controlled - may tend to correlate more with one or the other type, but these are only correlates, not necessary features.
Type 2 processing as composed of Type 1 components
In previous posts of my multi-agent minds sequence, I have been building up a model of mind that is composed of interacting components. How does it fit together with the proposed Type 1/Type 2 model?
Kahneman in Thinking Fast and Slow mentions that giving the answer to 2 + 2 = ? is a System (Type) 1 task, whereas calculating 17 * 24 is a System (Type) 2 task. This might be starting to sound familiar. In my post on subagents and neural Turing machines, I discussed Stanislas Dehane’s model where you do complex arithmetic by breaking up a calculation into subcomponents which can be done automatically, and then routing the intermediate results through working memory. You could consider this to also involve cognitive decoupling: for instance, if part of how you calculate 17 * 24 is by first noting that you can calculate 10 * 24, you need to keep the original representation of 17 * 24 intact in order to figure out what other steps you need to take.
To me, the calculation of 10 * 24 = 240 happens mostly automatically; like 2 + 2 = 4, it feels like a Type 1 operation rather than a Type 2 one. But what this implies, then, is that we carry out Type 2 arithmetic by chaining together Type 1 operations through Type 2 working memory.
I do not think that this is just a special case relating to arithmetic. Rather it seems like an implication of the Evans & Stanovich definition which they do not mention explicitly, but which is nonetheless relatively straightforward to draw: that Type 2 reasoning is largely built up of Type 1 components.
Under this interpretation, there are some components which are specifically dedicated to Type 2 processes: things like working memory storages and systems for manipulating their contents. But those components cannot do anything alone. The original input to be stored in working memory originates from Type 1 processes (and the act of copying it to working memory decouples it from the original process which produced it), and working memory alone could not do anything without those Type 1 inputs.
Likewise, there may be something like a component which is Type 2 in nature, in that it holds rules for how the contents of working memory should be transformed in different situations - but many of those transformations happen by firing various Type 1 processes which then operate on the contents of the memory. Thus, the rules are about choosing which Type 1 process to trigger, and could again do little without those processes. (My post on neural Turing machines explicitly discussed such rules.)
Looking through Kahneman’s examples
At this point, you might reasonably suspect that arithmetic reasoning is an example that I cherry-picked to support my argument. To avoid this impression, I’ll take the first ten examples of System 2 operations that Kahneman lists in the first chapter of Thinking, Fast and Slow and suggest how they could be broken down into Type 1 and Type 2 components.
Kahneman defines System 2 in a slightly different way than we have defined Type 2 operations - he talks about System 2 operations requiring attention - but as attention and working memory are closely related, this still remains compatible with our model. Most of these examples involve somehow focusing attention, and manipulating attention can be understood as manipulating the contents of working memory to ensure that a particular mental object remains in working memory. Modifying the contents of working memory was an important type of production rule discussed in my earlier post.
Starting with the first example in Kahneman’s list:
One tries to keep their body in such a position that it will be ready to run when the gun sounds; recognizing the feel of the correct position is a Type 1 operation. Type 2 rules are operating to focus attention on the output of the system which outputs proprioceptive data, allowing Type 1 processes to notice mismatches with the required body position and correct them. Additionally, Type 2 rules are focusing attention on the sound of the gun, so as to more quickly identify the sound when the gun fires (a Type 1 operation), causing the person to start running (also a Type 1 operation).
This involves Type 2 rules which focus attention on a particular sensory output, as well as keeping one’s eyes physically oriented towards the clowns. This requires detecting when one’s attention/eyes are on something else than the clowns and then applying an internal (in the case of attention) or external (in the case of eye position) correction. As Kahneman offers “orient to the source of a sudden sound”, “detect hostility in a voice”, “read words on large billboards”, and “understand simple sentences” as Type 1 operations, we can probably say that recognizing something as a clown or not-clown and moving one’s gaze accordingly are Type 1 operations.
As above, Type 2 rules check whether attention is on the voice of that person (a comparison implemented using a Type 1 process), and then adjust focus accordingly.
Similar to the clown example.
It’s unclear to me exactly what is going on here. But introspectively, this seems to involve something like keeping the sound in attention so as to feed it to memory processes, and then applying the rule of “whenever the memory system returns results, compare them against the sound and adjust the search based on how relevant they seem”. The comparison feels like it is done by something like a Type 1 process.
Walking: Similar to the “brace for the starter gun” example, Type 2 rules keep calling for a comparison of your current walking speed with the desired one (a Type 1 operation), passing any corrections resulting from that comparison to the Type 1 system controlling your walking speed.
Social behavior: maintain attention on a conscious representation of what you are doing, checking it against various Type 1 processes which contain rules about appropriate and inappropriate behavior. Adjust or block accordingly.
Focus attention on the letters of a text; when a Type 1 comparison detects the letter “a”, increment a working memory counter by one.
After a retrieval of the phone number from memory has been initiated, Type 2 rules use Type 1 processes to monitor that it is said in full.
Keeping attention focused on what you are doing to allow a series of evaluations, mental simulations, and cached (Type 1) procedural operations determining how to act in response to a particular situation in the parking process.
A general pattern in these examples is that Type 2 processing can maintain attention on something as well as hold the intention to invoke comparisons to use as the basis for behavioral adjustments. As comparisons involve Type 1 processes, Type 2 processing is fundamentally reliant on Type 1 processing to be able to do anything.
Consciousness and dual process theory
Alert readers might have noticed that focusing one’s attention on something involves keeping it in consciousness, whereas the previous Evans & Stanovich definition noted that consciousness is not a defining part of the Type 1/Type 2 classification. Is this a contradiction? Probably not, since as remarked previously, different aspects of the same process may be conscious and unconscious at the same time.
For example, if one intends to say something, one may be conscious of the intention while the actual speech production happens unconsciously; once they say it and they hear their own words, an evaluation process can run unconsciously but output its results into consciousness. With “conscious” being so multidimensional, it doesn’t seem like a good defining characteristic to use, even if some aspects of it did very strongly correlate with Type 2 processing.
Evans (2012) writes in a manner which seems to me compatible with the notion of there being many different kinds of Type 2 processing, with different processing resources being combined according to different rules as the situation warrants:
If you read my neural Turing machines post, you may recall that I noted that the rules which choose what becomes conscious operate below the level of conscious awareness. We may have the subjective experience of being able to choose what thoughts we think, but this is a post-hoc interpretation rather than a fact about the process.
Type 1/Type 2 and bias
People sometimes refer to Type 1 reasoning as biased, and to Type 2 reasoning as unbiased. But as this discussion should suggest, there is nothing that makes one of the two types intrinsically more or less biased than the other. The bias-correction power of Type 2 processing emerges from the fact that if Type 1 operations are known to be erroneous and a rule-based procedure for correcting them exists, a Type 2 operation can be learned which implements that rule.
For example, someone familiar with the substitution principle may know that their initial answer to a question like “how popular will the president be six months from now?” comes from a Type 1 process which actually answered the question of “how popular is the president right now?”.
They may then have a Type 2 rule saying something like “when you notice that the question you were asked is subject to substitution effects, replace the initial answer with one derived from a particular procedure”. But this still requires a) a Type 1 process recognizing the situation as one where the rule should be applied b) knowing a procedure which provides a better answer c) the cue-procedure rule having been installed previously, itself a process requiring a number of Type 1 evaluations (about e.g. how rewarding it would be to have such a rule in place).
There is nothing to say that somebody couldn’t learn an outright wrong Type 2 rule, such as “whenever you think of 2+2 = 4, substitute your initial answer of ‘4’ with a ‘5’”.
Often, it is also unclear of what the better Type 2 rule even should be. For instance, another common substitution effect is that when someone is asked “How happy are you with your life these days?”, they actually answer the question of “What is my mood right now?”. But what is the objectively correct procedure for evaluating your current happiness with life?
On the topic of Type 1/2 and bias, I give the final word to Evans (2012):
Summary and connection to the multiagent models of mind sequence
In this post, I have summarized some recent-ish academic discussion on dual-process models of thought, or what used to be called System 1 and System 2. I noted that the popular conception of them as two entirely distinct systems with very different properties is mistaken. While there is a defining difference between them - namely, the use of working memory resources to support hypothetical thinking and cognitive decoupling - they seem to rather refer to differences in two types of thought, either of which may use very different kinds of systems.
It is worth noting at this point that there are many different dual-process models in different parts of psychology. The Evans & Stanovich model which I have been discussing here is intended as a generalized model of them, but as they themselves (2013) write:
In previous posts of the multi-agent models of mind sequence, I have been building up a model of the mind being built up of a variety of subsystems (which might in some contexts be called subagents).
In my discussion of Consciousness and the Brain, I summarized some of its conclusions as saying that:
In Building up to an Internal Family Systems model, I used this foundation to discuss the IFS model of how various subagents manipulate consciousness in order to achieve various kinds of behavior. In Subagents, neural Turing machines, thought selection, and blindspots, I talked about the mechanistic underpinnings of this model and how processes like thought selection and firing of production rules might actually be implemented.
What had been lacking so far was a connection between these models and the Type 1/Type 2 typology. However, if we take something like the Evans & Stanovich model of Type 1/Type 2 processing to be true, then it turns out that our discussion has been connected with their model all along. Already in “Consciousness and the Brain”, I mentioned the “neural Turing machine” passing on results from one subsystem to another through working memory. That, it turns out, is the defining characteristic of Type 2 processing - with Type 1 processing simply being any process which does not do that.
Under this model, then, Type 2 processing is a particular way of chaining together the outputs of various Type 1 subagents using working memory. Some of the processes involved in this chaining are themselves implemented by particular kinds of subagents.
References
Evans, J. S. B. T. (2012). Dual process theories of deductive reasoning: facts and fallacies. The Oxford Handbook of Thinking and Reasoning, 115–133.
Evans, J. S. B. T., & Stanovich, K. E. (2013). Dual-Process Theories of Higher Cognition: Advancing the Debate. Perspectives on Psychological Science: A Journal of the Association for Psychological Science, 8(3), 223–241.
Melnikoff, D. E., & Bargh, J. A. (2018). The Mythical Number Two. Trends in Cognitive Sciences, 22(4), 280–293.
Pennycook, G., Neys, W. D., Evans, J. S. B. T., Stanovich, K. E., & Thompson, V. A. (2018). The Mythical Dual-Process Typology. Trends in Cognitive Sciences, 22(8), 667–668.