1.1 Summary & Table of Contents

This is the first of a series of eight blog posts, which I’ll be serializing over the next month or two. (Or email or DM me if you want to read the whole thing right now.) Here’s an overview of the whole series, and then we’ll jump right into the first post!

1.1.1 Summary & Table of Contents—for the whole series

This is a rather ambitious series of blog posts, in that I’ll attempt to explain what’s the deal with consciousness, free will, hypnotism, enlightenment, hallucinations, flow states, dissociation, akrasia, delusions, and more.

The starting point for this whole journey is very simple:

  • The brain has a predictive (a.k.a. self-supervised) learning algorithm.
  • This algorithm builds generative models (a.k.a. “intuitive models”) that can predict incoming data.
  • It turns out that, in order to predict incoming data, the algorithm winds up not only building generative models capturing properties of trucks and shoes and birds, but also building generative models capturing properties of the brain algorithm itself.

Those latter models, which I call “intuitive self-models”, wind up including ingredients like conscious awareness, deliberate actions, and the sense of applying one’s will.

That’s a simple idea, but exploring its consequences will take us to all kinds of strange places—plenty to fill up an eight-post series! Here’s the outline:

  • Post 1 (Preliminaries) gives some background on the brain’s predictive learning algorithm, how to think about the “intuitive models” built by that algorithm, how intuitive self-models come about, and the relation of this whole series to Philosophy Of Mind.
  • Post 2 (Awareness) proposes that our intuitive self-models include an ingredient called “conscious awareness”, and that this ingredient is built by the predictive learning algorithm to represent a serial aspect of cortex computation. I’ll discuss ways in which this model is veridical (faithful to the algorithmic phenomenon that it’s modeling), and ways that it isn’t. I’ll also talk about how intentions and decisions fit into that framework.
  • Post 3 (The Homunculus) focuses more specifically on the intuitive self-model that almost everyone reading this post is experiencing right now (as opposed to the other possibilities covered later in the series), which I call the Conventional Intuitive Self-Model. In particular, I propose that a key player in that model is a certain entity that’s conceptualized as actively causing acts of free will. Following Dennett, I call this entity “the homunculus”, and relate that to intuitions around free will and sense-of-self.
  • Post 4 (Trance) builds a framework to systematize the various types of trance, from everyday “flow states”, to intense possession rituals with amnesia. I try to explain why these states have the properties they do, and to reverse-engineer the various tricks that people use to induce trance in practice.
  • Post 5 (Dissociative Identity Disorder) (a.k.a. “multiple personality disorder”) is a brief opinionated tour of this controversial psychiatric diagnosis. Is it real? Is it iatrogenic? Why is it related to borderline personality disorder (BPD) and trauma? What do we make of the wild claim that each “alter” can’t remember the lives of the other “alters”?
  • Post 6 (Awakening / Enlightenment / PNSE) is a type of intuitive self-model, typically accessed via extensive meditation practice. It’s quite different from the conventional intuitive self-model. I offer a hypothesis about what exactly the difference is, and why that difference has the various downstream effects that it has.
  • Post 7 (Hearing Voices, and Other Hallucinations) talks about factors contributing to hallucinations—although I argue against drawing a deep distinction between hallucinations versus “normal” inner speech and imagination. I discuss both psychological factors like schizophrenia and BPD; and cultural factors, including some critical discussion of Julian Jaynes’s Origin of Consciousness In The Breakdown Of The Bicameral Mind.
  • Post 8 (De-homunculus-ifying Motivation) suggests that the way people normally talk about motivation and goals is excessively tied to the Conventional Intuitive Self-Model (CISM). This leads to muddled thinking even within the CISM, and leads to straight-up nonsense when applied to other intuitive self-models. I offer an alternative framework that I think works better.

1.1.2 Summary & Table of Contents—for this first post in particular

This post will lay groundwork that I’ll be using throughout the series.

  • Section 1.2 uses the fun optical-illusion-type phenomenon of “bistable perception” to illustrate the idea that the cortex learns and stores a space of generative models, and processes incoming data by searching through that space via probabilistic inference. (I’ll use “generative model” and “intuitive model” interchangeably in this series.)
  • Section 1.3 talks about three very different ways that people can judge each other’s intuitive models:
    • “Is the intuitive model real, or fake?” is about honestly reporting your own internal experience, versus lying or misremembering;
    • “Is the intuitive model veridical, or non-veridical?” is about map–territory correspondence: “veridical” means that the intuitive model (map) corresponds to some kind of external truth (territory).
    • “Is the intuitive model healthy, or pathological?” is about a person’s mental health and well-being.
  • Section 1.4 makes the jump from “intuitive model” to “intuitive self-model”: why and how does the brain’s predictive learning algorithm wind up building generative models of aspects of those very same brain algorithms?
  • Section 1.5 is an appendix going through some terminology and assumptions that I use when talking about probabilistic inference and intuitive models.
  • Section 1.6 is another appendix about how this series relates to Philosophy Of Mind. My opinion of Philosophy Of Mind is: I’m against it! Or rather, I’ll say plenty in this series that would be highly relevant to understanding the true nature of consciousness, free will, and so on, but the series itself is firmly restricted in scope to questions that can be resolved within the physical universe (including physics, neuroscience, algorithms, and so on). I’ll leave the philosophy to the philosophers.

1.2 Generative models and probabilistic inference

1.2.1 Example: bistable perception

Here are two examples of bistable visual perception:

Schröder's stairs: “A” can be seen as either closer or farther than “B”. (Source.)
undefined
The Spinning Dancer can be seen as either going clockwise or counterclockwise.

The steps are on the floor or the ceiling. The dancer is spinning clockwise or counterclockwise. You’ll see one or the other. Your mind may spontaneously switch after a while. If it doesn’t, you might be able to switch it deliberately, via hard-to-describe intuitive moves that involve attention control, moving your eyes, maybe blurring your vision at times, or who knows.

(Except that I can’t for the life of me get the dancer to go counterclockwise! I can sometimes get her to go counterclockwise in my peripheral vision, but as soon as I look right at her, bam, clockwise every time. Funny.)

1.2.2 Probabilistic inference

To understand what’s going on in bistable perception, a key is that perceptions involve probabilistic inference.[1] In the case of vision:

  • Your brain has a giant space of possible generative models[2] that map from underlying states of the world (e.g. “there’s a silhouette dancer with thus-and-such 3D shape spinning clockwise against a white background etc.”) to how the photoreceptor cells would send signals into the brain (“this part of my visual field is bright, that part is dark, etc.”)
  • …But oops, that map is the wrong way around, because the brain does not know the underlying state of the world, but it does know what the photoreceptor cells are doing.
  • …So your brain needs to invert that map. And that inversion operation is called probabilistic inference.

Basically, part of your brain (the cortex and thalamus, more-or-less) is somehow searching through the generative model space (which was learned over the course of prior life experience) until it finds a generative model (“posterior”) that’s sufficiently compatible with both what the photoreceptor cells are doing (“data”) and what is a priori plausible according to the generative model space (“prior”).

You might be wondering: How does that process work? Is that part of the brain literally implementing approximate probabilistic inference, or is it effectively doing probabilistic inference via some superficially-different process that converges to it (or has similar behavior for some other reason)? Exactly how is the generative model space learned and updated? …Sorry, but I won’t answer those kinds of questions in this series. Not only that, but I encourage everyone else to also not answer those kinds of questions, because I firmly believe that we shouldn’t invent brain-like AGI until we figure out how to use it safely.

1.2.3 The thing you “experience” is the generative model (a.k.a. “intuitive model”)

A key point (more on which in the next post) is that “experience” corresponds to the active generative model, not to what the photoreceptor cells are doing etc.

You might be thinking: “Wrong! I know exactly what my photoreceptor cells are doing! This part of the visual field is white, that part is black, etc.—those things are part of my conscious awareness! Not just the parameters of a high-level generative model, i.e. the idea that I’m looking at a spinning dancer!”

I have two responses:

  • First of all, you’re almost certainly more oblivious to what your photoreceptor cells are doing than you realize—see the checker shadow illusion, or the fact that most people don’t realize that their peripheral vision has terrible resolution and terrible color perception and makes faces look creepy, or the fact that the world seems stable even as your eyes saccade five times a second, etc.
  • Second of all, if you say “X is part of my conscious experience”, then I can and will immediately reply “OK, well if that’s true, then I guess X is part of your active generative model”. In particular, there’s no rule that says that the generative model must be only high-level-ish things like “dancer” or “chair”, and can’t also include low-level-ish things like “dark blob of such-and-such shape at such-and-such location”.

1.2.4 Explanation of bistable perception

Finally, back to those bistable perception examples from §1.2.1 above. These are constructed such that, when you look at them, there are two different generative models that are about equally good at explaining the visual data, and also similarly plausible a priori. So your brain can wind up settling on either of them. And that’s what you experience.

Again, the thing you “experience” is the generative model (e.g. the dancer spinning clockwise), not the raw sense data (patterns of light on your retina).

…And that’s the explanation of bistable perception! I don’t think there’s anything particularly deep or mysterious about it. That said, make sure you’re extremely comfortable with bistable perception, because we’re going to take that idea into much more mind-bending directions next.

1.2.5 Teaser: Unusual states of consciousness as a version of bistable perception

In this series, among other topics, I’ll be talking about “hearing voices”, “attaining enlightenment”, “entering trance”, dissociative identity disorder, and more. I propose that we can think of all these things as a bit like bistable perception—these are different generative models that are compatible with the same sensory data.

Granted, the analogy isn’t perfect. For one thing, the Spinning Dancer is just about perception, whereas a hypnotic trance (for example) impacts both perception and action. For another thing, the “sensory data” for the Spinning Dancer is just visual input—but what’s the analogous “sensory data” in the case of trance and so on? These are good questions, and I’ll get to them in future posts.

But meanwhile, I think the basic idea is absolutely right:

  • The Spinning Dancer is actually pixels on a screen, but we can experience it as either clockwise or counterclockwise motion—two different generative models.
  • Inner monologue is actually certain neurons and synapses doing a certain thing, but we can experience it as either coming from ourselves or coming from a disembodied voice—two different generative models.
  • Motor control is actually certain neurons and synapses doing some other thing, but we can experience it as either “caused by our own volition” or “caused by the hypnotist’s volition”—two different generative models.
  • Etc.

1.3 Casting judgment upon intuitive models

I know what you’re probably thinking: “OK, so people have intuitive models. How do I use that fact to judge people?” (Y’know, the important questions!) So in this section, I will go through three ways that people cast judgment upon themselves and others based on their intuitive models: “real” vs “fake”, “veridical” vs “non-veridical”, and “healthy” vs “pathological”. Only one of these three (veridical vs non-veridical) will be important for this series, but boy is that one important, so pay attention to that part.

1.3.1 “Is the intuitive model real, or is it fake?”

Fun fact: I consistently perceive the Spinning Dancer going clockwise.

…Is that fun fact “real”?

Yes!! I’m not trolling you—I’m telling you honestly what I experience. The dancer’s clockwise motion is my “real” experience, and it directly corresponds to a “real” pattern of neurons firing in my “real” brain, and this pattern is objectively measurable in principle (and maybe even in practice) if you put a sufficiently advanced brain scanner onto my head.

So, by the same token, suppose someone tells me that they hear disembodied voices, and suppose they’re being honest rather than trolling me. Is that “real”? Yes—in exactly the same sense.

1.3.2 “Is the intuitive model veridical, or is it non-veridical?”

When I say “X is a veridical model of Y”, I’m talking about map-territory correspondence:

  • X is part of the “map”—i.e., part of an intuitive model.
  • Y is part of the “territory”—i.e., it’s some kind of objective, observer-independent thing.
  • “X is a veridical model of Y” means roughly that there’s a straightforward, accurate correspondence between properties of X and corresponding properties of Y.

(Models can be more or less veridical, along a spectrum, rather than a binary.)

Simple example: I have an intuitive model of my sock. It’s pretty veridical! My sock actually exists in the world of atoms, and by and large there’s a straightforward and faithful correspondence between aspects of my intuitive model of the sock, and aspects of how those sock-related atoms are configured.

Conversely, Aristotle had an intuitive model of the sun, but in most respects it was not a veridical model of the sun. For example, his intuitive model said that the sun was smooth and unblemished and attached to a rotating transparent sphere made of aether.[3]

Here’s a weirder example, which will be relevant to this series. I have an intuitive concept of the mergesort algorithm. Is that intuitive model “veridical”? First we must ask: a veridical model of what? Well, it’s not a veridical model of any specific atoms in the real world. But it is a veridical model of a thing in the Platonic, ethereal realm of algorithms! That’s a bona fide “territory”, which is both possible and useful for us to “map”. So there’s a meaningful notion of veridicality there.

“The Platonic, ethereal realm of algorithms” (DALL-E 3)

When “veridical” needs nuance, I’ll just try to be specific in what I mean. For example, suppose intuitive concept X faithfully captures the behavior of algorithm Y, but X is intuitively conceptualized as a spirit floating in the room, rather than as an algorithm within the Platonic, ethereal realm of algorithms. Well then, I would just say something like: “X has good veridical correspondence to the behavior of algorithm Y, but the spirit- and location-related aspects of X do not veridically correspond to anything at all.”

(If that example seems fanciful, just wait for the upcoming posts!)

OK, hopefully you now know what I mean by the word “veridical”. I haven’t provided any rigorous and universally-applicable definition, because I don’t have one, sorry. But I think it will be clear enough.

Next, an important point going forward:

1.3.2.1 Non-veridical intuitive models are extremely common and unremarkable

…And I’m not just talking about your intuitive models of deliberate optical illusions, or of yet-to-be-discovered scientific phenomena. Here are some more everyday examples:

  • Do you see the Spinning Dancer going clockwise? Sorry, that’s not a veridical model of the real-world thing you’re looking at. Counter-clockwise? Sorry, that’s also not a veridical model of the real-world thing you’re looking at! A veridical model of the real-world thing you’re looking at would feel like a 2D pattern of changing pixels on a flat screen—after all, nothing in the real world of atoms is rotating in 3D!
    • If you know intellectually that you’re not looking at a 3D spinning thing, then good for you, see next subsection. But that doesn’t change the fact that you’re experiencing an intuitive model involving a 3D spinning thing.
  • Think of everything you’ve ever seen, or learned in school, that you found surprising, confusing, or unintuitive at the time you first saw it. Newton’s First Law! Magnets! Siphons! Gyroscopes! Quantum mechanics! Special Relativity! Antibonding orbitals! …If you found any of those unintuitive, then you must have had (and perhaps still have) non-veridical aspects of your intuitive models, that this new information was violating.
  • Does the moon seem to follow you when you walk at night? If so, then your intuitive model has the moon situated at the wrong distance from Earth.

1.3.2.2 …But of course it’s good if you’re intellectually aware of how veridical your various intuitive models are

It’s good to know intellectually that you have a non-veridical intuitive model, for the same humdrum reason that it’s good to know anything else that’s true. True beliefs are good. My intuitive models still say that the moon follows me at night, and that the spinning dancer spins, but I know intellectually that neither of those intuitive models are veridical. And that’s good.

By the same token, I endorse the conventional wisdom that if someone is “hearing voices”—in the sense of having an intuitive model that a disembodied voice is coming from 1 meter behind her head—then that’s pretty far on the non-veridical side of the spectrum. And if she denies that—if they say “the voice is there for sure—if you get the right scientific equipment and measure what’s happening 1 meter behind my head, then you’ll find it!”, then I say: “sorry but you’re wrong”.

Laugh at her all you want, but then go look in the mirror, because in my opinion everyone has not-terribly-veridical intuitive models of their metacognitive world—the world of consciousness, free will, desires, and so on—and practically all of us incorrectly believe those models to be more veridical than they really are, in various ways. Thus, in my spicy opinion, when Francis Crick (for example) says that phenomenal consciousness is in the claustrum, he’s being confused in a fundamentally similar way as that made-up person above who says that a disembodied voice is in the empty space behind her head. (I’m making a bold claim without defending it; more on this topic in later posts, but note that this kind of thing is not the main subject of this series.)

Source

1.3.3 “Is the intuitive model healthy, or is it pathological?”

As above, there’s nothing pathological about having non-veridical intuitive models.

In the case of the spinning dancer, it’s quite the opposite—if your intuitive model is a veridical model of the thing you’re looking at—i.e., if it doesn’t look like 3D spinning at all, but rather looks like patterns of pixels on a flat screen—then that’s the situation where you might consider checking with a neurologist!!

So, how about hearing voices? According to activists, people who hear voices can find it a healthy and beneficial part of their lives, as long as other people don’t judge and harass them for it—see Eleanor Longden’s TED talk. OK, cool, sounds great, as far as I know. As long as they’re not a threat to themselves or others, let them experience the world how they like!

I’ll go further: if those people think that the voices are veridical—well, they’re wrong, but oh well, whatever. In and of itself, a wrong belief is nothing to get worked up about, if it’s not ruining their life. Pick any healthy upstanding citizen, and it’s a safe bet that they have at least one strong belief, very important to how they live their life, that ain’t true.

(And as above, if you want to judge or drug people for thinking that their intuitive self-models are more veridical than the models really are, then I claim you’ll be judging or drugging 99.9…% of the world population, including many philosophers-of-mind, neuroscientists, etc.)

That’s how I think about all the other unusual states of consciousness too—trance, enlightenment, tulpas, you name it. Is it working out? Then great! Is it creating problems? Then try to change something! Don’t ask me how—I have no opinion.

1.4 Why does the predictive learning algorithm build generative models / concepts related to what’s happening in your own mind?

For example, I have a learned concept in my world-model of “screws”.

Sometimes I’ll think about that concept in the context of the external world: ”there are screws in the top drawer”.

But other times I’ll think about that concept in the context of my own inner world: ”I’m thinking about screws right now”, “I’m worried about the screws”, “I can never remember where I left the screws”, etc.

If the cortex “learns from scratch”, as I believe, then we need to explain how these models of my inner world get built by a predictive learning algorithm.

To start with: In general, if world-model concept X is active, it tends to invoke (or incentivize the creation of) an “explanation” of X—i.e., an upstream model that explains the fact that X is active—and which can thus help predict when X will be active again in the future.

This is just an aspect of predictive (a.k.a. self-supervised) learning from sensory inputs—the cortex learning algorithm sculpts the generative world-model to predict when the concept X is about to be active, just as it sculpts the world-model to predict when raw sensory inputs are about to be active.

For example, if a baby frequently sees cars, she would first learn a “car” concept that helps predict what car-related visual inputs are doing after they first appear. But eventually the predictive learning algorithm would need to develop a method for anticipating when the “car” concept itself was about to be active. For example, the baby would learn to expect cars when she looks at a street.

For a purely passive observer, that’s the whole story, and there would be no algorithmic force whatsoever for developing self-model / inner-world concepts. If the “car” concept is active right now, there must be cars in the sensory input stream, or at least something related to cars. Thus, when the algorithm finds models that help predict that the “car” concept is about to be active, those models will always be exogenous—they’ll track some aspect of how the world works outside my own head. “Cars are often found on highways.” “Cars are rarely found in somebody’s mouth.” Etc.

However, for an active agent, concepts in my world-model are often active for endogenous reasons. Maybe the “car” concept is active in my mind because it spontaneously occurred to me that it would be a good idea to go for a drive right about now. Or maybe it’s on my mind because I’m anxious about cars. Etc. In those cases, the same predictive learning algorithm as above—the one sculpting generative models to better anticipate when my “car” concept will be active—will have to construct generative models of what’s going on in my own head. That’s the only possible way to make successful predictions in those cases.

…So that’s all that’s needed. If any system has both a capacity for endogenous action (motor control, attention control, etc.), and a generic predictive learning algorithm, that algorithm will be automatically incentivized to develop generative models about itself (both its physical self and its algorithmic self), in addition to (and connected to) models about the outside world.

1.4.1 Further notes on the path from predictive learning algorithms to intuitive self-models

  • I said that, for any active agent with a predictive learning algorithm, that algorithm is incentivized to develop generative models of the self. But that doesn’t mean that the algorithm will necessarily succeed at developing such generative models. As Machine Learning researchers know, it’s perfectly possible to train a model to learn something, only to find that it can’t actually learn it. In fact, it’s impossible for the intuitive models to perfectly predict the outputs of the brain algorithm (of which they themselves are a part).[4] As a result, the generative models wind up modeling some aspects of brain algorithms, but also encapsulating other aspects into entities whose internals cannot themselves be deterministically modeled or predicted. As a teaser for future posts, these entities wind up closely related to key concepts like “my true self”, “free will”, and so on—but hold that thought until Post 3.
The predictive learning algorithm in the brain is given the objective of building a generative model of (aspects of) the brain algorithm itself. That’s an impossible task. So I think the generative model winds up modeling some aspects of the algorithm, while encapsulating other parts into entities which can be modeled probabilistically but not mechanistically—much like how you would intuitively model a living creature. More on this in Post 3.
  • The trained model that results from predictive learning is generally underdetermined—it can wind up depending on fine details, like the order that data is presented, learning algorithm hyperparameters that vary amongst different people, and so on. Indeed, as I’ll discuss in this series, different people can wind up with rather different intuitive self-models: there are people who hear voices, people who have achieved Buddhist enlightenment, and so on.
  • Relatedly, every vertebrate, and presumably many invertebrates too, are also active agents with predictive learning algorithms in their brain, and hence their predictive learning algorithms are also incentivized to build intuitive self-models. But I’m sure that the intuitive self-model in a zebrafish pallium (if any) is wildly different from the intuitive self-model in an adult human cortex.

1.5 Appendix: Some terminology I’ll be using in this series

I’m sure everything I write will be crystal clear in context (haha), but just in case, here are some of the terms and assumptions that I’ll be using throughout the series, related to how I think about probabilistic inference and generative modeling in the brain.

Learning algorithms and trained models

In this series, “learning algorithms” always means within-lifetime learning algorithms that are designed by evolution and built directly into the brain—I’m not talking about genetic evolution, or cultural evolution, or learned metacognitive strategies like spaced repetition.

In Machine Learning, people talk about the distinction between (A) learning algorithms versus (B) the trained models that those learning algorithms gradually build. I think this is a very useful distinction for the brain too—see my discussion at “Learning from scratch” in the brain. (The brain also does things that are related to neither learning algorithms nor trained models, but those things won’t be too relevant for this series.) If you’ve been reading my other posts, you’ll notice that I usually spend a lot of time talking about (certain aspects of) brain learning algorithms, and very little time talking about trained models. But this series is an exception: my main focus here is within the trained model level—i.e., the specific content and structure of the generative models in the cortex of a human adult.

Predictive learning (also called “self-supervised learning”) is any learning algorithm that tries to build a generative model that can predict what’s about to happen (e.g. imminent sensory inputs). When the generative model gets a prediction wrong (i.e., is surprised), the learning algorithm updates the generative model, or builds a new one, such that it will be less surprised in similar situations in the future. As I’ve discussed here, I think predictive learning is a very important learning algorithm in the brain (more specifically, in the cortex and thalamus). But it’s not the only learning algorithm—I think reinforcement learning is a separate thing, still centrally involving the cortex and thalamus, but this time also involving the basal ganglia.

(Learning from culture is obviously important, but I claim it’s not a separate learning algorithm, but rather an emergent consequence of predictive learning, reinforcement learning, and innate social drives, all working together.)

Concepts, models, thoughts, subagents

An “intuitive model”—which I’m using synonymously with “generative model”—constitutes a belief / understanding about what’s going on, and it issues corresponding predictions / expectations about what’s going to happen next. Intuitive models can say what you’re planning, seeing, remembering, understanding, attempting, doing, etc. If there’s exactly one active generative model right now in your cortex—as opposed to your being in a transient state of confusion—then that model would constitute the thought that you’re thinking.

The “generative model space” would be the set of all generative models that you’ve learned, along with how a priori likely they are. I have sometimes called that by the term “world-model”, but I’ll avoid that in this series, since it’s actually kinda broader than a world-model—it not only includes everything you know about the world, but also everything you know about yourself, and your habits, skills, strategies, implicit expectations, and so on.

“concept”, a.k.a. “generative model piece”, would be something which isn’t (usually) a generative model in and of itself, but more often forms a generative model by combining with other concepts in appropriate relations. For example, “hanging the dress in the closet” might be a generative model, but it involves concepts like the dress, its hanger, the closet, etc.

“Subagent” is not a term I’ll be using in this series at all, but I’ll mention it here in case you’re trying to compare and contrast my account with others’. Generative models sometimes involve actions (e.g. “I’m gonna hang the dress right now”), and sometimes they don’t (e.g. “the ball is bouncing on the floor”). The former, but not the latter, are directly[5] sculpted and pruned by not only predictive learning but also reinforcement learning (RL), which tries to find action-involving generative models that not only make correct predictions but also have maximally positive valence (as defined in my Valence series). Anyway, I think people usually use the word “subagent” to refer to a generative model that involves actions, or perhaps sometimes to a group of thematically-related generative models that involve thematically-related actions. For example, the generative model “I’m going to open the window now” could be reconceptualized as a “subagent” that sometimes pokes its head into your conscious awareness and proposes to open the window. And if the valence of that proposal is positive, then that subagent would “take over” (become and remain the active generative model), and then you would get up and actually open the window.

1.6 Appendix: How does this series fit into Philosophy Of Mind?

1.6.1 Introspective self-reports as a “straightforward” scientific question

Hopefully we can all agree that we live in a universe that follows orderly laws of physics, always and everywhere, even if we don’t know exactly what those laws are yet.[6] And hopefully we can all agree that those laws of physics apply to the biological bodies and brains that live inside that universe, just like everything else. After all, whenever scientists measure biological organisms, they find that their behavior can be explained by normal chemistry and physics, and ditto when they measure neurons and synapses.

So that gets us all the way to self-reports about consciousness, free will, and everything else. If you ask someone about their inner mental world, and they move their jaws and tongues to answer, then those are motor actions, and (via the laws of physics) we can trace those motor actions back to signals going down motoneuron pools, and those signals in turn came from motor cortex neurons spiking, and so on back through the chain. There’s clearly a systematic pattern in what happens—people systematically describe consciousness and the rest of their mental world in some ways but not others. So there has to be an explanation of those patterns, within the physical universe—principles of physics, chemistry, biology, neuroscience, algorithms, and so on.

“The meta-problem of consciousness” is a standard term for part of this problem, namely: What is the chain of causation in the physical universe that reliably leads to people declaring that there’s a “Hard Problem of Consciousness”? But I’m broadening that to include everything else that people say about consciousness—e.g. people describing their consciousness in detail, including talk about enlightenment, hypnotic trance, and so on—and then we can broaden it further to include everything people say about free will, sense-of-self, etc.

1.6.2 Are explanations-of-self-reports a first step towards understanding the “true nature” of consciousness, free will, etc.?

The broader research program would be:

Research program: 

  • STEP 1: Explain the chain-of-causation in the physical universe that leads to self-reports about consciousness, free will, etc.—and not just people’s declarations that those things exist at all, but also all the specific properties that people ascribe to those things.
  • STEP 2: Draw lessons about what (if anything) consciousness actually is, and what (if anything) free will actually is, etc., along with lessons about moral patienthood, and the meaning of life, or whatever else.

This series is focused on STEP 1, not STEP 2. Why? For one thing, STEP 1 is a pure question of physics and chemistry and biology and algorithms and neuroscience, which is in my comfort zone; whereas STEP 2 is a question of philosophy, which is definitely not. For another thing, I find that arguments over the true nature of consciousness are endless and polarizing and exhausting, going around in circles, whereas I hope we can all join together towards the hopefully-uncontroversial STEP 1 scientific endeavor.

But some readers might be wondering:

Is STEP 1 really relevant to STEP 2? 

The case for yes: If you ask me the color of my wristwatch, and if I answer you honestly from my experience (rather than trolling you, or parroting something I heard, etc.), then somewhere in the chain of causation that ultimately leads to me saying “it’s black”, you’ll find an actual wristwatch, and photons bouncing off of it, entering my eyes, etc.

What’s true for my wristwatch should be equally true of “consciousness”, “free will”, “qualia”, and so on: if those things exist at all, then they’d better be somewhere in the chain of causation that leads to us talking about them. When I ask you a question about those things, your brain needs to somehow sense their properties, and then actuate your tongue and lips accordingly. Otherwise you would have no grounds for saying anything about them at all! Everything you say about them would be just a completely uninformed shot in the dark! Indeed, as far as you know, everything you think you know about qualia is wrong, and in fact qualia are not unified, and not subjective, and not ineffable, but they are about two meters in diameter, with a consistency similar to chewing gum, and they ooze out of your bellybutton. Oh, that’s wrong? Are you sure it’s wrong? Now ask yourself what process you went through to so confidently confirm that in your mind. That process had to have somehow involved somehow “observing” your “qualia” to discern their properties, right?

Alternatively, see here for an argument for that same conclusion but framed in terms of Bayesian epistemology.

The case for no: You can find one in Chalmers 2003. He acknowledges the argument above (“It is certainly at least strange to suggest that consciousness plays no causal role in my utterances of ‘I am conscious’. Some have suggested more strongly that this rules out any knowledge of consciousness… The oddness of epiphenomenalism is exacerbated by the fact that the relationship between consciousness and reports about consciousness seems to be something of a lucky coincidence, on the epiphenomenalist view …”) But then Chalmers goes through a bunch of counterarguments, to which he is sympathetic. I won’t try to reproduce those here.

In any case, getting back to the topic at hand:

  • I obviously have opinions about who’s right in that back-and-forth above. And I also have opinions about The Hard Problem of Consciousness (probably not what you’d expect!). But I’ll leave them out, because that would be getting way off-topic.
  • Even if you don’t think STEP 1 provides any definitive evidence about STEP 2, I imagine you’d be at least curious about STEP 1, if only for its own sake. And if so, you should have no foundational objection to what I’m trying to do in this series.
  • Indeed, even if you think that the STEP 1 research program is doomed, because the laws of physics as we know them will be falsified (i.e. give incorrect predictions) in future experiments involving conscious minds—then you still ought to be interested in the STEP 1 research program, if only to see how far it gets before it fails!!

Yet again, I maintain that anyone with any opinion about the nature of consciousness, free will, etc., should be jointly interested in the STEP 1 research program.

…But as it happens, the people who have actually tried to work on STEP 1 in detail have disproportionately tended to subscribe to the philosophy-of-mind theory known as strong illusionism—see Frankish 2016. The five examples that I’m familiar with are: philosophers Keith Frankish, Daniel Dennett, and Thomas Metzinger, and neuroscientists Nicholas Humphrey and Michael Graziano. I’m sure there are others too.[7]

I think that some of their STEP-1-related discussions (especially Graziano’s) have kernels of truth, and I will cite them where applicable, but I find others to be wrong, or unhelpfully vague. Anyway, I’m mostly trying to figure things out for myself—to piece together a story that dovetails with my tons of opinions about how the human brain works.

1.7 Conclusion

Now that you better understand how I think about intuitive self-models in general, the next seven posts will dive into the specifics!

Thanks Thane Ruthenis, Charlie Steiner, Kaj Sotala, lsusr, Seth Herd, Johannes Mayer, Jonas Hallgren, and Justis Mills for critical comments on earlier drafts.

  1. ^

    In case you’re wondering, this series will centrally involve probabilistic inference, but will not involve “active inference”. I think most “active inference” discourse is baloney (see Why I’m not into the Free Energy Principle), and indeed I’m not sure how active inference ever became so popular given the obvious fact that things can be plausible but not desirable, and that things can be desirable but not plausible. I think “plausibility” involves probabilistic inference, while “desirability” involves valence—see my Valence series.

  2. ^

    I think it would be a bit more conventional to say that the brain has a (singular) generative model with lots of adjustable parameters / settings, but I think the discussion will sound more intuitive and flow better if I say that the brain has a whole space of zillions of generative models (plural), each with greater or lesser degrees of a priori plausibility. This isn’t a substantive difference, just a choice of terminology.

  3. ^

    We can’t read Aristotle’s mind, so we don’t actually know for sure what Aristotle’s intuitive model of the sun was; it’s technically possible that Aristotle was saying things about the sun that he found unintuitive but nevertheless intellectually believed to be true (see §1.3.2.2). But I think that’s unlikely. I’d bet he was describing his intuitive model.

  4. ^

    The idea that a simpler generative model can’t predict the behavior of a big complicated algorithm is hopefully common sense, but see Rice’s theorem for a related formalization.

  5. ^

    Reinforcement Learning (RL) is obviously indirectly relevant to the formation of generative models that don’t involve actions. For example, if I really like clouds, then I might spend all day watching clouds, and spend all night imagining clouds, and I’ll thus wind up with unusually detailed and accurate generative models of clouds. RL is obviously relevant in this story: RL is how my love of clouds influences my actions, including both attention control (thinking about clouds) and motor control (looking at clouds). And those actions, in turn, influence the choice of data that winds up serving as a target for predictive learning. But it’s still true that my generative models of clouds are updated only by predictive learning, not RL.

  6. ^

    “The Standard Model of Particle Physics including weak-field quantum general relativity (GR)” (I wish it was better-known and had a catchier name) appears sufficient to explain everything that happens in the solar system (ref). Nobody has ever found any experiment violating it, despite extraordinarily precise and varied tests. This theory can’t explain everything that happens in the universe—in particular, it can’t make any predictions about either (A) microscopic exploding black holes or (B) the Big Bang. Also, (C) the Standard Model happens to include 18 elementary particles (depending on how you count), because those are the ones we’ve discovered; but the theoretical framework is fully compatible with other particles existing too, and indeed there are strong theoretical and astronomical reasons to think they do exist. It’s just that those other particles are irrelevant for anything happening on Earth—so irrelevant that we’ve spent decades and billions of dollars searching for any Earthly experiment whatsoever where they play a measurable role, without success. Anyway, I think there are strong reasons to believe that our universe follows some set of orderly laws—some well-defined mathematical framework that elegantly unifies the Standard Model with all of GR, not just weak-field GR—even if physicists don’t know what those laws are yet. (I think there are promising leads, but that’s getting off-topic.) …And we should strongly expect that, when we eventually discover those laws, we’ll find that they shed no light whatsoever into how consciousness works—just as we learned nothing whatsoever about consciousness from previous advances in fundamental physics like GR or quantum field theory.

  7. ^

    Maybe Tor Norretranders’s The User Illusion (1999) belongs in this category, but I haven’t read it.

New Comment