All of EuanMcLean's Comments + Replies

Really cool post!

I can't help but compare it to john vervaeke's 4 ways of knowing: propositional, procedural, perspectival, and participatory (nice summary).

It's interesting how the two of you have made a categorization that can literally be described with the same words (4 ways of knowing), but they don't map onto each other much at all (maybe practical mastery ~ procedural, but a bit of a stretch). I guess yours is "4 ways of knowing a thing" and his is "4 ways of knowing things", but still!

I assume that phenomenal consciousness is a sub-component of the mind.

I'm not sure what is meant by this; would you mind explaining?

I mean something along the lines of "if you specify all aspects of the mind (e.g. using a program), you have also specified all aspects of the conscious experience"

Also, the in-post link to the appendix is broken; it's currently linking to a private draft.

Eek, thanks for the heads up, fixed!

MCMC is a simpler example to ensure that we’re on the same page on the general topic of how randomness can be involved in algorithms.

Thanks for clarifying :)

Are we 100% on the same page about the role of randomness in MCMC? Is everything I said about MCMC super duper obvious from your perspective?

Yes.

If I run MCMC with a PRNG given random seed 1, it outputs 7.98 ± 0.03. If I use a random seed of 2, then the MCMC spits out a final answer of 8.01 ± 0.03. My question is: does the random seed entering MCMC “have a causal effect on the execution of the

... (read more)

I'm not saying anything about MCMC. I'm saying random noise is not what I care about, the MCMC example is not capturing what I'm trying to get at when I talk about causal closure.

I don't disagree with anything you've said in this comment, and I'm quite confused about how we're able to talk past each other to this degree.

8Steven Byrnes
Yeah duh I know you’re not talking about MCMC. :) But MCMC is a simpler example to ensure that we’re on the same page on the general topic of how randomness can be involved in algorithms. Are we 100% on the same page about the role of randomness in MCMC? Is everything I said about MCMC super duper obvious from your perspective? If not, then I think we’re not yet ready to move on to the far-less-conceptually-straightforward topic of brains and consciousness. I’m trying to get at what you mean by: I don’t understand what you mean here. For example: * If I run MCMC with a PRNG given random seed 1, it outputs 7.98 ± 0.03. If I use a random seed of 2, then the MCMC spits out a final answer of 8.01 ± 0.03. My question is: does the random seed entering MCMC “have a causal effect on the execution of the algorithm”, in whatever sense you mean by the phrase “have a causal effect on the execution of the algorithm”? * My MCMC code uses a PRNG that returns random floats between 0 and 1. If I replace that PRNG with return 0.5, i.e. the average of the 0-to-1 interval, then the MCMC now returns a wildly-wrong answer of 942. Is that replacement the kind of thing you have in mind when you say “just take the average of those fluctuations”? If so, how do you reconcile the fact that “just take the average of those fluctuations” gives the wrong answer, with your description of that scenario as “what it’s meant to do”? Or if not, then what would “just take the average of those fluctuations” mean in this MCMC context?

The most obvious examples are sensory inputs—vision, sounds, etc. I’m not sure why you don’t mention those.

Obviously algorithms are allowed to have inputs, and I agree that the fact that the brain takes in sensory input (and all other kinds of inputs) is not evidence against practical CF. The way I'm defining causal closure is that the algorithm is allowed to take in some narrow band of inputs (narrow relative to, say, the inputs being the dynamics of all the atoms in the atmosphere around the neurons, or whatever). My bad for not making this more expli... (read more)

4Steven Byrnes
I’m confused by your comment. Let’s keep talking about MCMC. * The following is true: The random inputs to MCMC have “a causal effect on the execution of the algorithm such that the algorithm doesn't do what it's meant to do if you just take the average of those fluctuations”. * For example, let’s say the MCMC accepts a million inputs in the range (0,1), typically generated by a PRNG in practice. If you replace the PRNG by the function return 0.5 (“just take the average of those fluctuations”), then the MCMC will definitely fail to give the right answer. * The following is false: “the signals entering…are systematic rather than random”. The random inputs to MCMC are definitely expected and required to be random, not systematic. If the PRNG has systematic patterns, it screws up the algorithm—I believe this happens from time to time, and people doing Monte Carlo simulations need to be quite paranoid about using an appropriate PRNG. Even very subtle long-range patterns in the PRNG output can screw up the calculation. The MCMC will do a highly nontrivial (high-computational-complexity) calculation and give a highly non-arbitrary answer. The answer does depend to some extent on the stream of random inputs. For example, suppose I do MCMC, and (unbeknownst to me) the exact answer is 8.00. If I use a random seed of 1 in my PRNG, then the MCMC might spit out a final answer of 7.98 ± 0.03. If I use a random seed of 2, then the MCMC might spit out a final answer of 8.01 ± 0.03. Etc. So the algorithm run is dependent on the random bits, but the output is not totally arbitrary. All this is uncontroversial background, I hope. You understand all this, right? As it happens, almost all modern computer chips are designed to be deterministic, by putting every signal extremely far above the noise floor. This has a giant cost in terms of power efficiency, but it has a benefit of making the design far simpler and more flexible for the human programmer. You can write code without

If you believe there exists "a map between the third-person properties of a physical system and whether or not it has phenomenal consciousness" you believe you can define consciousness with a computation.

I'm not arguing against the claim that you could "define consciousness with a computation". I am arguing against the claim that "consciousness is computation". These are distinct claims.

So, most people who take the materialist perspective believe the material world comes from a sort of "computational universe", e.g. Tegmark IV.

Massive claim, nothing... (read more)

2James Camacho
I find it useful to take an axiom of extensionality—if I cannot distinguish between two things in any way, I may as well consider them the same thing for all that it could affect me. Given maths/computation/logic is the process of asserting things are the same or different, it seems to me to be tautologically true that maths and computaiton are the only symbols upon which useful discussion can be built. Maybe you want to include some undefinable aspect to consciousness. But anytime it functions differently, you can use that to modify your definition. I don't think the adherents for computational functionalism, or even a computational universe, need to claim it encapsulates everything there could possibly be in the territory. Only that it encapsulates anything you can perceive in the territory. I believe this is your definition of real consciousness? This tells me properties about consciousness, but doesn't really help me define consciousness. It's intrinsic and objective, but what is it? For example, if I told you that the Serpinski triangle is created by combining three copies of itself, I still don't know what it actually looks like. If I want to work with it, I need to know how the base case is defined. Once you have a definition, you've invented computational functionalism (for the Serpinski triangle, for consciousness, for the universe at large). Yes, exactly! To be precise, I don't consider an argument useful unless it is defined through a constructive logic (e.g. mathematics through ZF set theory). Note: this assumes computational functionalism. I haven't seen it written down explicitly anywhere, but I've seen echoes of it here and there. Essentially, in RL, agents are defined via their policies. If you want to modify the agent to be good at a particular task, while still being pretty much the "same agent", you add a KL-divergence anchor term: Loss(π)=Subgame Loss(π)+λKL(π,πoriginal). This is known as piKL and was used for Diplomacy, where it's importa

I don't really understand the point of this thought experiment, because if it wasn't phrased in such a mysterious manner, it wouldn't seem relevant to computational functionalism.

I'm sorry my summary of the thought experiment wasn't precise enough for you. You're welcome to read Chalmers' original paper for more details, which I link to at the top of that section.

I also don't understand a single one of your arguments against computational functionalism

I gave very brief recaps of my arguments from the other posts in the sequence here so I can connect... (read more)

-4James Camacho
Why is that the one effect? Jordan Peterson says that the one answer he routinely gives to Christians and atheists that piss them off is, "what do you mean by that?" In an interview with Alex O'Conner he says, Sure, this pisses off a lot of people, but it also gets some people thinking about what they actually mean. So, there's your answer: you're supposed to go back and figure out what you mean. A side benefit is if it pisses you off, maybe I won't see your writing anymore. I'm pretty annoyed at how the quality of posts has gone down on this website in the past few years.

Could you recommend any good (up-to-date) reading defending the neuron doctrine?

2Seth Herd
No. I'm not sure anyone has bothered to write one. There were only occasional halfhearted and poorly supported attacks on the neuron doctrine. The neuron doctrine is just not really debated because it's almost universally accepted. No reputable neuroscientist argued against it to any strong degree, just for additional supportive methods of information transmission. The closest reasonable question about it was active spikes at dendritic junctions. These are probably important, but they are akin to adding some extra layers of simple neurons to the network. They're using the same basic ionic gates to send signals. There's no reason to think those need to be modeled in anything like molecular detail; their function is well understood at a macro level.

How would the alien know when they've found the correct encoding scheme?

I'm not sure I understand this. You're saying the alien could look at the initial conditions, since they're much simpler than the quantum fields as the simulation runs? In that case, how could it track down those initial conditions and interpret them?

Ah I see, thanks for clarifying.

Perhaps I should have also given the alien access to infinite compute. I think the alien still wouldn't be able to determine the correct simulation.

And also infinite X if you hit me with another bottleneck of the alien not having enough X in practice.

The thought experiment is intended to be about in-principle rather than practical.

2Noosphere89
In general, I expect these sorts of constraint removals to make problems trivial, with exceptions being problems where you have to arbitrarily maintain a finite computational power, and a big problem of philosophy is not realizing how much their intuitions rests on constraints of our own world that don't have to hold when infinity is involved. More generally, a lot of our intuitions involve exploiting constraints on the world at large, which means that when you remove those constraints, our intuitions become false. I think Searle's Chinese Room argument is flawed for reasons similar to this, and more generally the use of idealizations/thought experiments make philosophers forget how wrong their intuition is when they consider the question (at least for non-moral and possibly non-identity cases, though I am much more fragile on the confidence of the non-identity case specifically.
2Steven Byrnes
I don’t think any of the challenges you mentioned would be a blocker to aliens that have infinite compute and infinite time. “Is the data big-endian or little-endian?” Well, try it both ways and see which one is a better fit to observations. If neither seems to fit, then do a combinatorial listing of every one of the astronomical number of possible encoding schemes, and check them all! Spend a trillion years studying the plausibility of each possible encoding before moving onto the next one, just to make sure you don’t miss any subtelty. Why not? You can do all sorts of crazy things with infinite compute and infinite time.

Hmm I guess I gave your original comment too shallow a reading, apologies for that.

So to be clear, are you saying that, if a half-awake version of you looks at a button saying "I am conscious", thinks to themselves "am I conscious? Yes I am!", and presses the button, whether or not that half-awake version was actually correct with that introspection is up to interpretation? In other words, you don't trust the report of your half-awake self?

My instinct is to say something like: if your half-awake self is actually capable of introspecting on their experience... (read more)

it does not boil down to Chalmer's argument.

As far as I can tell, Scott's argument does not argue against the possibility that a waterfall could execute a single forward pass of a chess playing algorithm, if you defined a gerrymandered enough map between the waterfall and logical states.

When he defines the waterfall as a potential oracle, implicit in that is that the oracle will respond correctly to different inputs - counterfactuals.

Viewing the waterfall's potential oracleness as an intrinsic property of that system is to view counterfactual waterfalls... (read more)

2Davidmanheim
I've written my point more clearly here: https://www.lesswrong.com/posts/zxLbepy29tPg8qMnw/refuting-searle-s-wall-putnam-s-rock-and-johnson-s-popcorn

Whether aliens can figure out that fact is irrelevant.

To be clear, would you say that you are disagreeing with "Premise 2" above here?

Premise 2: Phenomenal consciousness is a natural kind: There is an objective fact-of-the-matter whether a conscious experience is occurring, and what that experience is. It is not observer-dependent. It is not down to interpretation. It is an intrinsic property of a system. It is the territory rather than a map.

2Steven Byrnes
I don’t think Premise 2 is related to my comment. I think it’s possible to agree with premise 2 (“there is an objective fact-of-the-matter whether a conscious experience is occurring”), but also to say that there are cases where it is impossible-in-practice for aliens to figure out that fact-of-the-matter. By analogy, I can write down a trillion-digit number N, and there will be an objective fact-of-the-matter about what is the prime factorization of N, but it might take more compute than fits in the observable universe to find out that fact-of-the-matter.

Is this the passage you're referring to that means I'm "fundamentally misunderstanding computation"?

suppose we actually wanted to use a waterfall to help us calculate chess moves. [...] I conjecture that, given any chess-playing algorithm A that accesses a “waterfall oracle” W, there is an equally-good chess-playing algorithm A0, with similar time and space requirements, that does not access W. If this conjecture holds, then it gives us a perfectly observer-independent way to formalize our intuition that the “semantics” of waterfalls have nothing to do w

... (read more)
2Davidmanheim
Yes, and no, it does not boil down to Chalmer's argument. (as Aaronson makes clear in the paragraph before the one you quote, where he cites the Chalmers argument!) The argument from complexity is about the nature and complexity of systems capable of playing chess - which is why I think you need to carefully read the entire piece and think about what it says. But as a small rejoinder, if we're talking about playing a single game, the entire argument is ridiculous; I can write the entire "algorithm" a kilobyte of specific instructions. So it's not that an algorithm must be capable of playing multiple counterfactual games to qualify, or that counterfactuals are required for moral weight - it's that the argument hinges on a misunderstanding of how complex different classes of system need to be to do the things they do. PS. Apologies that the original response comes off as combative - I really think this discussion is important, and wanted to engage to correct an important point, but have very little time to do so at the moment!

I'm not talking about access consciousness here. I'm not talking about the ability to report. I'm talking about phenomenal consciousness.

Maybe I'm wrong, but I predict you're going to say "there's no difference", or "there's nothing to consciousness besides reporting" or something, which is a position I have sympathy for and is closely related to the I talk about at the end of the post. But reporting is not what I'm talking about here.

2Charlie Steiner
What I'm going to say is that I really do mean phenomenal consciousness. The person who turns off the alarm not realizing it's an alarm, poking at the loud thing without understanding it, is already so different from my waking self. And those are just the ones that I remember - the shape of the middle of the distribution implies the existence of an unremembered tail of the distribution. If I'm sleeping dreamlessly, and take a reflexive action such as getting goosebumps, am I having a kinesthetic experience? If you say yes here, then perhaps there is no mystery and you just use 'experience' idiosyncratically.

Perhaps the same calculation could simulate different real world phenomena, but it doesn't follow that the subjective experiences are different in each case.

I see what you mean I think - I suppose if you're into multiple realizability perhaps the set of all physical processes that the alien settles on all implement the same experience. But this just depends on how broad this set is. If it contains two brains, one thinking about the roman empire and one eating a sandwich, we're stuck.

This also does not follow. Both experiences could happen in the same

... (read more)
2simon
I suspect that if you do actually follow Aaronson (as linked by Davidmanheim) to extract a unique efficient calculation that interacts with the external world in a sensible way, that unique efficient externally-interacting calculation will end up corresponding to a consistent set of experiences, even if it could still correspond to simulations of different real-world phenomena. But I also don't think that consistent set of experiences necessarily has to be a single experience! It could be multiple experiences unaware of each other, for example.

Yea, you might be hitting on at least a big generator of our disagreement. Well spotted

4Rafael Harth
One thing I worry about is that the same disagreement happens with a lot of other users who, unlike Steven, just downvote the post rather than writing a comment. In general, when I've read through the entire LW catalogue of posts with the consciousness tag, I've noticed that almost all well received posts with the consciousness tag take what I call the camp #1 perspective (i.e., discuss consciousness from an illusionist lens, even if it's not always stated explicitly). Iirc the only major exceptions are the posts from Eliezer, which, well, are from Eliezer. So it could be that post who discuss consciousness from a realist PoV consistently receive certain amount of downvotes from camp #1 people to whom the post just seems gibberish/a waste of time. I don't have any data to prove that this is the mechanism, it's just a guess, but the pattern is pretty consistent. I also think you generally wouldn't predict this if you just read the comment sections. (And idk if clarifying the perspective would help since no one does it.)

fixed, thanks. Careless exponent juggling

Thanks for the comment Steven.

Your alternative wording of practical CF is indeed basically what I'm arguing against (although, we could interpret different degrees of the simulation having the "exact" same experience, and I think the arguments here don't only argue against the strongest versions but also weaker versions, depending on how strong those arguments are).

I'll explain a bit more why I think practical CF is relevant to CF more generally.

Firstly, functionalist commonly say things like

Computational functionalism: the mind is the software of the brai

... (read more)

I guess I shouldn’t put words in other people’s mouths, but I think the fact that years-long trains-of-thought cannot be perfectly predicted in practice because of noise is obvious and uninteresting to everyone, I bet including to the computational functionalists you quoted, even if their wording on that was not crystal clear.

There are things that the brain does systematically and robustly by design, things which would be astronomically unlikely to happen by chance. E.g. the fact that I move my lips to emit grammatical English-language sentences rather tha... (read more)

If I understand your point correctly, that's what I try to establish here

the speed of propagation of ATP molecules (for example) is sensitive to a web of more physical factors like electromagnetic fields, ion channels, thermal fluctuations, etc. If we ignore all these contingencies, we lose causal closure again. If we include them, our mental software becomes even more complicated.

i.e., the cost becomes high because you need to keep including more and more elements of the dynamics.

4Seth Herd
Only if all of those complex interactions are necessary to capture consciousness at all, not just to precisely reproduce the dynamics of one particular consciousness. If that were the case, the brain would be a highly unreliable mechanism. We'd lose consciousness when exposed to an external magnetic field, for instance, or when our electrolyte balance was off. Most of the brain's complexity contributes to maintaining information flow between neurons. The remainder is relatively cheap to simulate. See my other comment.

The statement I'm arguing against is:

Practical CF: A simulation of a human brain on a classical computer, capturing the dynamics of the brain on some coarse-grained level of abstraction, that can run on a computer small and light enough to fit on the surface of Earth, with the simulation running at the same speed as base reality, would cause the conscious experience of that brain.

i.e., the same conscious experience as that brain. I titled this "is the mind a program" rather than "can the mind be approximated by a program".

Whether or not a simulation ca... (read more)

Yes, perfect causal closure is technically impossible, so it comes in degrees. My argument is that the degree of causal closure of possible abstractions in the brain is less than one might naively expect.

Are there any measures of approximate simulation that you think are useful here?

I am yet to read this but I expect it will be very relevant! https://arxiv.org/abs/2402.09090

Especially if it's something as non-committal as "this mechanism could maybe matter". Does that really invalidate the neuron doctrine?

I agree each of the "mechanisms that maybe matter" are tenuous by themselves, the argument I'm trying to make here is hits-based. There are so many mechanisms that maybe matter, the chances of one of them mattering in a relevant way is quite high.

3Seth Herd
Sure, if you look at them in the abstract. Having considered lots of proposed mechanisms-that-matter over the course of a 23-year career in computational neuroscience, I still largely believe the neuron doctrine. The amount of information in neurons and the rate it changes seems quite adequate to explain the information density and rate-of-change of conscious experience. A rough simulation of brain function wouldn't precisely reproduce our conscious experience, but I see no reason to believe it wouldn't produce something very much like an average human experience. Assuming that every last molecule is functionally important (rather than very aggregated effects) seems so unlikely as to be irrelevant. Neuronal information transfer supports consciousness. Other mechanisms facilitate and regulate neuronal information transfer. Some transfer relatively small amounts of information. All of this can be simulated in the types of computers we'd have by around 2050, even if humans are still the only ones making better computuers - which now seems unlikely.
3Zac Hatfield-Dodds
I think your argument also has to establish that the cost of simulating any that happen to matter is also quite high. My intuition is that capturing enough secondary mechanisms, in sufficient-but-abstracted detail that the simulated brain is behaviorally normal (e.g. a sim of me not-more-different than a very sleep-deprived me), is likely to be both feasible by your definition and sufficient for consciousness.

Thanks for the feedback Garrett.

This was intended to be more of a technical report than a blog post, meaning I wanted to keep the discussion reasonably rigorous/thorough. Which always comes with the downside of it being a slog to read, so apologies for that!

I'll write a shortened version if I find the time!

Thanks James!

One failure mode is that the modification makes the model very dumb in all instances.

Yea, good point. Perhaps an extra condition we'd need to include is that the "difficulty of meta-level questions" should be the same before and after the modification - e.g. - the distribution over stuff it's good at and stuff its bad at should be just as complex (not just good at everything or bad at everything) before and after

Thanks Felix!

This is indeed a cool and surprising result. I think it strengthens the introspection interpretation, but without a requirement to make a judgement of the reliability of some internal signal (right?), it doesn't directly address the question of whether there is a discriminator in there.

Interesting question! I'm afraid I didn't probe the cruxes of those who don't expect hard takeoff. But my guess is that you're right - no hard takeoff ~= the most transformative effects happen before recursive self-improvement

Yea, I think you're hitting on a weird duality between setting and erasing here. I think I agree that setting is more fundamental than erasing. I suppose when talking about energy expenditure of computation, each set bit must be erased in the long run, so they're interchangeable in that sense.

Sorry for the delay. As both you and TheMcDouglas have mentioned; yea, this relies on $H(C|X) = 0$. The way I've worded it above is somewhere between misleading and wrong, have modified. Thanks for pointing this out!

Thanks for the comment, this is indeed an important component! I've added a couple of sentences pointing in this direction.