All of Shiroe's Comments + Replies

Shiroe10

The existence of God and Free Will feel like religious problems that philosophers took interest in, and good riddance to them.

Whether the experience of suffering/pain is fictional or not is a hot topic in some circles, but both sides are quite insistent about being good church-going "materialists" (whatever that means).

As for "knowledge", I agree that question falls apart into a million little subproblems. But it took the work of analytic philosophers to pull it apart, and after much labor. You're currently reaping the rewards of that work and the simplicity of hindsight.

Shiroe10

Do you know if Plato was claiming Euclidean geometry was physically true in that sense? Doesn't sound like something he would say.

Shiroe20

I'd like to see how this would compare to a human organization. Suppose individual workers or individual worker-interactions are all highly faithful in a tech company. Naturally, though, the entire tech company will begin exhibiting misalignment, tend towards blind profit seeking, etc. Despite the faithfulness of its individual parts.

Is that the kind of situation you're thinking of here? Is that why having mind-reading equipment that forced all the workers to dump their inner monologue wouldn't actually be of much use towards aligning the overall system, because the real problem is something like the aggregate or "emergent" behavior of the system, rather than the faithfulness of the individual parts?

2tailcalled
My threat model is entirely different: Even if human organizations are misaligned today, the human organizations rely primarily on human work, and so they pass tons of resources and discretion on to humans, and try to ensure that the humans are vigorous. Meanwhile, under @mattmacdermott 's alignment proposal, one would get rid of the humans and pass tons of resources and discretion on to LLMs. Whether one values the LLMs is up to you, but if human judgement and resource ownership is removed, obviously that means humans lose control, unless one can change organizations to give control to non-employees instead of employees (and it's questionable how meaningful that is).
Shiroe10

What do you mean by "over the world"? Are you including human coordination problems in this?

2tailcalled
CoT is generally done for one level of data processing, e.g. responding to a user in a chat, or performing one step of a task for an autonomous LLM agent. However usually when AIs are deployed, they are asked to do many levels of data processing, e.g. respond to many users or perform many steps sequentially. It doesn't matter if the chains of thoughts are highly faithful for pretty much all of these levels individually, what matters is if they are faithful in aggregate, i.e. for the few most important levels of data processing as well as in how the overall wave of smaller interactions add up.
Shiroe30

Did you end up writing the list of interventions? I'd like to try some of them. (I also don't want to commit to doing 3 hours a day for two weeks until I know what the interventions are.)

Shiroe32

It's very surprising to me that he would think there's a real chance of all humans collectively deciding to not build AGI, and successfully enforcing the ban indefinitely.

Shiroe10

Patternism is usually defined as a belief about the metaphysics of consciousness, but that boils down to incoherence, so it's better defined as a property of a utility function of agents not minding being subjected to major discontinuities in functionality, ie, being frozen, deconstructed, reduced to a pattern of information, reconstructed in another time and place, and resumed.

That still sounds like a metaphysical belief, and less empirical since consciousness experience isn't involved in it (instead it sounds like it's just about personal identity).

Shiroe20

Any suggestions for password management?

2Sinclair Chen
Whatever you use, remember to backup your vault regularly. A cautionary tale: I lost access to my bitwarden vault containing a private key to a few thousand $ worth of crypto, after changing my master password to something that I was then not able to recall perfectly.  And bitwarden's website / extension start to rate limit you client-side after failed attempts. So instead, after a lot of research I was able to find the bitwarden hashfile on my computer where chrome stores data for its extensions. I then downloaded hashcat and tried to do a dictionary attack and some other clever attacks that made use of what I thought my password was supposed to be, but to no success. Don't be me. Bitwarden lets you download your encrypted vault from the website or CLI. do that.
3kave
I really like 1Password, but my understanding is that Bitwarden has less frequent reported vulnerabilities
8JanGoergens
I recommend Keepass, but you might have different requirements. This video serves as a good comparison of your options and on this website you can find a list of recommendation for password managers and other privacy/security tools.
6hazel
I've been well served by Bitwarden: https://bitwarden.com/ It has a dark theme, apps for everything (including Linux commandline), the Firefox extension autofills with a keyboard shortcut, plus I don't remember any large data breaches.
Shiroe20

Because it's an individualized approach that is a WIP and if I just write it down 99% of people will execute it badly.

Why is that a problem? Do you mean this in the sense of "if I do this, it will lead to people making false claims that my experiment doesn't replicate" or "if I do this, nothing good will come of it so it's not even worth the effort of writing".

As someone who runs a lot of self-experiments and occasionally helps others, I'm disappointed in but sympathetic to this approach. People are complicated: the right thing to do probably is try a bunch of stuff and see what sticks. But people really, really want the answer to be simple, and will round down complicated answers until they are simple enough, then declare the original protocol a failure when their simplification doesn't work.  

I think it would be valuable for George to write up the list of interventions they considered, and a case report o... (read more)

2George3d6
I mean if I write this it will sound very weird and not be followable because it includes things like: I am trying to replicate this with more people right now so I'd rather not dilute the intervention specifically -- hence why this post was not about what I did as much as why one ought to expect increasing IQ, in general, works.
Shiroe30

I'm confused whether:

  1. the point of this article is that the IQ tests are broken, because some trivial life improvements (like doing yoga and eating blueberries) will raise your IQ score or whether:
  2. the point of this article is that you actually raised your "g" by doing trivial life improvements, and we should be excited by how easy it is to become more intelligent

Skimming it again I'm pretty sure you mean (2).

0George3d6
Somewhere in between actually, I tried to do something like (2) but in part I'm sure it's (1) I avoided any conceptual/learning tasks and just did brain stimulation, non-stimulant drugs and various physical practices to avoid (1) as much as possible You can toally do n-back training or take IQ tests to increase your IQ, and it's pretty boring.
Shiroe20

If I understand right the last sentence should say "does not hold".

2FlorianH
Thanks, corrected
Shiroe10

It's not easy to see the argument for treating your vales as incomparable with the values of other people, but seeing your future self's values as identical to your own. Unless you've adopted some idea of a personal soul.

Shiroe71

The suffering and evil present in the world has no bearing on God's existence. I've always failed to buy into that idea. Sure, it sucks. But it has no bearing on the metaphysical reality of a God. If God does not save children--yikes I guess? What difference does it make? A creator as powerful as has been hypothesised can do whatever he wants; any arguments from rationalism be damned.

Of course, the existence of pointless suffering isn't an argument against the existence of a god. But it is an old argument against the existence of a god who deserves to b... (read more)

Answer by Shiroe44

"tensorware" sprang to mind

2johnswentworth
This one independently sprang to mind for me too.
Shiroe10

Yeah, it's hard to say whether this would require restructuring the whole reward center in the brain or if the needed functionality is already there, but just needs to be configured with different "settings" to change the origin and truncate everything below zero.

My intuition is that evolution is blind to how our experiences feel in themselves. I think it's only the relative differences between experiences that matter for signaling in our reward center. This makes a lot of sense when thinking about color and "qualia inversion" thought experiments, but it's trickier with valence. My color vision could become inverted tomorrow, and it would hardly affect my daily routine. But not so if my valences were inverted.

Shiroe10

What about our pre-human ancestors? Is the twist that humans can't have negative valences either?

Shiroe42

I agreed up until the "euthanize everything that remains" part. If we actually get to the stage of having aligned ASI, there are probably other options with the same or better value. The "gradients of bliss" that I described in another comment may be one.

Shiroe51

Pearce has the idea of "gradients of bliss", which he uses to try to address the problem you raised about insensitivity to pain being hazardous. He thinks that even if all of the valences are positive, the animal can still be motivated to avoid danger if doing so yields an even greater positive valence than the alternatives. So the prey animals are happy to be eaten, but much more happy to run away.

To me, this seems possible in principle. When I feel happy, I'm still motivated at some low level to do things that will make me even happier, even though I was... (read more)

2RogerDearnaley
Whatever positive valence stopped when you were injured would need to be as extremely strong a motivator as pain is. So somewhere on the level of "I orgasm continuously unless I get hurt, then it stops!" That's just shifting the valence scale: I think by default it would fail due to hedonic adaptation — brains naturally reset their expectations. That's the same basic mechanism as opiate addition, and it's pretty innate to how the brain (or any complex set of biochemical pathways) works: they're full of long-term feedback loops evolved to try to keep them working even if one component is out-of-whack, say due to a genetic disease. This is related to a basic issue in the design of Utilitarian ethical systems. As is hopefully well-known, you need your AI to maximize the amount of positive utility (pleasure) not minimize the amount of negative utility (pain), otherwise it will just euthanize everyone before they can next stub their toe. (Obviously getting this wrong is an x-risk, as with pretty-much everything in basic ethical system design,) So you need to carefully set a suitable zero utility level, and that level needs to be low enough that you actually would want the AI to euthanize you if your future utility level for the rest of you life was going to be below that level. So that means the negative utility region is the sort of agonizing pain level where we put animals down, or allow people to sign paperwork for voluntary medical euthanasia. That's a pretty darned low valence level, well below what we start calling 'pain'. On a hospital numerical "how much pain are you in?" scale, it's probably somewhere around spending the rest of your life at an eight or worse: enough pain that you can't pay much attention to anything else ever. So my point is, if you just stubbed your toe and are in pain (say a six on the hospital pain scale), then by that offset scale of valence levels (which is what our AIs have to be using for utility in their ethical systems), your utility
Shiroe40

What are your thoughts on David Pearce's "abolitionist" project? He suggests genetically engineering wild animals to not experience negative valences, but still show the same outward behavior. From a sentientist stand-point, this solves the entire problem, without visibly changing anything.

2Shankar Sivarajan
Good news! That's already the way the world works.
7RogerDearnaley
I think it's basically impossible, using just genetic engineering. There are documented cases of humans born without the ability to feel pain, and they don't usually live long: they tend to die in stupid accidents, like jumping off a building, because they didn't learn the lesson as a kid that that hurts so you shouldn't do it. Or similarly in leprosy, where the ability to feel pain is lost due to bacterial damage to the nerves (typically as an adult once you have learnt not to do that sort of dumb stunt), the slow progressive disfiguring damage to hands and face from leprosy isn't directly bacterial, it's caused by the cumulative effect of a great many minor injuries that the patient doesn't notice in time, because they can't feel pain any more. So, producing the same behavior without negative valences would require a much larger, more detailed world model, able to correctly predict everything that would have hurt or been unpleasant and how the creature would have reacted to it, and then trigger that reaction. Even assuming you can somehow achieve that modelling task in a nervous system as a "philosophical zombie" involving no actual negative valences, just a prediction of their effects on an animal (it's very unclear to me how to even tell, I suspect "philosophical zombies" are a myth, and if they're not then they're a-priori indistinguishable), then we currently have no idea how to bioengineer something like that, and clearly the extra nervous tissue required to do all the extra processing would add a lot to physiological needs. The most plausible approach I can think of to achieve this would be some sort of nanotech cyborging where the extra processing was done in the cyborg parts, which would need to be much more compact and energy efficient than nervous tissue (i.e. roughly Borg-level technology). So it's an emotionally appealing idea, but I suspect actually even harder to implement than what I proposed. For largish animals, it might actually be technological
Shiroe43

Same. I feel somewhat jealous of people who can have a visceral in-body emotional reaction to X-risks. For most of my life I've been trying to convince my lizard brain to feel emotions that reflect my beliefs about the future, but it's never cooperated with me.

Shiroe10

You can compress huge prompts into metatokens, too (just run inference with the prompt to generate the training data)

I'm very curious about this technique but couldn't find anything about it. Do you have any references I can read?

4porby
Alas, nope! To my knowledge it hasn't actually been tried at any notable scale; it's just one of those super simple things that would definitely work if you were willing to spend the compute to distill the behavior.
Shiroe10

I see. Yes, "philosophy" often refers to particular academic subcultures, with people who do their philosophy for a living as "philosophers" (Plato had a better name for these people). I misread your comment at first and thought it was the "philosopher" who was arguing for the instrumentalist view, since that seems like their more stereotypical way of thinking and deconstructing things (whereas the more grounded physicist would just say "yes, you moron, electrons exist. next question.").

2dr_s
From the discussion it seemed that most physicists do take the realist view on electrons, but in general the agreement was that either view works and there's not a lot to say about it past acknowledging what everyone's favorite interpretation is. A question that can have no definite answer isn't terribly interesting.
Shiroe10

Do you have any examples of the "certain philosophers" that you mentioned? I've often heard of such people described that way, but I can't think of anyone who's insulted scientists for assuming e.g. causality is real.

2dr_s
For example there's recently been a controversy adjacent to this topic on Twitter involving one Philip Goff (philosopher) who started feuding over it with Sabine Hossenfelder (physicist, albeit with some controversial opinions). Basically Hossenfelder took up an instrumentalist position of "I don't need to assume that things described in the models we use are real in whatever sense you care to give to the word, I only need to know that those models' predictions fit reality" and Goff took issue with how she was brushing away the ontological aspects. Several days of extremely silly arguments about whether electrons exist followed. To me Hossenfelder's position seemed entirely reasonable, and yes, a philosophical one, but she never claimed otherwise. But Goff and other philosophers' position seemed to be "the scientists are ignorant of philosophy of science, if only they knew more about it, they would be far less certain about their intuitions on this stuff!" and I can't understand how they can be so confident about that or in what way would that precisely impact the scientists' actual work. Whether electrons "exist" in some sense or they are just a convenient mathematical fiction doesn't really matter a lot to a physicist's work (after all, electrons are nothing but quantized fluctuations of a quantum field, just like phonons are quantized fluctuations of an elastic deformation field; yet people probably feel intuitively that electrons "exist" a lot more than phonons, despite them being essentially the same sort of mathematical object. So maybe our intuitions about existence are just crude and don't well describe the stuff that goes on at the very edge of matter).
Shiroe20

On the contrary, it is my intention to illustrate that assertions of instances that have not been experienced (with respect to their assertion at t1) can be justified in the future in which they are observed (with respect to their observation at t2).

 

Sorry, I may not be following this right. I had thought the point of the skeptical argument was that you can't justify a prediction about the future until it happens. Induction is about predicting things that haven't happened yet. You don't seem to be denying the skeptical argument here, if we still need to wait for the prediction to resolve before it can be justified.

2Krantz
This is a good question. I agree that you can't justify a prediction until it happens, but I'm urging us to consider what it actually means for a prediction to happen.  This can become nuanced when you consider predictions that are statements which require multiple observations to be justified. If I predict that a box (that we all know contains 10 swans) contains 10 white swans (My prediction is 'There are ten white swans in this box.').  When does that prediction actually 'happen'?  When does it become 'justified'? I think we all agree that after we've witnessed the 10th white swan, my assertion is justified. But am I justified at all to believe I am more likely to be correct after I've only witnessed 8 or 9 white swans?  This is controversial.
Shiroe41

I've also noticed that scaffolded LLM agents seem inherently safer. In particular, deceptive alignment would be hard for one such agent to achieve, if at every thought-step it has to reformulate its complete mind state into the English language just in order to think at all.

You might be interested in some work done by the ARC Evals team, who prioritize this type of agent for capability testing.

Shiroe10

I'm sorry that comparing my position to yours led to some confusion: I don't deny the reality of 3rd person facts. They probably are real, or at least it would be more surprising if they weren't than if they were. (If not, then where would all of the apparent complexity of 1st person experience come from? It seems positing an external world is a good step in the right direction to answering this). My comparison was about which one we consider to be essential. If I had used only "pragmatist" and "agnostic" as descriptors, it would have been less confusing.

A... (read more)

1tangerine
I don't think I disagree with what you're saying here, though we may be using different terms to say the same thing. How does what you say here inform your thoughts about the Hard Problem?
Shiroe10

If I had to choose between those two phrasings I would prefer the second one, for being the most compatible between both of our notions. My notion of "emerges from" is probably too different from yours.

The main difference seems to be that you're a realist about the third-person perspective, whereas I'm a nominalist about it, to use your earlier terms. Maybe "agnostic" or "pragmatist" would be good descriptors too. The third-person is a useful concept for navigating the first-person world (i.e. the one that we are actually experiencing). But that it seems u... (read more)

1tangerine
That actually sounds more like the first phrasing to me. If you are a nominalist about the third-person perspective, then it seems that you think the third-person perspective does not actually exist and the concept of the third-person perspective is borne of the first-person perspective. I’m not sure whether this is a good double crux, because it’s not clear enough to me what we mean by first- and third-person perspectives. It seems conceivable to me that my conception of the third-person perspective is functionally equivalent to your conception of the first-person perspective. Let me expand on that below. If only the first-person perspective exists, then presumably you cannot be legitimately surprised, because that implies something was true outside of your first-person perspective prior to your experiencing it, unless you define that as being part of your first-person perspective, which seems contradictory to me, but functionally the same as just defining everything from the third-person perspective. The only alternative possibility that seems available is that there are no external facts, which would mean reality is actually an inconsistent series of experiences, which seems absurd; then we wouldn’t even be able to be sure of the consistency of our own reasoning, including this conversation, which defeats itself.
Shiroe10

I meant subjective in the sense of "pertaining to a subject's frame of reference", not subjective in the sense of "arbitrary opinion". I'm sorry if that was unclear.

Shiroe*10

But all of these observations are also happening from a third-person perspective, just like the rest of reality.

This is a hypothesis, based on information in your first-person perspective. To make arguments about a third-person reality, you will always have to start with first-person facts (and not the other way around). This is why the first person is epistemologically more fundamental.

It's possible to doubt that there is a third-person perspective (e.g. to doubt that there's anything like being God). But our first person perspective is primary, and ca... (read more)

1tangerine
Would I be correct to say that you think the third-person perspective emerges from the first-person perspective? Or would you say that they’re simply separate?
Shiroe10

You don't believe that all human observations are necessarily made from a first-person viewpoint? Can you give a counter-example? All I can think of are claims that involve the paranormal or supernatural.

1tangerine
Like TAG said, in a trivial sense human observations are made from a first-person, subjective viewpoint. But all of these observations are also happening from a third-person perspective, just like the rest of reality. The way I see it, the third-person perspective is basically the default, i.e., reality is like a list of facts, from which the first-person view emerges. Then of course the question is, how is that emergence possible? I can understand the intuition that the third-person and first-person view seem to be fundamentally different, but I think of it this way: all the thoughts you think and the statements you make are happening in reality and the structure of that reality determines your thoughts. This is where the illusion arguments become relevant; illusions, such as optical ones, demonstrate clearly that you can be made to believe things that are wrong because of convenience or simply brain malfunction. Changing the configuration of your brain matter can make you believe absolutely anything. The belief in the first-person perspective has evolved because it’s just very useful for survival and you can’t choose to disbelieve what your brain makes you believe. Given the above, to say that the first-person perspective is fundamentally different seems like the more supernatural claim to me.
2TAG
Observations being made from a first person perspective is a rather trivial definition of subjective,. because it's quite possible for different observers to agree on observations.(And for some aspects of the persepective to be predictable from objective facts).The forms of subjectivity that count are where they disagree, or where they can't even express their perceptions.
Shiroe10

I don't think I fall into either camp because I think the question is ambiguous. It could be talking about the natural structure of space and time ("mathematics") or it could be talking about our notation and calculation methods ("mathematics"). The answer to the question is "it depends what you mean".

The nominalist vs realist issue doesn't appear very related to my understanding of the Hard Problem, which is more about the definition of what counts as valid evidence. Eliminitivism says that subjective observations are problematic. But all observations are subjective (first person), so defining what counts as valid evidence is still unresolved.

1tangerine
What exactly do you mean by this? That nature is mathematical? This sounds like it could be a double crux, because if I believed this the Hard Problem would follow trivially, but I don’t believe it.
Shiroe10

I appreciate hearing your view; I don't have any comments to make. I'm mostly interested in finding a double crux.

This isn't really a double crux, but it could help me think of one:

If someone becomes convinced that there isn't any afterlife, would this rationally affect their behavior? Can you think of a case where someone believed in Heaven and Hell, had acted rationally in accordance with that belief, then stopped believing in Heaven and Hell, but still acted just the same way as they did before? We're assuming their utility function hasn't changed, just their ontology.

1tangerine
Well, for me, one crux is this question of nominalism vs philosophical realism. One way to investigate this question for yourself is to ask whether mathematics is invented (nominalism) or discovered (philosophical realism). I don’t often like to think in terms of -isms, but I have to admit I fall pretty squarely in the nominalist camp, because while concepts and words are useful tools, I think they are just that: tools, that we invented. Reality is only real in a reductionist sense; there are no people, no numbers and no consciousness, because those are just words that attempt to cope with the complexity of reality, so we just shouldn’t take them so seriously. If you agree with this, I don’t see how you can think the Hard Problem is worth taking seriously. If you disagree, I’m interested to see why. If you could convince me that there is merit to the philosophical realist position, I would strongly update towards the Hard Problem being worth taking seriously.
Shiroe10

Here are some cruxes, stated from what I take to be your perspective:

  1. That there's nothing at stake whether or not we have first person experiences of the kind that eliminitivists deny; it makes no practical difference to our lives whether we're so-called "automatons" or "zombies", such terms being only theoretical distinctions. Specifically it should make no difference to a rational ethical utilitarian whether or not eliminitivism happens to be true. Resources should be allocated the same way in either case, because there's nothing at stake.
  2. Eliminitivis
... (read more)
1tangerine
From my point of view, much or all of the disagreement around the existence of the Hard Problem seems to boil down to the opposition between nominalism and philosophical realism. I’ll discuss how I think this opposition applies to consciousness, but let me start by illustrating it with the example of money having value. In one sense, the value of money is not real, because it's just a piece of paper or metal or a number in a bank’s database. We have systems in place such that we can track relatively consistently that if I work some number of hours, I get some of these pieces of paper or metal or the numbers on my bank account change in some specific way and I can go to a store and give them some of these materials or connect with the bank’s database to have the numbers decrease in some specific way, while in exchange I get a coffee or a t-shirt or whatever. But this is a very obtuse way of communicating, so we just say that “money has value” and everybody understands that it refers to this system of exchanging materials and changing numbers. So in the case of money, we are pretty much all nominalists; we say that money has value as a shorthand and in that sense the value of money is real. On the other hand, a philosophical realist would say that actually the value of money is real independently from our definition of the words. (I view this idea similarly to how Eliezer Yudkowsky talks about buckets being “magical” in this story.) In the case of the value of money, philosophical realism does not seem to be a common position. However, when it comes to consciousness, the philosophical realist position seems much more common. This strikes me as odd, since both value and consciousness appear to me to originate in the same way; there is some physical system which we, through the evolution of language and culture generally, come to describe with shorthands (i.e., words), because reality is too complicated to talk about exhaustively and in most practical matters we all u
Shiroe10

because such sensations would be equivalent to predictions that I would be burning alive, which would be false and therefore interfere with my functioning

I don't see a necessary equivalence here. You could be fully aware that the sensations were inaccurate, or hallucinated. But it would still hurt just as much.

if you could have a body which doesn’t experience, then it’s not going to function as normal.

A human body, or any kind of body? It seems like a robot could engage in the same self-preservation behavior as a human without needing to have anythi... (read more)

1tangerine
The experience of hurting makes you respond as if you really were hurting; you have some voluntary control over your response by the frontal cortex’ modulation of pain signals, but it is very limited. Any control we exert over our experiences corresponds to physical interventions. The Hard Problem simply does not add anything of value here. That you can imagine such a prosthesis does not mean that it could exist. It depends on how such a prosthesis would work exactly. I suspect that the more such a prosthesis was able to mimick the normal response, the more its wearer would experience pain, i.e., inducing the normal response is equivalent to inducing the normal experience.
Shiroe10

You seem to be claiming that you have experiences, but that their role is purely functional. If you were to experience all tactile sensations as degrees of being burnt alive, but you could still make predictions just as well as before, it wouldn't make any difference to you?

3tangerine
It doesn’t make sense to say that I could make predictions just as well as before if I experienced all tactile sensations as degrees of being burnt alive, because such sensations would be equivalent to predictions that I would be burning alive, which would be false and therefore interfere with my functioning. You can’t separate experience from its consequences. That’s also why philosophical zombies are impossible; if you could have a body which doesn’t experience, then it’s not going to function as normal. If I were to experience all tactile sensations as degrees of being burnt alive, I would assume something was wrong with my body and I would want to alleviate that situation by making predictions about which actions would alleviate it by wielding only concepts related to the Easy Problem. How would the Hard Problem help me in that situation?
Shiroe10

It's plausible that reverse-engineering the human mind requires tools that are much more powerful than the human mind.

Shiroe1-2

So you don't believe there is such a thing as first-person phenomenal experiences, sort of like Brian Tomasik? Could you give an example or counterexample of what would or wouldn't qualify as such an experience?

3Gordon Seidoh Worley
I think that there's a process we can meaningfully point to and call qualia, and it includes all the things we think of as qualia, but qualia is not itself a thing per se but rather the reification of observations of mental processes that allows us to make sense of them. I have theories of what these processes are and how they work and they mostly line up with the what's pointed at by this book. In particular I think cybernetic models are sufficient to explain most of the interesting things going on with consciousness, and we can mostly think of qualia as the result of neurons in the brain hooked up in loops so that their inputs include information not only from other neurons but also from themselves, and these self-sensing loops provide the input stream of data that other neurons interpret as self-experience/qualia/consciousness.
Shiroe21

Doesn't "direct" have the implication of "certain" here?

2TAG
Some people think so, other don't. Indirectness adds extra uncertainty, but it's not the only possible source of uncertainty.
Shiroe10

Response in favor of the assumption that Signer said was detrimental.

Shiroe*30

but my current theory is that one such detrimental assumption is "I have direct knowledge of content of my experiences"

It's true this is the weakest link, since instances of the template "I have direct knowledge of X" sound presumptuous and have an extremely bad track record.

The only serious response in favor of the presumptuous assumption [edit] that I can think of is epiphenomenalism in the sense of "I simply am my experiences", with self-identity (i.e. X = X) filling the role of "having direct knowledge of X". For explaining how we're able to have co... (read more)

2Signer
Personally I wouldn't say "I am my experiences" is epiphenomenalism - I have a casual role.
2TAG
Response to what?
Shiroe20

The burden of proof is on those who assert that the Hard Problem is real. You can say what consciousness is not, but can you say what it is?

In the sense that you mean this, this is a general argument against the existence of everything, because ultimately words have to be defined either in terms of other words or in terms of things that aren't words. Your ontology has the same problem, to the same degree or worse. But we only need to give particular examples of conscious experience, like suffering. There's no need to prove that there is some essence of ... (read more)

1tangerine
Hi, please see my reply to gilch above. To add to that reply, an explanation only ever serves one function, namely to aid in prediction; every moment of our life, we try to achieve outcomes by predicting which action will lead to which outcome. An explanation to the Hard Problem doesn’t do that. Any state of consciousness that I try to achieve I do so with concepts related to the Easy Problem. I do have experiences (I don’t know what the word “phenomenal” would add to that), such as pain, but to the extent that I can predict and control these, I do so purely with solutions to the Easy Problem. And in my book, concepts that exist only in explanations that don’t aid in prediction are by definition not real. But the Hard Problem is even worse than that; it’s set up so that we can’t tell the difference between a correct and incorrect explanation in the first place, which means literally anything could be an explanation, which is equivalent to no explanation at all. Sure, you can choose to believe that something like panpsychism is real or that it’s not real, but because neither belief adds any predictive power, you’re better off just cutting it out, as per Occam’s Razor.
Shiroe10

Are you saying that you don't think there's any fact of the matter whether or not you have phenomenal experiences like suffering? Or do you mean that phenomenal experience is unreal in the same way that the hellscape described by Dante is unreal?

3Shankar Sivarajan
Closer to the latter, though I wouldn't call it "unreal." The experience of suffering exists in the same sense that my video game character is down to one life and has low HP: that state is legibly inspectable on screen, and is encoded in silicon. I only scorn the term "consciousness" to the extent it is used as woo. I think some version of the Hard Problem really was meaningful in the past, and that it was hard: it's far from obvious how "mere matter" could encode states as rich as what one … perceives, experiences, receives as the output of the brain's processes. Mills and clocks, as sophisticated as they may have been, were still far too simple. I consider modern technology to have demonstrated in sufficient detail precisely how it's possible. It didn't require anything conceptually new, so also I understand why some don't find the answer satisfying.
Shiroe12

I don't like "illusionism" either, since it makes it seem like illusionists are merely claiming that consciousness is an illusion, i.e., it is something different than what it seems to be. That claim isn't very shocking or novel, but illusionists aren't claiming that. They're actually claiming that you aren't having any internal experience in the first place. There isn't any illusion.

"Fictionalism" would be a better term than "illusionism": when people say they are having a bad experience, or an experience of saltiness, they are just describing a fictional character.

Shiroe11

Exactly. I wish the economic alignment issue was brought up more often.

Shiroe10

You're right. I'm updating towards illusionism being orthogonal to anthropics in terms of betting behavior, though the upshot is still obscure to me.

Shiroe10

I agree realism is underrated. Or at least the term is underrated. It's the best way to frame ideas about sentientism (in the sense of hedonic utilitarianism). On the other hand, you seem to be talking more about rhetorical benefits of normative realism about laws.

Most people seem to think phenomenal valence is subjective, but that's confusing the polysemy of the word "subjective", which can mean either arbitrary or bound to a first-person subject. All observations (including valenced states like suffering) are subjective in the second sense, but not in th... (read more)

Shiroe10

it is easy to cooperate on the shared goal of not dying

Were you here for Petrov Day? /snark

But I'm confused what you mean about a Pivotal Act being unnecessary. Although both you and a megacorp want to survive, you each have very different priors about what is risky. Even if the megacorp believes your alignment program will work as advertised, that only compels them to cooperate with you if they are (1) genuinely concerned about risk in the first place, (2) believe alignment is so hard that they will need your solution, and (3) actually possess the institutional coordination abilities needed.

And this is just for one org.

Shiroe10

World B has a 1, maybe minus epsilon chance of solving alignment, since the solution is already there.

That is totum pro parte. It's not World B which has a solution at hand. It's you who have a solution at hand, and a world that you have to convince to come to a screeching halt. Meanwhile people are raising millions of dollars to build AGI and don't believe it's a risk in the first place. The solution you have in hand has no significance for them. In fact, you are a threat to them, since there's very little chance that your utopian vision will match up wit... (read more)

1Noosphere89
I do not think a pivotal act is necessary, primarily because it's much easier to coordinate around negative goals like preventing their deaths than positive goals. That's why I'm so optimistic, it is easy to cooperate on the shared goal of not dying even if value differences after that are large.
Shiroe*10

Okay, let's operationalize this.

Button A: The state of alignment technology is unchanged, but all the world's governments develop a strong commitment to coordinate on AGI. Solving the alignment problem becomes the number one focus of human civilization, and everyone just groks how important it is and sets aside their differences to work together.

Button B: The minds and norms of humans are unchanged, but you are given a program by an alien that, if combined with an AGI, will align that AGI in some kind of way that you would ultimately find satisfying.

World ... (read more)

1Noosphere89
I actually think A or B is a large improvement compared to the world as it exists today, but B wins due to the stakes and the fact that they already have the solution, but world A doesn't have the solution pre-loaded, and with extremely important decisions, B wins over A. World A is much better than today, to the point that a civilizational scale effort would probably succeed about 95-99.9% of the time, primarily because they understand deceptive alignment. World B has a 1, maybe minus epsilon chance of solving alignment, since the solution is already there. Both, of course are far better than our world.
Shiroe10

I agree that the political problem of globally coordinating non-abuse is more ominous than solving technical alignment. If I had the option to solve one magically, I would definitely choose the political problem.

What it looks like right now is that we're scrambling to build alignment tech that corporations will simply ignore, because it will conflict with optimizing for (short-term) profits. In a word: Moloch.

1Noosphere89
I would choose the opposite, because I think the consequences of the first drastically outweigh the second.
Load More