All of MattJ's Comments + Replies

MattJ11

That doesn’t make sense to me. If someone wants to fool me that I’m looking att a tree he has to paint a tree in every detail. Depending on how closely I examine this tree he has to match my scrutiny to the finest detail. In the end, his rendering of a tree will be indistinguishable from an actual tree even at the molecular level.

2AnthonyC
This does not imply that the simulation is run entirely in linear time, or at a constant frame rate (or quivalent), or that details are determined a priori instead of post hoc. It is plausible such a system could run a usually-convincing-enough simulation at lower fidelity, back-calculate details as needed, and modify memories to ignore what would have been inconsistencies when doing so is necessary or just more useful/tractable. 'Full detail simulation at all times' is not a prerequisite for never being able to find and notice a flaw, or for getting many kinds of adequate high level macroscopic outputs. In other words: If I want to convince you something is a real tree, it needs to look and feel like a tree, but it doesn't need an exact, well-defined wave-function. Classical approximations at tens of microns scale are about the limit of unaided human perception. If you pull out a magnifying glass or a scanning electron microscope, then you can fill in little pieces of the remaining whole, but you still aren't probing the whole tree down to the Planck scale.

In your dreams do you ever see trees you think are real? I doubt your brain is simulating the trees at a very high level of detail, yet this dream simulation can fool you.

MattJ1416

We don’t want an ASI to be ”democratic”. We want it to be ”moral”. Many people in the West conflate the two words thinking that democratic and moral is the same thing but it is not. Democracy is a certain system of organizing a state. Morality is how people and (in the future) an ASI behave towards one another.

There are no obvious reasons why an authocratic state would care more or less about a future ASI being immoral, but an argument can be made that autocratic states will be more cautious and put more restrictions on the development of an ASI because autocrats usually fear any kind of opposition and an ASI could be a powerful adversary of itself or in the hands of powerful competitors.

2GeneSmith
I'm not sure I buy that they will be more cautious in the context of an "arms race" with a foreign power. The Soviet Union took a lot of risks their bioweapons program during the cold war. My impression is the CCP's number one objective is preserving their own power over China. If they think creating ASI will help them with that, I fully expect them to pursue it (and in fact to make it their number one objective)

I think "democratic" is often used to mean a system where everyone is given a meaningful (and roughly equal) weight into it decisions. People should probably use more precise language if that's what they mean, but I do think it is often the implicit assumption.

And that quality is sort of prior to the meaning of "moral", in that any weighted group of people (probably) defines a specific morality - according to their values, beliefs, and preferences. The morality of a small tribe may deem it as a matter of grave importance whether a certain rock has been tou... (read more)

MattJ10

The actual peace deal will be something for the Ukraine to agree to. It is not up to Trump to dictate the terms. All Trump should do is to stop financing the war and we will have peace.

Having said that, if it is somehow possible for Trump to pressure Ukraine into agreeing to become a US colony, my support for Trump was a mistake. The war would be preferable to the peace.

MattJ-1-5

Good post! We will soon have very powerful quantum computers that probably could simulate what will happen if a mirror bacteria is confronted with the human immune system. Maybe there is no risk at all or an existential risk to humanity. This should be a prioritized task for our first powerful quantum computer to find out.

7k64
For me, a candidate's claim of what they will do is sufficient when they have unilateral control over doing it.  For instance, I believe a claim to sign or veto a specific type of bill.  I don't tend to believe that they will make the economy good, avoid recession, close all the tax loopholes, etc.   Do you: a) believe candidates when they claim they will be successful at things not entirely in their control b) believe Trump but not others (like Kamala) when they claim they'll do things not entirely in their control c) think that a Russia-Ukraine peace negotiation would be entirely in Trump's control d) see some actions that are entire in Trump's control that you are confident would cause peace e) other
MattJ40

I’m not allowed to vote in the election but I hope Trump wins because I think he will negotiate a peace in Ukraine. If Harris wins I think the war will drag on for another couple of years at worst.

I have no problem getting pushback.

3Anon User
It would seem that my predictions of how Trump would approach this were pretty spot on... @MattJ I am curious what's your current take on it?
5Anon User
Do you care about what kind of peace it is, or just that there is some sort of peace? If latter, I might agree with you on Trump being more likely to quickly get us there. For former, Trump is a horrible choice. On of the easiest way for a US President to force a peace agreement in Ukraine is probably to privately threaten Ukranians to withhold all support, unless they quickly agree to Russian demands. IMHO, Trump is very likely to do something like that. The huge downside is that while this creates a temporary peace, it would encourage Russia to go for it again with other neighbors,and to continue other destabilizing behaviors across the globe (in collaboration with China, Iran, North Korea, etc). Also increases the chances of China going at Taiwan.
6k64
Why do you believe that Trump will negotiate a peace?
MattJ-10

I guess it could be a great tool to help people quickly learn to converse in a foreign language.

MattJ10

The ”eternal recurrence” is surprisingly the most attractive picture of the ”afterlife”. The alternatives: the annihilation of the self or eternal life in heaven are both unattractive, for different reasons. Add to this that Nietzsche is right to say that the eternal recurrence is a view of the world which seems compatible with a scientific cosmology.

MattJ30

I remember I came up with a similar thought experiment to explain the Categorical Imperative.

Assume there is only one Self-Driving Car on the market, what principle would you want it to follow?

The first priciple we think of is: ”Always do what the driver would want you to do”.

This would certainly be the principle we would want if our SDC was the only car on the road. But there are other SDCs and so in a way we are choosing a principle for our own car which is also at the same time a ”universal law”, valid for every car on the road.

With this in mind, it is ... (read more)

MattJ2-1

Yes, but that ”generative AI can potentially replace millions of jobs” is not contradictory to the statement that it eventually ”may turn out to be a dud”.

I initially reacted in the same way as you to the exact same passage but came to the conclusion that it was not illogical. Maybe I’m wrong but I don’t think so.

MattJ10

I think the auther ment that there was a perception that it could replace millions of jobs, and so an incentive for business to press forward with their implementation plans, but that this would eventually back fire if the hallucination problem is insoluble.

2Quintin Pope
Perhaps, but that's not the literal meaning of the text.
MattJ1-1

I am a Kantian and believe that those a priori rules have already been discovered.

But my point here was merely that you can isolate the part that belongs to pure ethics from evererything empirical, like in my example what a library is; why do people go to libraries; what is a microphone and what is it’s purpose and so on. What makes an action right or wrong at the most fundamental level however is independent of everything empirical and simply an a priori rule.

I guess also my broader point was that Stephen Wolfram is far too pessimistic about the prospects... (read more)

2Mitchell_Porter
Does it boil down to the categorical imperative? Where is the best exposition of the rules, and the argument for them? 
MattJ21

You can’t isolate individual ”atoms” in ethics, according to Wolfram. Let’s put that to the test. Tell me if the following ”ethical atoms” are right or wrong:

  1. I will speak in a loud voice

2…on a monday

3…in a public library

4…where I’ve been invited to speak about my new book and I don’t have a microphone.

Now, (1) seems morally permissible, and (2) doesn’t change the evaluation. (3) does make my action seem morally impermissible, but (4) turns it around again. I’m convinced all of this was very simple to everyone.

Ethics is the science about the a priori ru... (read more)

3Mitchell_Porter
An optimistic view. Any idea how to figure out what they are?
MattJ10

I think that was the point. Comedians of the future will be performers. They will not write their own jokes but be increasingly good at reading the lines written by AI.

When Chat GPT came out I asked it to write a Seinfeld episode about taking an IQ-test. In my judgment it was just as good and funny as every other Seinfeld episodes I’ve watched. Mildly amusing. Entertaining enough not to stop reading.

MattJ10

This answer is a little bit confusing to me. You say that ”agency” may be an important concept even if we don’t have a deep understanding of what it entails. But how about a simple understanding?

I thought that when people spoke about ”agency” and AI, they meant something like ”a capacity to set their own final goals”, but then you claim that Stockfish could best be understood by using the concept of ”agency”. I don’t see how.

I myself kind of agree with the sentiment in the original post that ”agency” is a superfluous concept, but want to understand the opposite view?

1Max H
I didn't claim that Stockfish was best understood by using the concept of agency. I claimed agency was one useful model. Consider the first diagram in this post on embedded agency: you can regard Stockfish as Alexi playing a chess video game. By modeling Stockfish as an agent in this situation, you can abstract its internal workings somewhat and predict that it will beat you at chess, even if you can't predict the exact moves it makes, or why it makes those moves. > I thought that when people spoke about ”agency” and AI, they meant something like ”a capacity to set their own final goals” I think this is not what is meant by agency. Do I have the capacity to set (i.e. change) my own final goals? Maybe so, but I sure don't want to. My final goals are probably complex and I may be uncertain about what they are, but one day I might be able to extrapolate them into something concrete and coherent. I would say that an agent is something which is usefully modeled as having any kind of goals at all, regardless of what those goals are. As the ability of a system to achieve those goals increases, the model as an agent gets more useful and predictive relative to other models based on (for example) the system's internal workings. If I want to know whether Kasparov is going to beat me at chess, a detailed model of neuroscience and a scan of his brain is less useful than modeling him as an agent whose goal is to win the game. Similarly, for Mu Zero, a mechanistic understanding of the artificial neural networks it is comprised of is probably less useful than modeling it as a capable Go agent, when predicting what moves it will make.
2Chris_Leong
To clarify, I wasn't arguing it was a superficial concept, just trying to collect all the possible reasons together and nudge people towards making the use cases/applications more precise.
MattJ10

You seem to hold the position that:

  1. Scientists and not philosophers should do meta-ethics and normative ethics, until
  2. AGIs can do it better at which point we should leave it to them.

I don’t believe that scientists either have the inclination or the competence to do what you ask of them, and secondly that letting AGIs decide right and wrong would be a nightmare scenario for the human race.

1Roman Leventov
Normative ethics -- yes, because I gravitate towards ethical naturalism myself (see my discussion of scale-free ethics here), which is a part of the "package" of scientism. A scientist (as a role, not a person!) "shouldn't" do meta-ethics (that is, decide that ethical naturalism is the way to go), because the question of meta-ethics, and acceptance or rejection of such fundamental philosophical stances as scientism, or idealism, or postmodernism is outside of the scope of science, that is, cannot be settled with methods of science. Ultimately, every scientist must do at least a little bit of philosophy (of science), at which moment they assume the role of a philosopher. Scientism is a philosophy that maximises the scope of science as much as possible and minimises the scope of philosophy as much as possible, but not to zero. But regardless of who should or shouldn't do meta-ethics, I claim that technical alignment is impossible with anything except naturalistic ethics. That is, to successfully technically align AI to anyone or anything, one must take on a naturalistic theory of ethics. This is because engineering success is defined in scientific terms, thus if you don't treat ethics as a science (which is a synonym for ethical naturalism), you can't say that you technically succeeded at alignment. From the practical point of view, attempting to align AI to haphazard "values" or arbitrary "philosophical" theory of ethics rather than a coherent scientific theory of ethics seems bonkers, too. AGI can definitely do it much faster. And it seems that this is the strategy of both OpenAI and Conjecture and quite possibly other AGI labs too, to first build AGIs and then task them with "solving alignment" rather than recursive self-improvement. I don't try to estimate whether this strategy is better or worse than other strategies (at least in this post), I just take it as a premise because it seems very unlikely to me at this point that the aforementioned AGI labs will ch
MattJ10

”Whom the gods would destroy, they first make mad.”

One way to kill everyone would be to turn the planet into a Peoples Temple and have everyone drink the kool-aid.

I think even those of us who think of ouselves as psychologically robust will be defenseless against the manipulations of a future GPT-9.

2avturchin
AI can make drink everyone kool-aid, but soon its computers will stop so utility of doing that is small.
MattJ10

I thought that was what was meant. The question is probably the easiest one to answer affirmatively with a high degree of confidence. I can think of several ongoing ”moral catastrophs”.

MattJ51

Very good, fundamental questions.. I don’t understand question 85 though. Here are two more good questions.

  1. Are human beings aligned?
  2. Is human alignment, insofar as it exists, a property of the goals we have when we act or a property of the actions themselves?
1RohanS
Basic idea of 85 is that we generally agree there have been moral catastrophes in the past, such as widespread slavery. Are there ongoing moral catastrophes? I think factory farming is a pretty obvious one. There's a philosophy paper called "The Possibility of an Ongoing Moral Catastrophe" that gives more context.
MattJ128

I don’t think the response to Covid should give us reason to be optimistic about our effectiveness at dealing with the threat from AI. Quite the opposite. Much of the measures taken were known to be useless from the start, like masks, while others were ineffective or harmful like shutting down schools or giving vaccine to young people who were not at risk of dying from Covid.

Everything can be explained by the incentives our politicians have to do anything. They want to be seen to take important questions seriously while not upsetting their doners in the pharma industry.

I can easily imagine something similar happening if the voters becomes concerned about AI. Some ineffective legislation dictated by Big tech.

Max_He-Ho1510

I think most comments regarding the covid analogy miss the point made in the post. Leopold makes the case that there will be a societal moment of realization and not that specific measures regarding covid were good and this should give us hope. 

Right now talking about AI risk is like yelling about covid in Feb 2020.

I agree with this & there likely being a wake-up moment. This seems important to realize!

I think unless one has both an extremely fast takeoff model and doesn’t expect many more misaligned AI models with increases in capabilities to be ... (read more)

8AnthonyC
This. IIRC by ~April 2020 there were some researchers and experts asking why none of the prevention measures were focused on improving ventilation in public spaces. By ~June the same was happening for the pretty clear evidence that covid was airborne and not transmitted by surfaces (parks near me were closing off their outdoor picnic tables as a covid measure!).  And of course, we can talk about "warp speed" vaccine development all we like, but if we had had better public policy over the last 30 years, Moderna would likely have been already focusing on infectious disease instead of cancer, and had multiple well-respected, well-tested, well-trusted mRNA vaccines on the market, so that the needed regulatory and physical infrastructures could have already been truly ready to go in January when they designed their covid vaccine. We haven't learned these lessons even after the fact. We haven't improved our institutions for next time. We haven't educated the public or our leaders. We seem to have decided to pretend covid was close to a worst case scenario for a pandemic, instead of realizing that there can be much more deadly diseases, and much more rapidly spread diseases. AI seems...about the same to me in how the public is reacting so far? Lots of concerns about job losses or naughty words, so that's what companies and legislators are incentivized to (be seen as trying to) fix, and most people either treat very bad outcomes as too outlandish to discuss, or treat not-so-very-small probabilities as too unlikely to worry about regardless of how bad they'd be if they happened.
MattJ43

It is somewhat alarming that many participants here appear to accept the notion that we should cede political decision-making to an AGI. I had assumed that this was a widely-held view that such a course of action was to be avoided, yet it appears that I may be in the minority.

1twkaiser
The question I'm currently pondering is do we have any other choice? As far as I see, we have three options to deal with AGI risks: A: Ensure that no AGI is ever built. How far are we willing to go to achieve this outcome? Can anything short of burning all GPUs accomplish this? Is that even enough or do we need to burn all CPUs in addition to that and go back to a pre-digital age? Regulation on AI research can help us gain some valuable time, but not everyone adheres to regulation, so eventually somebody will build an AGI anyway. B: Ensure that there is no AI apocalypse, even if a misaligned AGI is built. Is that even possible? C: Ensure that every AGI created is aligned. Can we somehow ensure that there is no accident with misaligned AGIs? What about bad actors that build a misaligned AGI on purpose? D: What I describe in this post - actively build one aligned AGI that controls all online devices and eradicates all other AGIs. For that purpose, the aligned AGI would need to at least control 51% of the world’s total computing power. While that doesn’t necessarily mean total control, we’d already give away a lot of autonomy by just doing that. And surely, some human decision-makers will turn their duties over to the AGI. Eventually, all or most decision-making will be either AGI-guided or fully automated, since it’s more efficient. Am I overlooking something?
1Noosphere89
For me, it's primarily due to way more optimistic models of AI risk than most LWers have, so I count as one of the people who would accept political decision making to an AGI.
MattJ2-1

Someone used the metaphore of Plato’s cave to describe LLMs. The LLM is sitting in cave 2, unable to see the shadows on the wall but can only hear the voices of the people in cave 1 talking about the shadows.

The problem is that we people in cave 1 are not only talking about the shadows but also telling fictional stories, and it is very difficult for someone in cave 2 to know the difference between fiction and reality.

If we want to give a future AGI the responsibility to make important decisions I think it is necessary that it occupies a space in cave 1 and not just being a statistical word predictor in cave 2. They must be more like us.

2Paul Tiplady
I buy this. I think a solid sense of self might be the key missing ingredient (though it’s potentially a path away from Oracles toward Agents). A strong sense of self would require life experience, which implies memory. Probably also the ability to ruminate and generate counterfactuals. And of course, as you say, the memories and “growing up” would need to be about experiences of the real world, or at least recordings of such experiences, or of a “real-world-like simulation”. I picture an agent growing in complexity and compute over time, while retaining a memory of its earlier stages. Perhaps this is a different learning paradigm from gradient descent, relegating it to science fiction for now.
MattJ37

Running simulations of other people’s preferences is what is usually called ”empathy”so I will use that word here.

To have empathy for someone, or an intuition about what they feel is a motivational force to do good in most humans, but it can also be used to be better at deceving and take advantage of others. Perhaps high functioning psychopaths work in this way.

To build an AI that knows what we think and feel, but without having moral motivation would just lead to a world of superintelligent psychopaths.

P.s. I see now that kibber is making the exact same point.

MattJ10

Consciousness is a red herring. We don’t even know if human beings are conscious. You may have a strong belief that you are yourself a conscious being, but how can you know if other people are conscious? Do you have a way to test if other people are conscious?

A superintelligent, misaligned AI poses an existential risk to humanity quite independantly of whether it is conscious or not. Consciousness is an interesting philosophical topic, but has no relevance to anything in the real world.

1[anonymous]
I'm not sure how we could say that there's no phenomenon that the word "consciousness" refers to, it seems to me that it's like questioning if reality itself exists: the point of "reality" is referred to the consistency of things we perceive, if we question if reality 'exists', we still find that consistency of things we perceive regardless, it seems to me that it's analogous to questioning consciousness. If I can identify the referent of the word "consciousness" at all, then I can see if the way other people speak about their experiences matches with that concept of "consciousness", and they do. That's evidence in favour of then being conscious.  And we can actually detect empirical differences between consciousness and non-consciousness, because there are people that perceive visual stimuli who say that are not aware of seeing anything (even while they could at some point of their lifes).
1Roman Leventov
You are talking about what I would call a phenomenological, or "philosophical-in-the-hard-problem-sense" consciousness ("phenomenological" is also not quite right the word because psychology is also phenomenology, relative to neuroscience, but this is an aside). "Psychological" consciousness (specifically, two kinds of it: affective/basal/core consciousness, and access consciousness) is not mysterious at all. These are just normal objects in neuropsychology. Corresponding objects could also be found in AIs, and called "interpretable AI consciousness". "Psychological" and "interpretable" consciousness could be (maybe) generalised in some sort of "general consciousness in systems". (Actually, Fields et al. already proposed such a theory, but their conception of general consciousness surely couldn't serve as a basis of ethics.) The proper theory of non-anthropocentric ethics, shall it be based in some way on consciousness (which I'm actually doubtful about; I will write a post about this soon), surely should use "psychological" and "interpretable" rather than "philosophical-in-the-hard-problem-sense" consciousness.