If you found yourself interested in advocacy, the largest AI Safety protest ever is happening Saturday, October 21st!
https://www.lesswrong.com/posts/abBtKF857Ejsgg9ab/tomorrow-the-largest-ai-safety-protest-ever
I think there’s an aesthetic clash here somewhere. I have an intuition or like... an aesthetic impulse, telling me basically… “advocacy is dumb”. Whenever I see anybody Doing An Activism, they're usually… saying a bunch of... obviously false things? They're holding a sign with a slogan that's too simple to possibly be the truth, and yelling this obviously oversimplified thing as loudly as they possibly can? It feels like the archetype of overconfidence.
This is exactly the same thing that I have felt in the past. Extremely well said. It is worth pointing out explicitly that this is not a rational thought - it's an Ugh Field around advocacy, and even if the thought is true, that doesn't mean all advocacy has to be this way.
Sometimes such feelings are your system 1 tracking real/important things that your system 2 hasn’t figured out yet.
I just want to say that I thought this was an excellent dialogue.
It is very rare for two people with different views/perspectives to come together and genuinely just try to understand each other, ask thoughtful questions, and track their updates. This dialogue felt like a visceral, emotional reminder that this kind of thing is actually still possible. Even on a topic as "hot" as AI pause advocacy.
Thank you to Holly, Rob, and Jacob.
I'll also note that I've been proud of Holly for her activism. I remember speaking with her a bit when she was just getting involved. I was like: "she sure does have spirit and persistence– but I wonder if she'll really be able to make this work." But so far, I think she has. I'm impressed with how far she's come.
I think she's been doing an excellent and thoughtful job so far. And this is despite navigating various tradeoffs, dealing with hostile reactions, being a pioneer in this space, and battling the lonely dissent that Rob mentioned.
I don't know what the future of AI pause advocacy will look like, and I'm not sure what the movement will become, but I'm very glad that Holly has emerged as one of its leaders.
Curated.
I think "whether to engage in advocacy for AI, and how to go about it" is a pretty important topic. I do get the sense of LessWrong-type folk being really selected for finding it aesthetically ughy, and it seems like a lot of this is a bias.
I think there are separately real arguments about the risks and costs of advocacy, or how to figure out how to have an "advocacy arm" of the AI safety movement without warping the epistemic culture we've built here. I'd be interested in a followup dialogue that somehow approaches that with an "okay, how can we make this work?" attitude.
Update: Will curate this in 2-3 days instead. Looks like curation emails currently look a bit broken for dialogues, and we should fix that before we send out an email to 10k+ people.
Fwiw, since we decided to delay a couple days in curating, a thing I think would be cool for this one is to have either a "highlights" section at the beginning, or maybe a somewhat gearsier "takeaways" at the end.
Maybe this is more useful for someone else to do since it may be harder for you guys to know what felt particularly valuable for other people.
Just in case you missed that link at the top:
This is a historic event, the first time hundreds of people are coming out in 8 countries to protest AI.
I'm helping with logistics for the San Francisco one which you can join here. Feel free to contact me or Holly on DM/email for any reason.
I think it could be misinterpreted to mean "pause all AI development and deployment", which results in a delayed deployment of "sponge safe" narrow AI systems that would improve or save a large number of people's lives. There's a real cost to slowing things down.
This cost is trivial compared to the cost of AGI Ruin. It's like going on a plane to see your family on a plane where the engineers say they think there's a >10% chance of catastrophic failure. Seeing your family is cool, but ~nobody would think it's reasonable to go on such a plane. There are other ways to visit your family, they just take longer.
The analogy breaks down when it comes to trying to fix the plane. We understand how airplanes work; we do not understand how AI works. It makes sense to ground the plane until we have such understanding, despite the benefits of transportation.
I would love to have all the cool AI stuff too, but I don't think we're capable of toeing the line between safe and ruinous AI at acceptable risk levels.
I think in this analogy the narrow AI models would be like texting your parents instead of flying to see them. Obviously not as good as visiting them in person, but you avoid the 10% x-risk. I think Rob is saying let's make sure we don't stop the development of texting/calling as collateral.
So yes, don't get on the plane, but let's be very specific about what we're trying to avoid.
There seems to be a trade-off in policy-space between attainability and nuance (part of what this whole dialogue seems to be about). The point I was trying to make here is that the good of narrow AI is such a marginal gain relative to the catastrophic ruin of superintelligent AI that it's not worth being "very specific" at the cost of potentially weaker messaging for such a benefit.
Policy has adversarial pressure on it, so it makes sense to minimize the surface area if the consequence of a breach (e.g. "this is our really cool and big ai that's technically a narrow ai and which just happens to be really smart at lots of things...") is catastrophic.
One thing that me more comfortable with making statements that are less nuanced in some circumstances is Wittgenstein's idea of language games. Rationalists have a tendency of taking words literally, whilst Wittgenstein views statements as moves in a language games where there are a host of different language games for different situations and people can generally figure it out. Specifically, there seems to be some distinct language games associated with protests where people understand that your sign or slogan doesn't cover everything in complete nuance. At the same time, I think we should be trying to raise the bar in terms of the epistemics/openness of our advocacy work and I do see risks in people taking this reasoning too far.
There is a massive tradeoff between nuance/high epistemic integrity and reach. The general population is not going to engage in complex nuanced arguments about this, and prestigious or high-power people who are able to understand the discussion and potentially steer government policy in a meaningful way won't engage in this type of protest for many reasons, so the movement should be ready for dumbing-down or at least simplifying the message in order to increase reach, or risk remaining a niche group (I think "Pause AI" is already a good slogan in that sense).
I agree that there is a trade-off here, however:
a) Dumbing down the message will cost us support from ML engineers and researchers.
b) If the message is dumbed down too much, then the public is unlikely to create pressure towards the kinds of actions that will actually help as opposed to pressuring politicians to engage in shallow, signaling-driven responses.
I think the idea we're going to be able to precisely steer government policy to achieve nuanced outcomes is dead on arrival - we've been failing at that forever. What's in our favor this time is that there are many more ways to cripple advance than to accelerate it, so it may be enough for the push to be simply directionally right for things to slow down (with a lot of collateral damage).
Our inner game policy efforts are already bearing fruit. We can't precisely define exactly what will happen, but we certainly can push for nuance via this route than we would be able to through the public outreach route.
I can see why you would be a lot more positive on advocacy if you thought that crippling advances is a way out of our current crisis. Unfortunately, I fear that will just result in AI being built by whichever country/actor cares the least about safety. So I think we need more nuance than this.
We learned the hard way that "solving the technical problem is way easier than solving the political one". But group project technical problems are artificially easy.
I think we've mostly learned this not from group projects, but rather from most of history. E.g. just look at "covid as technical problem" vs "covid as political problem".
Covid was a big learning experience for me, but I'd like to think about more than one example. Covid is interesting because, compared to my examples of birth control and animal-free meat, it seems like with covid humanity smashed the technical problem out of the park, but still overall failed by my lights because of the political situation.
How likely does it seem that we could get full marks on solving alignment but still fail due to politics? I tend to think of building a properly aligned AGI as a straightforward win condition, but that's not a very deeply considered view. I guess we could solve it on a whiteboard somewhere but for political reasons it doesn't get implemented in time?
I think this is a potential scenario, and if we remove existential risk from the equation, it is somewhat probable as a scenario, where we basically have solved alignment, and yet AI governance craps out in different ways.
I think this way primarily because I tend to think that value alignment is really easy, much easier than LWers generally think, and I think this because most of the complexity of value learning is offloadable to the general learning process, with only very weak priors being required.
Putting it another way, I basically disagree with the implicit premise on LW that being capable of learning is easier than being aligned to values, at most they're comparably or a little more difficult.
More generally, I think it's way easier to be aligned with say, not killing humans, than to actually have non-trivial capabilities, at least for a given level of compute, especially at the lower end of compute.
In essence, I believe there's simple tricks to aligning AIs, while I see no reason to expect a simple trick to make governments be competent at regulating AI.
it's one toe out of line and they can never sincerely value and seek truth again.
That's almost my position, but not quite: mistakes, born of passion (anger, fear, etc.) aren't fatal. But deliberately deciding to sacrifice truth to achieve something … yeah, there's no coming back from that.
Does that really seem true to you? Do you have no memories of sacrificing truth for something else you wanted when you were a child, say? I'm not saying it's just fine to sacrifice truth but it seems false to me to say that people never return to seeking the truth after deceiving themselves, much less after trying on different communication styles or norms. If that were true I feel like no one could ever be rational at all.
I think that's a misunderstanding of what I mean by "sacrificing truth." Of course I have lied: I told my mom I didn't steal from the cookie jar. I have clicked checkboxes saying "I am over 18" when I wasn't. I enjoy a game of Mafia as much as the next guy. Contra Kant, I wholeheartedly endorse lying to your enemies to protect your friends.
No, sacrificing truth is fundamentally an act of self-deception. It is making yourself a man who believes a falsehood, or has a disregard for the truth. It is Gandhi taking the murder-pill. That is what I consider irreversible. It's not so easy that I worry I might do it to myself by accident, so I'm not paranoid about it or anything.
(One way to go about doing this would be to manipulate your language, redefining words as convenient: "The sky is 'green.' My definition of the word 'green' includes that color. It has always included that color. Quid est veritas?" Doing such things for a while until it becomes habitual should do it.)
In this sense, no, I don't think I have ever done this. By the time I conceived of the possibility, I was old enough to resolve never to do it.
Of course, the obvious counter is that if you had scifi/magic brain surgery tech, you could erase and rewrite your mind and memories as you wished, and set it to a state where you still sincerely valued truth, so it's not technically irreversible. My response to that is that a man willing to rewrite his own brain to deceive himself is certainly not one who values truth, and the resultant amnesiac is essentially a different person. But okay, fair enough, if this tech existed, I would reconsider my position on the irreversibility of sacrificing truth via self-deception.
No, sacrificing truth is fundamentally an act of self-deception. It is making yourself a man who believes a falsehood, or has a disregard for the truth. It is Gandhi taking the murder-pill. That is what I consider irreversible.
This is what I was talking about, or the general thing I had in mind, and I think it is reversible. Not a good idea, but I think people who have ever self-deceived or wanted to believe something convenient have come back around to wanting to know the truth. I also think people can be truthseeking in some domains while self-deceiving in others. Perhaps if this weren’t the case, it would be easier to draw lines for acceptable behavior, but I think that unfortunately it isn’t.
Very beside my original point about being willing to speak more plainly, but I think you get that.
The LessWrong Review runs every year to select the posts that have most stood the test of time. This post is not yet eligible for review, but will be at the end of 2024. The top fifty or so posts are featured prominently on the site throughout the year.
Hopefully, the review is better than karma at judging enduring value. If we have accurate prediction markets on the review results, maybe we can have better incentives on LessWrong today. Will this post make the top fifty?
Why no discussion of the world view? IMHO AI cannot be paused because, if the US & EU pause AI development to pursue AI safety research then other state actors such as Russia & China will just see this as an opportunity to get ahead with the added benefit that the rest of the world will freely give away their safety research. It's a political no-brainer unless the leaders of those countries have extreme AI safety concerns themselves. Does anyone really believe that the US would either go to war or impose serious economic sanctions on countries that did not pause?
This is a pretty standard joke/stereotype, the XKCD protest where the signs are giant and full of tiny text specifying exactly what we mean.
Can somebody link to this comic? My attempts to Google different combinations of "XKCD", "protests", and "giant sign" were unsuccessful.
Reflecting on what I found while Googling, I think the phrase “XKCD protest” refers not to a protest portrayed in an XKCD comic but rather to an XKCD comic that was used in a protest: https://www.reddit.com/r/xkcd/comments/6uu5wj/xkcd_1357_in_the_boston_protest_xpost_rpics/
I couldn't find any xkcd even remotely similar. What this does remind me of though is the "left wing memes vs. right wing memes" (meta)meme.
Holly is an independent AI Pause organizer, which includes organizing protests (like this upcoming one). Rob is an AI Safety YouTuber. I (jacobjacob) brought them together for this dialogue, because I've been trying to figure out what I should think of AI safety protests, which seems like a possibly quite important intervention; and Rob and Holly seemed like they'd have thoughtful and perhaps disagreeing perspectives.
Quick clarification: At one point they discuss a particular protest, which is the anti-irreversible proliferation protest at the Meta building in San Francisco on September 29th, 2023 that both Holly and Rob attended.
Also, the dialogue is quite long, and I think it doesn't have to be read in order. You should feel free to skip to the section title that sounds most interesting to you.
(Meta-note: At this point in the dialogue, there was a brief pause, after which Rob and Holly switched to having a spoken conversation, that was then transcribed.)
Techno-optimism about everything... except AGI
Cluelessness and robustly good plans
Taking beliefs seriously, and technological progress as enabler of advocacy
The aesthetics of advocacy, and showing up not because you're best suited, but because no one else will
Closing thoughts