I feel like I have to wholeheartedly agree with this section below, because I agree with the claim that international cooperation is way better than racing for even very basic forms of cooperation that don't involve slowdowns or pauses, and in general I am alarmed at a lot of the nationalist rhetoric that says we should race to deal with threat of China, because as I see it, US companies are the ones pushing the state of the art, and there's some appetite for China to cooperate on AI, and while I think the evidence for China seriously pausing/slowing down is overstated right now, I think that even without slowing down or pausing that international cooperation is better than international competition on AI safety and progress:
This is the full text of a post from "The Obsolete Newsletter," a Substack that I write about the intersection of capitalism, geopolitics, and artificial intelligence. I’m a freelance journalist and the author of a forthcoming book called Obsolete: Power, Profit, and the Race for Machine Superintelligence.
OpenAI is disbanding its AGI readiness team — and its Senior Advisor for AGI readiness, Miles Brundage, just resigned after over six years at the organization.
The company defines artificial general intelligence (AGI) as “a highly autonomous system that outperforms humans at most economically valuable work.” And Brundage defines “ready for AGI” as roughly “readiness to safely, securely, and beneficially develop, deploy, and govern increasingly capable AI systems.”
So is OpenAI prepared for the thing it’s explicitly trying to build?
No, according to Brundage.
In an accompanying Substack post, he writes, “In short, neither OpenAI nor any other frontier lab is ready, and the world is also not ready.”
His jarring answer is followed immediately by a less than reassuring clarification:
Buried in Brundage’s announcement tweet is the news that the AGI readiness team is being disbanded and absorbed into other teams.
So what?
Brundage was one of the last of an old guard of AI safety and governance employees of the AI company. His departure is yet another sign that OpenAI, as we knew it, is dead.
“Miles is leaving OpenAI = end of an era,” tweeted Haydn Belfield, a Cambridge existential risk researcher and long-time collaborator of Brundage, whose departure announcement prompted a deluge of well-wishes from AI figures who usually agree on very little.
At a glance, there doesn’t appear to be much bad blood between Brundage and OpenAI. He writes:
However, it’s hard not to think that the timing of his exit might have something to do with the recent reports that OpenAI is trying to restructure as a for-profit public benefit corporation (PBC) as a condition of keeping the $6.6 billion it just raised in its latest funding round.
For instance, here's what Brundage had to say when OpenAI first spun up its for profit arm in 2019:
Over five years later, things look very different.
The first major test of the nonprofit board’s power came and went with November’s short-lived firing of CEO Sam Altman.
(Brundage was one of the earliest signatures calling for Altman’s reinstatement, however, over 700 of OpenAI’s 770 employees signed that letter, and I’ve spoken to at least one signatory who changed their mind after more information came to light.)
In response to Altman’s abrupt firing, employees rallied behind the slogan: “OpenAI is nothing without its people.” But nearly a year later, the people that OpenAI is purportedly “nothing without” are mostly different. The company has since grown its headcount by nearly 1,000 employees, to 1,700.
In that same period, OpenAI has lost the majority of its most safety conscious employees. Fortune reported in August that more than half of the employees focused on AGI safety had left the company in just the previous few months.
Additionally, a close read of Brundage’s Substack post hints at some larger frustrations. For example, when explaining why he left, he writes:
In other words, OpenAI was policing his publications too harshly, which shouldn’t come as a surprise when you remember that Altman reportedly tried to fire Helen Toner from the nonprofit board over an obscure academic paper she co-authored that he felt was too critical of OpenAI’s commitment to safety.
Brundage also writes that he’s done much of what he’s set out to do at OpenAI and that he’s “already told executives and the board (the audience of my advice) a fair amount about what I think OpenAI needs to do and what the gaps are.”
Sounds nice enough, right? But as @renormalized tweeted:
Brundage is also careful not to single out OpenAI when he complains about corner-cutting on safety — there are plenty of examples of this from rival AI companies — but he also didn’t say that OpenAI was better on this front either.
There are some years where decades happen
Some think automated AI scientists will make decades of discoveries in just years. OpenAI appears to be trying to demonstrate this principle by compressing decades of corporate scandals into a single year.
In May, the company disbanded its Superalignment team, which was tasked with determining how to make superintelligent systems safe in under four years. The two team leads, Ilya Sutskever and Jan Leike, left the company. On his way out the door, Leike publicly criticized OpenAI, tweeting:
Fortune subsequently reported that the Superalignment team never got the 20 percent of compute it was promised.
The day after Leike departed, OpenAI was roiled by the revelation that it required outgoing employees to sign draconian, lifetime non-disparagement and non-disclosure agreements to retain their vested equity, often the vast majority of their total compensation. The practice had been in place since at least 2019 and was only revealed because of Daniel Kokotajlo’s courageous refusal to sign the agreements and his subsequent choice to share details with Kelsey Piper of Vox.
And in a much less reported story, OpenAI removed Aleksander Madry from a senior safety role in July. His safety responsibilities were given to a different team and effectively deprioritized.
Brundage’s farewell
Brundage’s farewell post is long and worth reading in full, but I’ve pulled out some of the most significant bits here.
As previously mentioned, he mainly says that going independent will allow him more freedom and time to work on “issues that cut across the whole AI industry.” He plans to start or join an existing nonprofit focused on AI policy research and advocacy, “since I think AI is unlikely to be as safe and beneficial as possible without a concerted effort to make it so.”
Brundage is careful to emphasize that “OpenAI remains an exciting place for many kinds of work to happen, and I’m excited to see the team continue to ramp up investment in safety culture and processes.”
It’s completely possible that Brundage genuinely has few gripes with OpenAI and remains optimistic about its attitude toward AI safety research.
But here’s a different hypothesis: If you worked at OpenAI on safety and governance and found that you were not getting the resources and prioritization you thought were appropriate, there’s value in telling the rest of the world, as doing so increases pressure on the organization to change its practices and may make regulation more likely. However, this will also predictably dissuade other safety-minded people from joining going forward. If the culture and leadership were immovable on safety policies, then scaring committed people off won’t matter too much — their presence wouldn’t have made a difference.
But if the culture and leadership aren’t immovable, then making it more likely that all the really serious safety people avoid OpenAI going forward could be bad! Especially since there aren’t many regulations on frontier AI systems and that may not change any time soon. If you think AGI could come in a matter of years, as many at OpenAI do, the people in the room at the first place to build it could have a massively outsized impact on how well it goes.
There is some indirect evidence that this was a consideration for Brundage, who writes, “Culture is important at any organization, but it is particularly important in the context of frontier AI since much of the decision-making is not determined by regulation, but rather up to the people at the company.”
He also calls on Congress to:
Brundage expresses concern about “misalignment between private and societal interests, which regulation can help reduce,” and corner-cutting that can happen without deliberate efforts.
Despite a long record of rhetorically supporting regulation, OpenAI formally opposed the first major binding AI safety bill in the US, California’s SB 1047, which was vetoed by governor Gavin Newsom on September 29th. You can check out my extensive coverage of the epic fight over the bill here.
A call for cooperation with China
Perhaps the most significant part of Brundage’s post is his dissent from the growing chorus of powerful voices (including Sam Altman) pushing for an AI arms race against China:
And even though he thinks the West will continue to outcompete China on AI, Brundage writes that autocratic countries have:
I think the next major fault line in the AI debates will be posture toward China. Altman, Anthropic CEO Dario Amodei, and former OpenAI researcher Leopold Aschenbrenner have prominently advanced varieties of the idea that democracies must ‘prevail’ over China in a race to build superhuman AI systems.
Others, like Brundage, prefer a cooperative approach to China.
Existential clash of civilizations thinking motivated the most dangerous choices ever made, from the nuclear arms buildup to the Soviets’ illegal, secret bioweapons program. The mere belief in this narrative with respect to AGI could motivate profoundly destabilizing actions, like a nuclear first-strike.
We can and should instead approach the challenge of safely building AGI as one of global, rather than national security.
I’ll have a lot more to say on this topic soon.
For now, I’ll just say that we should be very wary of letting science guys with toy models of the world dictate international relations. And we should think harder about how China is likely to react if the “free world” moves mountains to kneecap their AI progress.
I’m also looking forward to seeing what Brundage writes next, unencumbered by corporate overseers.
If you enjoyed this post, please subscribe to The Obsolete Newsletter. You can also find my accompanying Twitter thread here.
Appendix:
Where did the co-authors go?
I checked out four of the papers Brundage co-authored in 2019 and 2020. Of the 29 authors at OpenAI at the time of publication, only five were still there (there may be some overlap).
Five years is an eternity in startup world, so this doesn’t mean a ton on its own. However, the ways in which these folks left are, shall we say, atypical.
For example, this February 2019 paper on the policy implications of GPT-2 names these authors: Alec Radford, Jeffrey Wu, Dario Amodei, Daniella Amodei, Jack Clark, Miles Brundage, and Ilya Sutskever.
Dario and Daniela Amodei started Anthropic with Jack Clark and other OpenAI defectors after reportedly trying to push Altman out over safety concerns in 2021 (Anthropic officially disputes this account).
Sutskever voted to fire Altman in November 2023, caved under immense pressure and reversed his decision. He never returned to the office and formally left OpenAI in May to launch Safe Super Intelligence.
Alec Radford and Jeffrey Wu were the lead authors on the GPT-2 technical paper.
Wu left in July and said this to Vox in September following reports that OpenAI was planning to transform from a nonprofit to a for-profit:
Radford is the only author from both GPT-2 papers still at OpenAI.
Full announcement text
I asked Claude to transcribe Brundage’s full announcement message from the screenshots he posted: