[I privately wrote the following quick summary of some publicly-available information on (~safety-relevant) talent leaving OpenAI since the founding of Anthropic. Seems worth pasting here since it already exists but I'd have been more careful if I wrote it with public sharing in mind, it's not comprehensive, and I don't have time to really edit. I'd advise against updating too hard on it because:
The main group of people working on alignment (other than interpretability) at OpenAI at the time of the Anthropic split at the end of 2020 was the Reflection team, which has since been renamed to the Alignment team. Of the 7 members of the team at that time (who are listed on the summarization paper), 4 are still working at OpenAI, and none are working at Anthropic.
Like Habryka, I believe it's literally true that nobody from the "Alignment team" left for Anthropic and 4/7 are still working at OpenAI. But it seems possible that things look different if you weight by seniority and account for potential contributions to OpenAI's attention to existential safety made by people who weren’t technical safety researchers, who were researchers on another team, etc.
Important: I don't know why the below people left OpenAI and their inclusion doesn't mean there's any bad blood between them or that they necessarily have criticisms of OpenAI's attitude toward safety.
If I understand correctly,
1 The alignment team lost its team lead (Paul).
2 Two senior people who weren’t counted as on the team but oversaw it or helped with its research direction left for Anthropic.
3 The head of the interpretability team (Chris Olah), which is one of the other teams that seems most relevant to existential safety, left for Anthropic.
4 Other Anthropic co-founders who left OpenAI include
5 Another person who worked on technical safety at OpenAI and left for Anthropic
6 Several people on the policy team left for Anthropic including the director and two EAs who are interested in alignment.
7 Another EA who I believe cares about alignment and left OpenAI for Anthropic:
8 Other people I don’t know who left for Anthropic
9 Holden is no longer on OpenAI's board (though Helen Toner now is).
On the other hand, they’ve also hired some EAs who care about alignment since then. I believe examples include:
Thanks for writing this. I found it useful and have shared it with others.
I’ve also heard a number of people tell me that EA or AI safety efforts caused them to lose the ability to have serious hobbies, or serious intellectual interests, and I would guess this was harmful to long-term AI safety potential in most cases.
If you'd be up for sharing, I'd be pretty interested in a rough estimate of how many specific people you know of who have had this experience (and maybe also how many were "people who (IMO) have a better shot and a better plan than most for reducing AI risk.")
To be clear, I totally buy that this happens and there's a problem here. I just find it useful to get some sense of the prevalence of this kind of thing.
FYI - the footnotes seem broken. Footnote 6 links to halfway through footnote 4.