That’s useful analysis. Focusing so heavily on evals seems like a mistake given how AI Safety Institutes are focused on evals.
(I work at Open Phil on TAIS grantmaking)
I agree with most of this. A lot of our TAIS grantmaking over the last year was to evals grants solicited through this RFP. But I want to make a few points of clarification:
I appreciate you examining our work and giving your takes!
Thanks for this analysis! A minor note: you're probably aware of this, but OpenPhil funds a lot of technical AI safety field-building work as part of their "Global Catastrophic Risks Capacity Building" grants. So the proportion of field-building / talent-development grants would be significantly higher if those were included.
Thanks for this post!
This looks much worse than I thought it would, both in terms of funding underdeployment, and in terms of overfocusing on evals.
TL;DR
I spent a few hours going through Open Philanthropy (OP)'s grant database. The main findings were:
My overall takeaway was that very few TAIS grants are directly focused on making sure systems are aligned / controllable / built safely.[1]
Method
I:
Results
Grants by Research Agenda
Grants by Cluster
Full data available here
Key Findings
(1) Evaluations & Benchmarking make up 2/3rds of all OP TAIS funding in 2024
Most of these grants are related to the RFP on LLM Benchmarks.
(2) Excluding Evaluations & Benchmarking, OP grants for TAIS have fallen significantly
(3) Most TAIS funding is focused on "investment" rather than direct approaches to AI safety
I classify grants into two broad buckets:
More cynically: I worry the heavy focus on evaluations will give us a great understanding of how / when scary AI systems emerge. But that won't (a) prevent the scary AI systems being built, or (b) cause any action if they are built (see also Would catching your AIs trying to escape convince AI developers to slow down or undeploy?)