Roko would probably call "the most important century" work "building a stable equilibrium to land an AGI/ASI on".
I broadly agree with you and Roko that this work is important and that it would often make more sense for people to do this kind of work than "narrowly-defined" technical AI safety.
An aspect for why this may be the case that you didn't mention is money: technical AI safety is probably bottlenecked on funding, but more of the "most important century/stable equilibrium" are more amenable to conventional VC funding, and the funders shouldn't even be EA/AI x-risk/"most important century"-pilled.
In a comment to Roko's post, I offered my classification of this "stable equilibrium" systems and work that should be done. Here I will reproduce it, with extra directions that appeared to me later:
The list above covers all the directions mentioned in the post, and there are a few more important ones.
I originally wrote this for the Meta Coordination Forum. The organizers were interested in a memo on topics other than alignment that might be increasingly important as AI capabilities rapidly grow — in order to inform the degree to which community-building resources should go towards AI safety community building vs. broader capacity building. This is a lightly edited version of my memo on that. All views are my own.
Some example neglected topics (without much elaboration)
Here are a few example topics that could matter a lot if we’re in the most important century, which aren’t always captured in a normal “AI alignment” narrative:
(More elaboration on these below.)
Here are a few examples of somewhat-more-concrete things that it might (or might not) be good for some people to do on these (and related) topics:
Implications for community building?
…with a focus on “the extent to which community-building resources should go towards AI safety vs. broader capacity building”.
Elaborating on the example topics
Elaborating on the topics I mentioned above.
Nick Bostrom and Carl Shulman’s propositions concerning digital minds and society has some good discussion of a lot of this stuff.
How ITN are these issues?
How good do these topics look in an importance/neglectedness/tractability framework? In my view, they look comparable to alignment on importance, stronger on neglectedness (if we consider only work that’s been done so far), and pretty unclear on tractability (though probably less tractable than alignment).
For example, let’s consider “human deliberation could go poorly (without misalignment or other blatant x-risks”).
So let’s briefly talk about a specific argument for why these neglected topics might not be so great: That if we solve alignment, AI will help us deal with these problems. Or phrased differently: Why spend precious hours on these problems now, when cognitive resources will be cheap and plentiful soon enough.[3]
I think this argument is pretty good. But I don’t think it’s overwhelmingly strong:
So I don't think “AI will help us deal with these problems” is decisive. I’d like to see more attempted investigations to learn about these issues’ tractability.
P.S. Feel free to DM me if you’re interested in working on any of these topics.
Including both: Welfare of digital minds, and whether there’s any types of misaligned AI that would be relatively better to get, if we fail to get intent-alignment.
This could be: tech that (if proliferated) would give vast destructive power to millions of people, or that would allow & encourage safe “first strikes” against other countries, or that would allow the initial developers of that tech to acquire vast power over the rest of the world. (C.f.: vulnerable world hypothesis.)
Annoyingly — that could be counted into either lower importance (if we restrict our attention to the part of the problem that needs to be addressed before sufficiently good AI), lower neglectedness (if we take into account all of the future labor that will predictably be added to the problem), or lower tractability (it’s hard to make an impact by doing research on questions that will mainly be determined by research that happens later-on).