I think if the problem turns out to be too difficult to solve for humanity right now, the right strategy seems pretty straightforward:
My current model is that there are roughly two promising ways to make people smarter:
Both of these seem pretty promising and I am in favor of work on both of these.
Re: 2
Most promising way is just raising children better.
See (which I'm sure you've already read): https://www.lesswrong.com/posts/CYN7swrefEss4e3Qe/childhoods-of-exceptional-people
Alongside that though, I think the next biggest leverage point would be something like nationalising social media and retargeting development/design toward connection and flourishing (as opposed to engagement and profit).
This is one area where, if we didn't have multiple catastrophic time pressures, I'd be pretty optimistic about the future. These are incredibly high impact and t...
Curious about the 'delay the development' via regulation bit.
What is your sense of what near-term passable regulations would be that are actually enforceable? It's been difficult for large stakeholder groups facing threatening situations to even enforce established international treaties, such as the Geneva convention or the Berne three-step test.
Here are dimensions I've been thinking need to be constrained over time:
If it requires big datacenters then I think folks will hear about it and stop it. We're not the only country with a CIA. A datacenter can be destroyed without even killing any people (so less risk of retaliation). Let's hope it requires a big obvious datacenter.
Meanwhile, people were on track to invent new technology at increasing speed every year without the AI's help. Personally, I don't mind it taking 10x longer to reach the stars etc.
Very good question. It is awful that we find ourselves in a situation in which there are only tiny shreds of hope for our species's surviving "the AI program" (the community pushing AI capabilities as far as they will go).
One tiny shred of hope is some "revolution in political affairs" that allow a small group to take over the world, and this small group understands how dangerous the AI program is. One way such a revolution might come about is the creation of sufficiently good technology to "measure loyalty" by scanning the brain somehow: the first group to use the brain-scanning tech to take over the world (by giving police, military and political power only to loyal people) can (hopefully) prevent other groups from using the brain-scanning tech, yielding a stable regime, and hopefully they use their stable hegemony to shut down the AI labs, to make it illegal to teach or publish about AI and to stop "progress" in GPU tech.
The greater the number of "centers of autonomous power" on Earth, the greater the AI extinction risk because the survival of our species basically requires every center of autonomous power to choose to refrain from continuing the AI program and because when there is only one dominant center of power, that center's motive to use AI to gain an advantage over rival centers is greatly diminished relative to the current situation. Parenthetically, this is why I consider the United States the greatest source of AI extinction risk: Russia and China are arranged so as to make it easy or at least possible for a few people in the central government to shut down things going on in the country whereas the United States is arranged to keep power as dispersed as practical a la, "the government that governs least governs best".
Another tiny shred of hope is our making contact with an alien civilization and asking them to help us out of the situation. This need not entail the arrival of an alien ship or probe because a message from the aliens can contain a computer program and that computer program might be (and AFAICT probably would be) an AGI which (after we receive it) we can run or emulate on our computers. Yes, doing that definitely does give an alien civilization we know almost nothing about complete power over us, but the situation around human-created AI is so dire that I'm tentatively tepidly in favor of putting our hope in the possibility that the alien civilization that sent the message will turn out to be nice or if it is not intrinsically nice at least has some extrinsic motive to treat us well. (One possible extrinsic motive would be protecting its reputation among civilizations that might receive the message that aren't as helpless as we are and consequently might be worthwhile to trade with. I can expand on this if there is interest.)
IIUC the Vera Rubin telescope coming online in a few years will increase our civilization's capacity to search for messages in the form laser beams by at least 3 orders of magnitude.
I can imagine other tiny shreds of hope, but they have the property that if the people (i.e., most people with power or influence) who don't understand the severity of AI extinction risk knew about them, they'd probably try to block them, and blocking them would probably be fairly easy for them to do, so it doesn't make any sense to talk about them on a public forum.
Unlike most of the technical alignment/safety research going on these days, pursuing these tiny shreds of hope at least doesn't make the problem worse by providing (unintended) assistance to the AI project.
Is this an accurate summary of your suggestions?
Realistic actions an AI Safety researcher can take to save the world:
Given we’re not on track to control fully autonomous systems to stay safe, what do we do?
Let’s say alignment research will not catch up with capability development. For any combination of reasons: corporations are scaling too fast, there are too many lethal-if-unsolved subproblems we have at most made partial process on, some subproblems can only be solved in sequential order, or there are hard limits capping the progress we can make on the control problem.
What do we do? With ‘we’, you can consider yourself, specific organisations, or the community coordinating roughly as a whole.