Slowing down AI progress is an underexplored alignment strategy
The emotional burden of watching the world end My current beliefs about AI timelines have made my life significantly worse. I find thoughts about the ever-shrinking timeline to AGI invading nearly every aspect of my life. Every choice now seems to be affected by the trajectory of this horrible technology....
I wrote this a year ago and was not really thinking about this topic as carefully and was feeling quite emotional about the lack of effort. At the time people mostly thought slowing down AI was impossible or undesirable for some good reasons and a lot of reasons that in hindsight looked pretty dumb.
I think a better strategy would look more like “require new systems guarantee a reasonable level of interpretability and pass a set of safety benchmarks”
And eventually, if you can actually convince enough people of the danger, there should be a hard cap on the amount of compute that can be used in training runs that decreases over time to compensate for algorithmic improvements.