Yes, it's a great topic. The aspect which seems to be missing from "AI capabilities can be significantly improved without expensive retraining", https://arxiv.org/abs/2312.07413 is that post-training is a particularly fertile ground for rapid turnaround self-modification and recursive self-improvement, as post-training tends to be rather lightweight and usually does not include a delay of training a novel large model.
Some recent capability works in that direction include, for example
"Self-Taught Optimizer (STOP): Recursively Self-Improving Code Generation", https://arxiv.org/abs/2310.02304
"Language Agents as Optimizable Graphs", https://arxiv.org/abs/2402.16823
People who are specifically concerned with rapid foom risks might want to focus on this aspect of the situation. These self-improvement methods currently saturate in a reasonably safe zone, but they are getting stronger both due to novel research, and due to improvements of the underlying LLMs they tend to rely upon.
An important and neglected topic!
Also, a challengingly complicated topic. Been thinking a lot about this myself recently, in looking at possible interactions between general models and use-case-specific software or models. For instance, in biology, if an LLM agent can query an API for a tool like AlphaFold, or search records produced by such a tool, and then add the results of the query to their context before answering a user's question... The result is a much more powerful and easy-to-use system than either LLM or narrow tool alone.
This work has been done in the context of SaferAI’s work on risk assessment. Equal contribution by Eli and Joel. I'm sharing this writeup in the form of a Google Doc and reproducing the summary below.
Disclaimer: this writeup is context for upcoming experiments, not complete work. As such it contains a lot of (not always well-justified) guess-work and untidy conceptual choices. We are publishing now despite this to get feedback.
If you are interested in this work — perhaps as a future collaborator or funder, or because this work could provide helpful input into e.g. risk assessments or RSPs — please get in touch with us at joel@qallys.com and/or simeon@safer-ai.org.
Summary
In this write-up, we will: