This definition of gain of function is about gaining very specific functions of pathogenesis, transmissibility, or host range. It does not include a function like a bacteria getting the capability to produce insulin.
When it comes to AI research, the corresponding approach would be to think about which function we consider to be dangerous to develop. You likely don't want DeepMind's Gato to have finding zero-day exploits as one of his training tasks.
It might be worth thinking about which tasks for Gato might be analogous to the functions we mean when we speak about gain-of-function research in biology.
All technological development is gain-of-function research on matter. What matters is how dangerous the gained function is.
Epistemic status: highly uncertain and fairly confused, seeking out opinions
First let me just post the definition of (biological) gain-of-function research from Wikipedia, so that everyone is on the same page:
The question I have is, should we have similar intuitions about certain kinds of AI and/or AI safety research? This is related in a complicated way to AI capabilities research, so let's define some terms carefully:
This is more of a spectrum than it is two distinct classes; for instance, which class "recursive self-improvement" belongs to probably depends on how seriously you take AI risk in the first place. And one could imagine a sufficiently edgy or depressed person wanting to program AI's to do all manner of nonsense.
There are then a number of questions:
Why does this matter? Suppose that we identify some capability (or group of capabilities) that would be necessary and/or sufficient for a model to disempower humans. (One might call such a thing an "ascension kit", after the strategic term in NetHack lore.) How would we detect if a system has or is trying to assemble an ascension kit? Implementing parts of an ascension kit seems like it's obviously gain-of-function research, but how else would we be able to test any such detection toolkit?
Lastly, if there are related questions about this topic, I would love it if people posted them in the comments.