If you're making AI for study it shouldn't be super-intelligent at all, ideally it should be dumber than you. I can imagine an AGI that can usefully perform some tasks but is too stupid to self-modify itself into fooming if constrained. You can let it be in charge of opening and closing doors!
Well, I definitely agree that we should make non-super intelligent AIs for study, and also for a great many other reasons. But it's perhaps less clear what 'too stupid to foom' actually means for an AGI. There was a moment when a hominid brain crossed an invisible line and civilization became possible; but the mutation precipitating that change may not have obviously been a major event from the perspective of an outside observer. It may just have looked like another in a sequence of iterative steps. Is the foom line in about the same place as the agric...
If it's worth saying, but not worth its own post (even in Discussion), then it goes here.
Notes for future OT posters:
1. Please add the 'open_thread' tag.
2. Check if there is an active Open Thread before posting a new one.
3. Open Threads should be posted in Discussion, and not Main.
4. Open Threads should start on Monday, and end on Sunday.