Let's suppose we have an AGI running on a perfectly physically isolated computer that may be misalligned. It only communicates with select humans through a perfectly secure text channel.
How can such an AGI escape?
While don't think that AGI can manipulate EVERY person to do what AGI wants, it's better to be cautious.
But recently, I thought of steganography. An AGI can tell a human to convey a message to the public that looks helpful at the first glance, but in fact contains a steganographic message with instructions for building antimatter fusion reactor/nanobots/bioweapon/whatever.
What other ways do you think can such an AGI escape?
Hmm, I was somewhat worried about that, but there are way more dangerous things for AI to see written on the internet.
If you're trying to create AGI by training it on a large internet crawl dataset, you have bigger problems...
To fix something, we need to know what to fix first.