I would like to be given pointers to prior art on the following question: how do we communicate about potential risks from AI in a net-positive way, taking into account the negative that is giving people ideas? I can easily see how someone describing steps an AI takeover could happen influences someone who is malicious or just cares less about safety. This directly relates to the research I intend to pursue, relating specific capabilities AI systems do have or can acquire and the risk factors associated with them.
No answer for you yet but I'm trying to achieve something similar in my book. I want to avoid infohazards but communicate the threat, and also provide hope. A tricky thing to navigate for sure, especially with diverse audiences.