Here is another example of an outsider perspective on risks from AI. I think such examples can serve as a way to fathom the inferential distance between the SIAI and its target audience as to consequently fine tune their material and general approach.
This shows again that people are generally aware of potential risks but either do not take them seriously or don't see why risks from AI are the rule rather than an exception. So rather than making people aware that there are risks you have to tell them what are the risks.
If we make an AI like that in the first place, then we've pretty much already lost. If it ever had to destroy us to protect its ability to carry out its top-level goals, then its top-level goals are clearly something other than humaneness, in which case it'll most likely wipe us out when it's powerful enough whether or not we ever felt threatened by it.