Here is another example of an outsider perspective on risks from AI. I think such examples can serve as a way to fathom the inferential distance between the SIAI and its target audience as to consequently fine tune their material and general approach.
This shows again that people are generally aware of potential risks but either do not take them seriously or don't see why risks from AI are the rule rather than an exception. So rather than making people aware that there are risks you have to tell them what are the risks.
I think these are two important points. In particular, the idea of "mind design space" deserves some serious criticism. (I mentioned one point here.) It assumes a particular representation (one bit representing each possible binary feature of a mind), but reasoning about a different representation might lead to very different conclusions.