In this post I propose an (allegedly...) secure AI boxing method. For reasons elaborated at the end of the post, this proposal is probably unsafe, and is certainly not practical. Still, I've elected to share my thoughts on this in the hope of receiving valuable feedback, as well a hope that discussing these ideas in public will provoke more useful thoughts in the minds of others.
Thank you for reading.
-- Xavier O'Rourke
Yudkowsky, Bostrom and others have illustrated the fundamental security problems which arise when attempting to mitigate the influence of a smarter-than-human intelligence by putting the agent inside a “box”.
“Boxing” an artificial intelligence means placing the agent in a virtual... (read 1626 more words →)
What if people are required to try a tiny bit of each before deciding which one they want a full helping of? In that setup I'm still extremely confident the ice cream wins - it would be crazy imo for anyone to update meaningfully due to this survey.
Of course, as we dive more into the nitty gritty of the setup it becomes relevant how the product/stunt is positioned etc. - but the fact that there are all these considerations that need to be made serves to further reinforce the point that it's hard to predict human preferences!