Tenoke comments on xkcd on the AI box experiment - Less Wrong

15 Post author: FiftyTwo 21 November 2014 08:26AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (229)

You are viewing a single comment's thread. Show more comments above.

Comment author: Brillyant 21 November 2014 07:14:07PM 6 points [-]

In his defense, is it possible EY can't win at this point, regardless of his approach? Maybe the internet has grabbed this thing and the PR whirlwinds are going to do with it whatever they like?

I've read apologies from EY where he seems to admit pretty clearly he screwed up. He comes off as defensive and pissy sometimes in my opinion, but he seems sincerely irked about how RW and other outlets have twisted to whole story to discredit LW and himself. From my recall, one comment he made on the reddit sub dedicated to his HP fanfic indicated he was very hurt by the whole kerfuffle, in addition to his obvious frustration.

Comment author: Yvain 22 November 2014 02:39:54AM *  51 points [-]

At this point I think the winning move is rolling with it and selling little plush basilisks as a MIRI fundraiser. It's our involuntary mascot, and we might as well 'reclaim' it in the social justice sense.

Then every time someone brings up "Less Wrong is terrified of the basilisk" we can just be like "Yes! Yes we are! Would you like to buy a plush one?" and everyone will appreciate our ability to laugh at ourselves, and they'll go back to whatever they were doing.

Comment author: Tenoke 22 November 2014 10:41:35AM 9 points [-]

Blasphemy, our mascot is a paperclip.

Comment author: chaosmage 22 November 2014 11:11:24AM 11 points [-]

I'd prefer a paperclip dispenser with something like "Paperclip Maximizer (version 0.1)" written on it.

Comment author: philh 22 November 2014 03:15:38PM 4 points [-]

But a plush paperclip would probably not hold its shape very well, and become a plush basilisk.

Comment author: Tenoke 22 November 2014 05:38:15PM 26 points [-]

Close enough

Comment author: Error 22 November 2014 10:06:08PM 2 points [-]

I feel the need to switch from Nerd Mode to Dork Mode and ask:

Which would win in a fight, a basilisk or a paperclip maximizer?

Comment author: Dallas 22 November 2014 11:21:38PM 0 points [-]

Paperclip maximizer, obviously. Basilisks typically are static entities, and I'm not sure how you would go about making a credible anti-paperclip 'infohazard'.

Comment author: ThisSpaceAvailable 26 November 2014 09:00:20AM 3 points [-]

That depends entirely on what the PM's code is. If it doesn't include input sanitizers, a buffer overflow attack could suffice as a basilisk. If your model of a PM basilisk is "Something that would constitute a logical argument that would harm a PM", then you're operating on a very limited understanding of basilisks.

Comment author: lmm 25 November 2014 10:52:57PM 3 points [-]

The same way as an infohazard for any other intelligence: acausally threaten to destroy lots of paperclips, maybe even uncurl them, maybe even uncurl them while they were still holding a stack of pap-ARRRRGH I'LL DO WHATEVER YOU WANT JUST DON'T HURT THEM PLEASE