Crossposting from my substack. I figured I would crosspost to LessWrong since electoral systems seem like the type of thing rationalists would be interested in. This is the latest post in my constitutional design series. It is the quintessential topic of constitutional design, the one that everyone loves to talk...
Does a paperclip-maximizing AI care about the actual number of paperclips being made, or does it just care about its perception of paperclips? If the latter, I feel like this contradicts some of the AI doom stories: each AI shouldn’t care about what future AIs do (and thus there is...
Crossposted from the EA Forum AI Safety is hot right now. The FLI letter was the catalyst for most of this, but even before that there was the Ezra Klein OpEd piece in the NYTimes. (Also general shoutout to Ezra for helping bring EA ideas to the mainstream - he's...
I've always been pretty confused about this. The standard AI risk scenarios usually (though I think not always) suppose that advanced AI wants not to be shut down. As commonly framed, the AI will fool humanity into believing it is aligned so as not to be turned off, until -...