Posts

Sorted by New

Wiki Contributions

Comments

Sorted by
mtreder-40

Yes, but after the AGI finds out what a paperclip is, it will then, if it is an AGI, start questioning why it was designed with the goal of building paperclips in the first place. And that's where the friendly AI fallacy falls apart.