I wanted to share a new paper from the special issue on AI safety that I'm editing, which takes up the influential idea that evolutionary theory gives us some reason to think that the project of value alignment is bound to fail and (in my opinion) shows that it has serious problems.
I wanted to share a new paper from the special issue on AI safety that I'm editing, which takes up the influential idea that evolutionary theory gives us some reason to think that the project of value alignment is bound to fail and (in my opinion) shows that it has serious problems.
If you don't have institutional access to the article, I'm also hosting it on my personal website here: https://www.cd.kg/wp-content/uploads/2024/10/selfish_machine.pdf