You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

eli_sennesh comments on Steelmaning AI risk critiques - Less Wrong Discussion

26 Post author: Stuart_Armstrong 23 July 2015 10:01AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (98)

You are viewing a single comment's thread. Show more comments above.

Comment author: [deleted] 03 August 2015 03:59:33AM 1 point [-]

Okay, so we just have to determine human terminal values in detail, and plug them into a powerful maximizer.

Why do you even go around thinking that the concept of "terminal values", which is basically just a consequentialist steelmanning Aristotle, cuts reality at the joints?

For starters, you want to be able to prove formally that its goals will remain stable as it self-modifies

That part honestly isn't that hard once you read the available literature about paradox theorems.