Eliezer_Yudkowsky comments on Timeless Decision Theory: Problems I Can't Solve - Less Wrong

39 Post author: Eliezer_Yudkowsky 20 July 2009 12:02AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (153)

You are viewing a single comment's thread. Show more comments above.

Comment author: Eliezer_Yudkowsky 20 July 2009 01:52:43AM 4 points [-]

"I believe X to be like me" => "whatever I decide, X will decide also" seems tenuous without some proof of likeness that is beyond any guarantee possible in humans...

Maybe you could aspire to such determinism in a proven-correct software system running on proven-robust hardware.

Well, yeah, this is primarily a theory for AIs dealing with other AIs.

You could possibly talk about human applications if you knew that the N of you had the same training as rationalists, or if you assigned probabilities to the others having such training.

Comment author: Tedav 28 February 2014 04:44:00PM 0 points [-]

For X to be able to model the decisions of Y with 100% accuracy, wouldn't X require a more sophisticated model?

If so, why would supposedly symmetrical models retain this symmetry?

Comment author: Eliezer_Yudkowsky 28 February 2014 08:10:49PM 3 points [-]

For X to be able to model the decisions of Y with 100% accuracy, wouldn't X require a more sophisticated model?

Nope. http://arxiv.org/abs/1401.5577