MattMahoney comments on Reply to Holden on 'Tool AI' - Less Wrong

94 Post author: Eliezer_Yudkowsky 12 June 2012 06:00PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (348)

You are viewing a single comment's thread. Show more comments above.

Comment author: MattMahoney 05 July 2012 05:38:30PM 0 points [-]

If we were smart enough to understand its policy, then it would not be smart enough to be dangerous.

Comment author: wedrifid 05 July 2012 06:13:31PM 2 points [-]

If we were smart enough to understand its policy, then it would not be smart enough to be dangerous.

That doesn't seem true. Simple policies can be dangerous and more powerful than I am.

Comment author: Nebu 17 February 2016 11:24:41AM 0 points [-]

To steelman the parent argument a bit, a simple policy can be dangerous, but if an agent proposed a simple and dangerous policy to us, we probably would not implement it (since we could see that it was dangerous), and thus the agent itself would not be dangerous to us.

If the agent were to propose a policy that, as far as we could tell, appears safe, but was in fact dangerous, then simultaneously:

  1. We didn't understand the policy.
  2. The agent was dangerous to us.