XiXiDu comments on One Medical? Expansion of MIRI? - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (8)
I did not follow the interviews in detail. But I doubt that most of these AI safety engineers believe that you could achieve AI software that can drive trains and fly planes without crashing, but which nonetheless drives and flies people to locations they do not desire. In other words, my guess is that these people believe that without being able to prove that programs meet certain conditions you won't achieve FOOM in the first place. What they probably do not believe is MIRI's idea of an AI that works perfectly along a huge number of dimensions (e.g. making itself superhuman smart, solving the protein folding problem etc.), but which nonetheless fails at doing what people designed it to do (except that it does not fail at all the aforementioned tasks).
The problem isn't so much that the AI doesn't do what is was designed to do, it's that what you implemented is subtly different from what you designed. This is something that commonly happens in programming, not just a hypothetical concern.