eli_sennesh comments on MIRI's technical research agenda - LessWrong

33 Post author: So8res 23 December 2014 06:45PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (52)

You are viewing a single comment's thread. Show more comments above.

Comment author: [deleted] 12 January 2015 05:53:54AM 1 point [-]

While I actually agree that tool AI goals can be programmed, if you want to keep the whole thing from turning unsafely agenty, you're going to have to strictly separate the inductive reasoning from the actual tool run: run induction for a while, then use tool-mode to compose plans over the induced models of the world, potentially after censoring those models for safety.