You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

paulfchristiano comments on Restrictions that are hard to hack - Less Wrong Discussion

6 Post author: Stuart_Armstrong 09 March 2015 01:52PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (8)

You are viewing a single comment's thread.

Comment author: paulfchristiano 28 October 2015 06:33:13PM 0 points [-]

If you want to talk about the behavior of the AI being uninformative, you need to talk about the distribution over possible values over R. If the distribution is just "it exists" or "it doesn't," then it's clear that the AI will just have to satisfy R in every case, and you don't get anything beyond the restriction itself.

If there is some broader distribution, then it's less clear what happens, but as far as I can tell this is no better than simply having the AI care about an unknown requirement from that distribution.

Comment author: Stuart_Armstrong 28 October 2015 06:39:39PM 0 points [-]

There is an R, it's given. There is a distribution over possible R's for agents that only know the data E, F, B(), and v.

But this approach seems very wobbly to me; I no longer give it much potential.