You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Manfred comments on The AI That Pretends To Be Human - Less Wrong Discussion

1 Post author: Houshalter 02 February 2016 07:39PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (69)

You are viewing a single comment's thread.

Comment author: Manfred 03 February 2016 08:32:27PM 0 points [-]

I think the "second AI" really should just be an algorithm that the first AI runs in order to evaluate actions (it should not have to learn to predict the second AI based on signals in a reward channel). A logical rather than physical connection. Otherwise bad behavior is incentivized, to control the reward channel.

GANs are neat, but their highest-scoring images aren't all that natural - I'd be worried about any implementation of this using current ideas about supervised learning. Certainly if you desire reasoning like "this action would lead to the AI taking over the world, and that's not something a human would do," you'll need some futuristic AI design.