You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

jsteinhardt comments on AGI/FAI Theorist for Hire - Less Wrong Discussion

12 Post author: Peter_de_Blanc 15 July 2011 03:50PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (5)

You are viewing a single comment's thread.

Comment author: jsteinhardt 15 July 2011 06:47:14PM 14 points [-]

I'm also interested in repurposing machine learning algorithms used for finding plausible hypotheses about data distributions into algorithms for finding action policies with high expected utility.

While I'm not in a position to hire you, I think that this is an extremely important problem. In the case where the utility function is known, I think there is lots of low-hanging fruit that will lead to progress in more "physical" application areas of machine learning like computer vision and robotics. In the case where the utility function is unknown, I think the problem is harder (at the level say of a PhD thesis), but would be a crucial step towards making progress on FAI.

If you're interested in talking to me about either of these then I'd be happy to, assuming you have enough of a statistical background for me to get my thoughts across without too much of an explanatory burden. Assuming you haven't already decided on a specific set of algorithms, I have some ideas here that I don't currently have time to pursue myself that I think could lead to a publication in a good machine learning journal if done well.

I've moved in the direction of predicting AGI soonish (5-20 years)

This timeline is much sooner than I would predict. Could you perhaps point me to a few sources that you think would cause me to update my estimate towards yours?

Comment author: Peter_de_Blanc 17 July 2011 08:33:30AM 1 point [-]

There was a specific set of algorithms that got me thinking about this topic, but now that I'm thinking about the topic I'd like to look at more stuff. I would proceed by identifying spaces of policies within a domain, and then looking for learning algorithms that deal with those sorts of spaces. For sequential decision-making problems in simple settings, dynamic bayesian networks can be used both as models of an agent's environment and as action policies.

I'd be interested in talking. You can e-mail me at peter@spaceandgames.com.