You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Vaniver comments on Q&A with experts on risks from AI #1 - Less Wrong Discussion

29 Post author: XiXiDu 08 January 2012 11:46AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (66)

You are viewing a single comment's thread. Show more comments above.

Comment author: Vaniver 21 January 2013 11:01:57PM 1 point [-]

Although it's possible that I just abused the one minimization-like part you accidentally left in there and there is some relatively simple patch that I'm not seeing.

I meant "resources used" in the sense of "resources directed towards this goal" rather than "resources drawn from the metropolitian utility company"- if the streetlamps play the stock market and accumulate a bunch of money, spending that money will still decrease their utility, and so unless they can spend the money in a way that improves the illumination cost-effectively they won't.

Now, defining "resources directed towards this goal" in a way that's machine-understandable is a hard problem. But if we already have an AI that thinks causally- such that it can actually make these plans and enact them- then it seems to me like that problem has already been solved.

Comment author: Elithrion 21 January 2013 11:58:36PM 0 points [-]

Hm, all right, fair enough. That actually sounds plausible, assuming we can be sure that the AI appropriately takes account of something vaguely along the lines of "all resources that will be used in relation to this problem", including, for example, creating a copy of itself that does not care about resources used and obfuscates its activities from the original. Which will probably be doable at that point.