Nick_Tarleton comments on Open Thread: June 2009 - Less Wrong

4 Post author: Cyan 01 June 2009 06:46PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (142)

You are viewing a single comment's thread. Show more comments above.

Comment author: Nick_Tarleton 02 June 2009 01:44:10AM *  1 point [-]

I understand the general point, but "AI is disabled" seems like a special case, in that an AI able to do any sort of reasoning about itself, allocate internal resources, etc. (I don't know how necessary this is for it to do anything useful), will have to have concepts in its qualitative ontology of, or sufficient to define, its disability – though perhaps not in a way easily available for framing a goal system (e.g. if it developed them itself, assuming it could build up to them in their absence), and probably complicated in some other ways that haven't occurred to me in two minutes.

Comment author: Eliezer_Yudkowsky 02 June 2009 05:03:23AM 3 points [-]

Suppose the AI builds devices in the environment, especially computational devices designed to offload cognitive labor. What do you want to happen when the AI is "switched off"? Hence, magical category.

Comment author: Vladimir_Nesov 02 June 2009 09:46:20PM *  0 points [-]

Interesting, I didn't think of this situation. How do you define "lack of intelligence" or "removal of the effect of intelligence" in the environment, so that an AI can implement that state? How is this state best achieved?

Once the system is established, the world will ever be determined by a specific goal system, even if the goal is for the world to appear as if no AI is present, starting from a certain time. The best solution is for AI to pretend of not being present, "pulling the planets along their elliptic orbits".

Comment author: Nick_Tarleton 02 June 2009 05:25:40AM 0 points [-]

D'oh. Yes, of course, that breaks it.

Comment author: SoullessAutomaton 02 June 2009 09:45:39PM 4 points [-]

As an aside, "waiting for Eliezer to find a loophole" probably does not constitute a safe and effective means of testing AI utility functions. This is something we want provable from first principles, not "proven" by "well, I can't think of a counterexample".

Comment author: Nick_Tarleton 02 June 2009 10:01:45PM 1 point [-]

Of course, hence "...and probably complicated in some other ways that haven't occurred to me in two minutes.".

Comment author: SoullessAutomaton 02 June 2009 11:00:03PM *  1 point [-]

Right. I know you realize this, and the post was fine in the context of "random discussion on the internet". However, if someone wants to actually, seriously specify a utility function for an AI any description that starts with "here's a high-level rule to avoid bad things" and then works from there looking for potential loopholes is deeply and fundamentally misguided completely independently of the rule proposed.