Unknowns comments on Two straw men fighting - Less Wrong

2 Post author: JanetK 09 August 2010 08:53AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (157)

You are viewing a single comment's thread. Show more comments above.

Comment author: Unknowns 09 August 2010 02:51:44PM -1 points [-]

That isn't an algorithm for making decisions.

Comment author: wedrifid 09 August 2010 03:17:12PM *  2 points [-]

That isn't an algorithm for making decisions.

No, but note the text:

This is false for some algorithms, and so I imagine it would be false for the entirety of the AI's source code.

It is, incidently, trivial to alter the code to an algorithm for making decisions and also simple to make it an algorithm that can predict it's decision before making it.

do_self_analysis();
unsigned long i;
unsigned long j;
for (i=0; i<ULONG_MAX-1; i++)
for (j=0; j<ULONG_MAX-1; j++);
if(i > 2) return ACTION_DEFECT;
return ACTION_COOPERATE;

The doselfanalysis method (do they call them methods or functions? Too long since I've used C) can browse the entire source code of the AI, determine that the above piece of code is the algorithm for making the relevant decision, prove that do_self_analysis doesn't change anything or perform any output and does return in finite time and then go on to predict that the AI will behave like a really inefficient defection rock. Quite a while later it will actually make the decision to defect.

All rather pointless but the concept is proved.

Comment author: Unknowns 09 August 2010 03:26:37PM -2 points [-]

When the AI runs the code for predicting it's action, it will have the subjective experience of making the decision. Later "it will actually make the decision to defect" only in the sense that the external result will come at that time. If you ask it when it made it's decision, it will point to the time when it analyzed the code.

Comment author: wedrifid 09 August 2010 03:31:34PM *  2 points [-]

You are mistaken. I consider the explanations given thus far by myself and others sufficient. (No disrespect intended beyond that implicit in the fact of disagreement itself and I did not vote on the parent.)

Comment author: Unknowns 09 August 2010 03:33:29PM 0 points [-]

The explanations given say nothing about the AI's subjective experience, so they can't be sufficient to refute my claim about that.

Comment author: wedrifid 09 August 2010 03:55:52PM 2 points [-]

Consider my reply to be to the claim:

If you ask it when it made it's decision, it will point to the time when it analyzed the code.

If you ask the AI when it made its decision it will either point to the time after the analysis or it will be wrong.

I avoided commenting on the 'subjective experience' side of things because I thought it was embodying a whole different kind of confusion. It assumes that the AI executes some kind of 'subjective experience' reasoning that is similar to that of humans (or some subset thereof). This quirk relies on lacking any strong boundaries between thought processes. People usually can't predict their decisions without making them. For both the general case and the specific case of the code I gave a correctly implemented module that could be given the label 'subjective experience' would see the difference between prediction and analysis.

I upvoted the parent for the use of it's. I usually force myself to write its in that context but cringe while doing so. The syntax of the English language is annoying.

Comment author: thomblake 09 August 2010 04:51:50PM 1 point [-]

I upvoted the parent for the use of it's. I usually force myself to write its in that context but cringe while doing so. The syntax of the English language is annoying.

Really? Do you also cringe when using theirs, yours, ours, mine, and thine?

Comment author: wedrifid 09 August 2010 05:16:05PM *  1 point [-]

Mine and thine? They don't belong in the category. The flaw isn't that all words about possession should have an apostrophe. The awkwardness is that the pattern of adding the "s" to the end to indicate ownership is the same from "Fred's" to "its" but arbitrarily not punctuated in the same way. The (somewhat obsolete) "ine" is a distinct mechanism of creating a possessive pronoun which while adding complexity at least doesn't add inconsistency.

As for "theirs, yours and ours", they prompt cringes in decreasing order of strength (in fact, it may not be a coincidence that you asked in that order). Prepend "hers" to the list and append "his". "Hers" and "theirs" feel more cringe-worthy, as best as I can judge, because they are closer in usage to "Fred's" while "ours" is at least a step or two away. "His" is a special case in as much as it is a whole different word. It isn't a different mechanism like "thine" or "thy" but it isn't "hes" either. I have never accidentally typed "hi's".

Comment author: thomblake 09 August 2010 05:19:32PM *  0 points [-]

You're just reading the wrong pattern. There are simple, consistent rules:

  1. When making a noun possessive, EDIT: <strike>add 's</strike> use the appropriate possessive form with an apostrophe
  2. When making a pronoun possessive, use the appropriate possessive pronoun (none of which have an apostrophe)

EDIT: <strike>Leaving out " Jesus' " for the moment...</strike>

Comment author: wedrifid 09 August 2010 05:51:09PM 1 point [-]

You're just reading the wrong pattern.

No, I'm not reading the wrong pattern. I'm criticising the pattern in terms of the objective and emotional-subjective criteria that I use for evaluating elements of languages and communication patterns in general. I am aware of the rules in question and more than capable of implementing it and the hundreds of other rules that go into making our language.

The undesirable aspect of this part of the language is this: It is not even remotely coincidental that we add the "ss" sound to the end of a noun to make it possessive and that most modern possessive pronouns are just the pronoun with a "ss" sound at the end. Nevertheless, the rule is "use the appropriate possessive pronoun"... that's a bleeding lookup table! A lookup table for something that is nearly always an algorithmic modification is not something I like in a language design. More importantly, when it comes to the spoken word the rule for making *nouns possessive is "almost always add 'ss'". 'Always' is better than 'almost always' (but too much to ask). Given 'almost always' , the same kind of rule for converting them all to written form would be far superior.

According to subjectively-objective criteria, this feature of English sucks. If nothing else it would be fair to say that my 'subjective' is at least not entirely arbitrary, whether or not you share the same values with respect to language.

Comment author: NancyLebovitz 09 August 2010 05:21:26PM 0 points [-]

There's one more rule-- if the noun you're making possessive ends with an s (this applies to both singular and plural nouns), just add an apostrophe.

Comment author: Unknowns 09 August 2010 04:02:50PM *  -1 points [-]

"If you ask the AI when it made its decision it will either point to the time after the analysis or it will be wrong."

I use "decision" precisely to refer the experience that we have when we make a decision, and this experience has no mathematical definition. So you may believe yourself right about this, but you don't have (and can't have) any mathematical proof of it.

(I corrected this comment so that it says "mathematical proof" instead of proof in general.)

Comment author: thomblake 09 August 2010 04:15:21PM 2 points [-]

So you may believe yourself right about this, but you don't have (and can't have) any proof of it.

If you believe that we can't have any proof of it, then you're wasting our time with arguments.

Comment author: Unknowns 09 August 2010 04:20:36PM *  -2 points [-]

You might have a proof of it, but not a mathematical proof.

Also note that your comment that I would be "wasting our time" implies that you think that you couldn't be wrong.

Comment author: Emile 09 August 2010 04:14:33PM *  2 points [-]

I think most people on LessWrong are using "decision" in the sense used in Decision Theory.

Making a claim, and then, when given counter-arguments, claiming that one was using an exotic definition seems close to logical rudeness to me.

Comment author: wedrifid 09 August 2010 04:51:11PM *  3 points [-]

Making a claim, and then, when given counter-arguments, claiming that one was using an exotic definition seems close to logical rudeness to me.

It also does his initial position a disservice. Rereading the original claim with the professed intended meaning changes it from "not quite technical true" to, basically, nonsense (at least in as much as it claims to pertain to AIs).

Comment author: Unknowns 09 August 2010 04:22:00PM -2 points [-]

I don't think my definition is either exotic or inconsistent with the sense used in decision theory.

Comment author: wedrifid 09 August 2010 04:53:52PM 3 points [-]

I don't think my definition is ... inconsistent with the sense used in decision theory.

You defined decision as a mathematical undefinable experience and suggested that it cannot be subject to proofs. That isn't even remotely compatible with the sense used in decision theory.

Comment author: wedrifid 09 August 2010 04:38:16PM 0 points [-]

How many legs does an animal have if I call a tail a leg and believe all animals are quadrupeds?

Comment author: thomblake 09 August 2010 03:18:50PM 1 point [-]

No, but surely some chunks of similarly-transparent code would appear in an algorithm for making decisions. And since I can read that code and know what it outputs without executing it, surely a superintelligence could read more complex code and know what it outputs without executing it. So it is patently false that in principle the AI will not be able to know the output of the algorithm without executing it.

Comment author: Unknowns 09 August 2010 03:27:51PM 1 point [-]

Any chunk of transparent code won't be the code for making an intelligent decision. And the decision algorithm as a whole won't be transparent to the same intelligence, but perhaps only to something still more intelligent.

Comment author: thomblake 09 August 2010 03:41:40PM 0 points [-]

Any chunk of transparent code won't be the code for making an intelligent decision.

Do you have a proof of this statement? If so, I will accept that it is not in principle possible for an AI to predict what its decision algorithm will return without executing it.

Of course, logical proof isn't entirely necessary when you're dealing with Bayesians, so I'd also like to see any evidence that you have that favors this statement, even if it doesn't add up to a proof.

Comment author: Unknowns 09 August 2010 03:54:53PM *  0 points [-]

It's not possible to prove the statement because we have no mathematical definition of intelligence.

Eliezer claims that it is possible to create a superintelligent AI which is not conscious. I disagree with this because it is basically saying that zombies are possible. True, he would say that he only believes that human zombies are impossible, not that zombie intelligences in general are impossible. But in that case he has no idea whatsoever what consciousness corresponds to in the physical world, and in fact has no reason not to accept dualism.

My position is more consistent: all zombies are impossible, and any intelligent being will be conscious. So it will also have the subjective experience of making decisions. But it is essential to this experience that you don't know what you're going to do before you do it; when you experience knowing what you're going to do, you experience deciding to do it.

Therefore any AI that runs code capable of predicting its decisions, will at that very time subjectively experience making those decisions. And on the other hand, given that a block of code will not cause it to feel the sensation of deciding, that block of code must be incapable of predicting its decision algorithm.

You may still disagree, but please note that this is entirely consistent with everything you and wedrifid have argued, so his claim that I have been refuted is invalid.

Comment author: Randaly 09 August 2010 04:52:35PM 1 point [-]

As I recall, Eliezer's definition of consciousness is borrowed from GEB- it's when the mind examines itself, essentially. That has very real physical consequences, so the idea of non-conscious AGI doesn't support the idea of zombies, which require consciousness to have no physical effects.

Comment author: Unknowns 09 August 2010 04:57:54PM 0 points [-]

Any AGI would be able to examine itself, so if that is the definition of consciousness, every intelligence would be conscious. But Eliezer denies the latter, so he also implicitly denies that definition of consciousness.

Comment author: JoshuaZ 09 August 2010 05:02:20PM 0 points [-]

Any AGI would be able to examine itself, so if that is the definition of consciousness, every intelligence would be conscious. But Eliezer denies the latter, so he also implicitly denies that definition of consciousness.

I'm not sure I am parsing correctly what you've wrote. It may rest with your use of the word "intelligence"- how are you defining that term?

Comment author: Unknowns 09 August 2010 05:03:31PM *  0 points [-]

You could replace it with "AI." Any AI can examine itself, so any AI will be conscious, if consciousness is or results from examining itself. I agree with this, but Eliezer does not.

Comment author: LucasSloan 10 August 2010 08:49:16AM *  0 points [-]

we have no mathematical definition of intelligence.

Yes we do, ability to apply optimization pressure in a wide variety of environments. The platonic ideal of which is AIXI.

Comment author: torekp 10 August 2010 01:43:46AM 0 points [-]

Eliezer claims that it is possible to create a superintelligent AI which is not conscious.

Can you please provide a link?

Comment author: Eliezer_Yudkowsky 10 August 2010 03:04:25AM 0 points [-]
Comment author: torekp 22 August 2010 03:35:56PM 0 points [-]

Thank you. I agree with Eliezer for reasons touched on in my comments to simplicio's Consciousness of simulations & uploads thread.

Comment author: thomblake 09 August 2010 04:07:24PM 0 points [-]

My position is more consistent: all zombies are impossible, and any intelligent being will be conscious. So it will also have the subjective experience of making decisions. But it is essential to this experience that you don't know what you're going to do before you do it; when you experience knowing what you're going to do, you experience deciding to do it.

Therefore any AI that runs code capable of predicting its decisions, will at that very time subjectively experience making those decisions. And on the other hand, given that a block of code will not cause it to feel the sensation of deciding, that block of code must be incapable of predicting its decision algorithm.

I don't have any problem granting that "any intelligent being will be conscious", nor that "It will have the subjective experience of making decisions", though that might just be because I don't have a formal specification of either of those - we might still be talking past each other there.

But it is essential to this experience that you don't know what you're going to do before you do it

I don't grant this. Can you elaborate?

when you experience knowing what you're going to do, you experience deciding to do it.

I'm not sure that's true, or in what sense it's true. I know that if someone offered me a million dollars for my shoes, I would happily sell them my shoes. Coming to that realization didn't feel to me like the subjective feeling of deciding to sell something to someone at the time, as compared to my recollection of past transactions.

Therefore any AI that runs code capable of predicting its decisions, will at that very time subjectively experience making those decisions.

Okay, that follows from the previous claim.

And on the other hand, given that a block of code will not cause it to feel the sensation of deciding, that block of code must be incapable of predicting its decision algorithm.

If I were moved to accept your previous claim, I would now be skeptical of the claim that "a block of code will not cause it to feel the sensation of deciding". Especially since we've already shown that some blocks of code would be capable of predicting some decision algorithms.

that block of code must be incapable of predicting its decision algorithm.

This follows, but I draw the inference in the opposite direction, as noted above.

Comment author: Unknowns 09 August 2010 04:19:59PM 0 points [-]

I would distinguish between "choosing" and "deciding". When we say "I have some decisions to make," we also mean to say that we don't know yet what we're going to do.

On the other hand, it is sometimes possible for you to have several options open to you, and you already know which one you will "choose". Your example of the shoes and the million dollars is one such case; you could choose not to take the million dollars, but you would not, and you know this in advance.

Given this distinction, if you have a decision to make, as soon as you know what you will or would do, you will experience making a decision. For example, presumably there is some amount of money ($5? $20? $50? $100? $300?) that could be offered for your shoes such that you are unclear whether you should take the offer. As soon as you know what you would do, you will feel yourself "deciding" that "if I was offered this amount, I would take it." It isn't a decision to do something concretely, but it is still a decision.