thomblake comments on Two straw men fighting - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (157)
As I've stated before, no AI can predict its own decisions in that sense (i.e. in detail, before it has made them.) Knowing its source code doesn't help; it has to run the code in order to know what result it gets.
This is false for some algorithms, and so I imagine it would be false for the entirety of the AI's source code. For example (ANSI C):
I know that i is equal to 5 after this code is executed, and I know that without executing the code in any sense.
That isn't an algorithm for making decisions.
No, but note the text:
It is, incidently, trivial to alter the code to an algorithm for making decisions and also simple to make it an algorithm that can predict it's decision before making it.
The doselfanalysis method (do they call them methods or functions? Too long since I've used C) can browse the entire source code of the AI, determine that the above piece of code is the algorithm for making the relevant decision, prove that do_self_analysis doesn't change anything or perform any output and does return in finite time and then go on to predict that the AI will behave like a really inefficient defection rock. Quite a while later it will actually make the decision to defect.
All rather pointless but the concept is proved.
When the AI runs the code for predicting it's action, it will have the subjective experience of making the decision. Later "it will actually make the decision to defect" only in the sense that the external result will come at that time. If you ask it when it made it's decision, it will point to the time when it analyzed the code.
You are mistaken. I consider the explanations given thus far by myself and others sufficient. (No disrespect intended beyond that implicit in the fact of disagreement itself and I did not vote on the parent.)
The explanations given say nothing about the AI's subjective experience, so they can't be sufficient to refute my claim about that.
Consider my reply to be to the claim:
If you ask the AI when it made its decision it will either point to the time after the analysis or it will be wrong.
I avoided commenting on the 'subjective experience' side of things because I thought it was embodying a whole different kind of confusion. It assumes that the AI executes some kind of 'subjective experience' reasoning that is similar to that of humans (or some subset thereof). This quirk relies on lacking any strong boundaries between thought processes. People usually can't predict their decisions without making them. For both the general case and the specific case of the code I gave a correctly implemented module that could be given the label 'subjective experience' would see the difference between prediction and analysis.
I upvoted the parent for the use of it's. I usually force myself to write its in that context but cringe while doing so. The syntax of the English language is annoying.
Really? Do you also cringe when using theirs, yours, ours, mine, and thine?
Mine and thine? They don't belong in the category. The flaw isn't that all words about possession should have an apostrophe. The awkwardness is that the pattern of adding the "s" to the end to indicate ownership is the same from "Fred's" to "its" but arbitrarily not punctuated in the same way. The (somewhat obsolete) "ine" is a distinct mechanism of creating a possessive pronoun which while adding complexity at least doesn't add inconsistency.
As for "theirs, yours and ours", they prompt cringes in decreasing order of strength (in fact, it may not be a coincidence that you asked in that order). Prepend "hers" to the list and append "his". "Hers" and "theirs" feel more cringe-worthy, as best as I can judge, because they are closer in usage to "Fred's" while "ours" is at least a step or two away. "His" is a special case in as much as it is a whole different word. It isn't a different mechanism like "thine" or "thy" but it isn't "hes" either. I have never accidentally typed "hi's".
"If you ask the AI when it made its decision it will either point to the time after the analysis or it will be wrong."
I use "decision" precisely to refer the experience that we have when we make a decision, and this experience has no mathematical definition. So you may believe yourself right about this, but you don't have (and can't have) any mathematical proof of it.
(I corrected this comment so that it says "mathematical proof" instead of proof in general.)
If you believe that we can't have any proof of it, then you're wasting our time with arguments.
I think most people on LessWrong are using "decision" in the sense used in Decision Theory.
Making a claim, and then, when given counter-arguments, claiming that one was using an exotic definition seems close to logical rudeness to me.
How many legs does an animal have if I call a tail a leg and believe all animals are quadrupeds?
No, but surely some chunks of similarly-transparent code would appear in an algorithm for making decisions. And since I can read that code and know what it outputs without executing it, surely a superintelligence could read more complex code and know what it outputs without executing it. So it is patently false that in principle the AI will not be able to know the output of the algorithm without executing it.
Any chunk of transparent code won't be the code for making an intelligent decision. And the decision algorithm as a whole won't be transparent to the same intelligence, but perhaps only to something still more intelligent.
Do you have a proof of this statement? If so, I will accept that it is not in principle possible for an AI to predict what its decision algorithm will return without executing it.
Of course, logical proof isn't entirely necessary when you're dealing with Bayesians, so I'd also like to see any evidence that you have that favors this statement, even if it doesn't add up to a proof.
It's not possible to prove the statement because we have no mathematical definition of intelligence.
Eliezer claims that it is possible to create a superintelligent AI which is not conscious. I disagree with this because it is basically saying that zombies are possible. True, he would say that he only believes that human zombies are impossible, not that zombie intelligences in general are impossible. But in that case he has no idea whatsoever what consciousness corresponds to in the physical world, and in fact has no reason not to accept dualism.
My position is more consistent: all zombies are impossible, and any intelligent being will be conscious. So it will also have the subjective experience of making decisions. But it is essential to this experience that you don't know what you're going to do before you do it; when you experience knowing what you're going to do, you experience deciding to do it.
Therefore any AI that runs code capable of predicting its decisions, will at that very time subjectively experience making those decisions. And on the other hand, given that a block of code will not cause it to feel the sensation of deciding, that block of code must be incapable of predicting its decision algorithm.
You may still disagree, but please note that this is entirely consistent with everything you and wedrifid have argued, so his claim that I have been refuted is invalid.
As I recall, Eliezer's definition of consciousness is borrowed from GEB- it's when the mind examines itself, essentially. That has very real physical consequences, so the idea of non-conscious AGI doesn't support the idea of zombies, which require consciousness to have no physical effects.
Any AGI would be able to examine itself, so if that is the definition of consciousness, every intelligence would be conscious. But Eliezer denies the latter, so he also implicitly denies that definition of consciousness.
I'm not sure I am parsing correctly what you've wrote. It may rest with your use of the word "intelligence"- how are you defining that term?
Yes we do, ability to apply optimization pressure in a wide variety of environments. The platonic ideal of which is AIXI.
Can you please provide a link?
http://lesswrong.com/lw/x5/nonsentient_optimizers/
Thank you. I agree with Eliezer for reasons touched on in my comments to simplicio's Consciousness of simulations & uploads thread.
I don't have any problem granting that "any intelligent being will be conscious", nor that "It will have the subjective experience of making decisions", though that might just be because I don't have a formal specification of either of those - we might still be talking past each other there.
I don't grant this. Can you elaborate?
I'm not sure that's true, or in what sense it's true. I know that if someone offered me a million dollars for my shoes, I would happily sell them my shoes. Coming to that realization didn't feel to me like the subjective feeling of deciding to sell something to someone at the time, as compared to my recollection of past transactions.
Okay, that follows from the previous claim.
If I were moved to accept your previous claim, I would now be skeptical of the claim that "a block of code will not cause it to feel the sensation of deciding". Especially since we've already shown that some blocks of code would be capable of predicting some decision algorithms.
This follows, but I draw the inference in the opposite direction, as noted above.
I would distinguish between "choosing" and "deciding". When we say "I have some decisions to make," we also mean to say that we don't know yet what we're going to do.
On the other hand, it is sometimes possible for you to have several options open to you, and you already know which one you will "choose". Your example of the shoes and the million dollars is one such case; you could choose not to take the million dollars, but you would not, and you know this in advance.
Given this distinction, if you have a decision to make, as soon as you know what you will or would do, you will experience making a decision. For example, presumably there is some amount of money ($5? $20? $50? $100? $300?) that could be offered for your shoes such that you are unclear whether you should take the offer. As soon as you know what you would do, you will feel yourself "deciding" that "if I was offered this amount, I would take it." It isn't a decision to do something concretely, but it is still a decision.
Now, I am not certain about this, but we have to examine that code before we know it's outcome.
While this isn't "Running" the code in the traditional sense of computation as we are familiar with it today, it does seem that the code is sort of run by our brains as a simulation as we scan it.
As sort of meta-process if you will...
I could be so wrong about that though... eh...
Also, that code is useless really, except maybe as a wait function... It doesn't really do anything (Not sure why Unknowns gets voted up in the first post above, and down below)...
Also, leaping from some code to the Entirety of an AI's source code seems to be a rather large leap.
"some code" is part of "the entirety of an AI's source code" - if it doesn't need to execute some part of the code, then it doesn't need to execute the entirety of the code.