You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

The Kolmogorov complexity of a superintelligence

2 Post author: Thomas 26 June 2011 12:11PM

The most simple, able to self-improve "seed" for a superintelligence must be how complex?

Give your (wild) estimation or a pure guess in the terms of bits.

Do you reckon it is 1000 bits enough? At least 1000000 is required? More than 10^20 bits?

I wonder what your approximate numbers are.

Thank you!

Comments (30)

Comment author: cousin_it 26 June 2011 02:33:24PM *  5 points [-]

The K-complexity of a superintelligence can't be higher than the K-complexity of a program that searches for a superintelligence, enumerating all programs in order and asking them hard questions to check their abilities. I don't see any reason why the latter should be over a megabyte, probably much less.

Comment author: MrMind 30 June 2011 02:58:14PM 1 point [-]

It must be, because this verification procedure also produces infinites false positives, for example all the hash tables which happen to have the correct answers by chance. That is, the procedure doesn't produce more information than simply asking "Are you a superintelligence?".

Comment author: Zetetic 26 June 2011 08:53:38PM *  1 point [-]

I don't see any reason why the latter should be over a megabyte, probably much less.

In contrast, it seems to me like the verification procedure for a friendly superintelligence could be quite long. Do you agree with this? It would definitely be longer than for just any superintelligence, but I haven't got a clue about how much; just that we could potentially be adding some pretty heavy constraints.

Comment author: cousin_it 26 June 2011 09:05:02PM *  0 points [-]

Yes, I agree. An upper bound on the complexity of a friendly superintelligence would be the total information content of all human brains, but you could probably compress that a lot.

Comment author: wedrifid 26 June 2011 09:17:51PM *  4 points [-]

Yes, I agree. An upper bound on the complexity of a friendly superintelligence would be the total information content of all human brains, but you could probably compress that a lot.

False. The upper bound on the complexity of a Friendly<all humans> superintelligence is [(total information content of all brains) + (minimum complexity of a what it takes to identify a superintelligence with a defined goal of deriving its objective from a set of agents according to some mechanism that represents what it would mean to be 'friendly' to them)].

ie. Just having all the information content of all human brains is not enough. You can't avoid defining Friendliness in the search program.

Comment author: Wei_Dai 26 June 2011 09:57:09PM *  1 point [-]

I agree with wedrifid here. We don't seem to have a valid argument showing that "an upper bound on the complexity of a friendly superintelligence would be the total information content of all human brains". I would like to point out that if the K-complexity of friendly superintelligence is greater than that, then there is no way for us to build a friendly superintelligence except by luck (i.e., most Everett branches are doomed to fail to build a friendly superintelligence) or by somehow exploiting uncomputable physics.

Comment author: cousin_it 26 June 2011 10:11:35PM *  0 points [-]

Technically you can cheat by using the information in human brains to create upload-based superintelligences along the lines of Stuart's proposal, make them do the research for you, etc., so it seems likely to me that the upper bound should hold... but I appreciate your point and agree that my comment was wrong as stated.

Comment author: wedrifid 27 June 2011 12:57:37AM *  2 points [-]

When talking about upper bounds we cannot afford to just cheat saying humans can probably figure it out. That isn't an upper bound - it is an optimistic prediction about human potential. Moreover we still need a definition of Friendliness built in so we can tell whether this thing that the human researchers come up with is Friendliness or some other thing with that name. (Even an extremely 'meta' reference to a definition is fine but still requires more bits to point to which part of the humans is able to define Friendliness.)

Upper bounds are hard. But yes, I know your understanding of the area is solid and your ancestor comment serves as the definitive answer to the question of the post. I disagree only with the statement as made.

Comment author: Vladimir_Nesov 26 June 2011 09:56:04PM 0 points [-]

False.

You are most likely not disagreeing with the intended meaning, so using words like "false" to motivate the clarification you were making is wrong.

Comment author: wedrifid 27 June 2011 02:43:02AM *  1 point [-]

You are most likely not disagreeing with the intended meaning, so using words like "false" to motivate the clarification you were making is wrong.

No Vladimir. My disagreement was with the statement and the statement was, indeed false. That doesn't mean I disagree with cousin_it's overall philosophy. It just means I am calling a single false claim false.

You are wrong.

Comment author: ciphergoth 28 June 2011 09:15:46AM 1 point [-]

I'm confused - a Friendly AI doesn't start with information about the insides of brains, it only starts with enough information to correctly identify human beings in order to know whose preferences to infer its values from.

Comment author: Zetetic 26 June 2011 09:12:12PM 0 points [-]

but you could probably compress that a lot.

This definitely fits my intuition. It seems like CEV style friendliness, for example, could at least be reduced to something on the order of complexity of a single "averaged" human brain. This is all very vague of course, I'm mostly just trying to feel out some intuitions on this.

Comment author: wedrifid 26 June 2011 09:27:53PM -2 points [-]

It seems like CEV style friendliness, for example, could at least be reduced to something on the order of complexity of a single "averaged" human brain.

That doesn't sound like it passes the adequately-friendly-to-wedrifid test. I'd thermite such an AI rather than press the on button.

Comment author: Zetetic 26 June 2011 09:37:05PM -1 points [-]

That seems difficult to determine without unpacking what I mean by "averaged"; how did you come to that conclusion? (I'm wondering if you have a clearer concept of what it would mean to "average" over brains, my intuition is nothing more than a sort of vague feeling that makes me think that I should study certain topics and ask certain questions) I don't even know enough to accurately convey my intuition behind my use of "averaged", I was just hoping that it might elicit a response from someone that would give me some useful information that would, in turn, help me in my task of understanding CEV better.

Comment author: wedrifid 26 June 2011 09:45:50PM 1 point [-]

Unfortunately nobody has a good answer to your question. At least none that would satisfy me. The aggregation problem is the hardest part of the problem and hasn't been answered adequately yet. Without such an answer CEV remains basically undefined. (This is something that troubles me.)

Comment author: Zetetic 26 June 2011 10:39:16PM 0 points [-]

I see. Thanks; that is enlightening. I had gotten the impression that this was the case, but wasn't sure.

Comment author: Miller 26 June 2011 06:01:46PM *  1 point [-]

Interesting. How does the program determine hard questions (and their answers) without qualifying as generating them itself? That is, the part about enumerating other programs seems superfluous.

[Edit added seconds later] Ok, I see perhaps it could ask something like 'predict what's gonna happen 10 seconds from now', and then wait to see if the prediction is correct after the universe runs for that long. In that case, the parent program isn't computing the answer itself.

Comment author: cousin_it 26 June 2011 06:20:23PM *  1 point [-]

You don't need to be able to generate the answer to your hard problem yourself, only to check that the superintelligence's offered answer is correct. These two abilities are equivalent if computing resources are unlimited, but you could run the superintelligence for a limited time... This line of thought seems to lead into the jungle of complexity theory and you should probably take my comments with a grain of salt :-)

Comment author: gwern 26 June 2011 04:55:01PM 3 points [-]

Legg's 2006 "Is There an Elegant Universal Theory of Prediction?" may be relevant.

(The answer, BTW, is 'no'; seems to be in the usual Godelian limit vein of thought: "In this paper, it is shown that although highly powerful algorithms exist, they are necessarily highly complex.")

Comment author: timtyler 27 June 2011 06:36:21AM 0 points [-]

The paper seems not very quantative. It is not obvious from it whether a human needs a thousand bits, a million bits, a trillion bits - or whatever.

Comment author: gwern 27 June 2011 01:17:15PM 0 points [-]

It would be quite impressive if it were able to...

My point was that Legg has shown, as I understand it, that any powerful prediction algorithm which is powerful enough to predict most/all of the universe (as one would expect a fearsome AGI to able to do) will be at least as complex as the universe it's predicting.

Comment author: timtyler 26 June 2011 12:22:03PM *  3 points [-]

Some entertain the hypothesis that the K-complexity of the entire universe may be under 1000 bits - though nobody knows if that is true or not.

A few bacteria will have created the explosion that leads to any future superintelligence. Much of the resulting complexity came not from them, but from their environment, though. Also, that took a while. It seems as though time limits would be needed if you are looking for a different kind of answer.

Comment author: wedrifid 26 June 2011 09:42:26PM 5 points [-]

Some entertain the hypothesis that the K-complexity of the entire universe may be under 1000 bits - though nobody knows if that is true or not.

More can be said about a single apple than about all the apples in the world.

Handing me an entire universe and saying "here you go, there is a superintelligence in there somewhere, find it yourself" does not qualify as giving me a superintelligence (in the information sense). In fact I need the same amount of information to find a superintelligence in the "1,000 bit universe" as the minimum K-complexity of the superintelligence itself.

The K-complexity of the entire universe is basically irrelevant.

Comment author: timtyler 27 June 2011 06:32:04AM *  1 point [-]

In fact I need the same amount of information to find a superintelligence in the "1,000 bit universe" as the minimum K-complexity of the superintelligence itself.

Well, perhaps minus 1,000 bits. Those 1,000 bits might be screening off a lot of universes with no superintelligences in them, so they could matter a great deal. If so, that's a smaller space to search - by a factor of 2^1,000.

Comment author: paulfchristiano 27 June 2011 01:33:27AM 0 points [-]

A description of a superintelligence might be quite easy to find in our universe; for example, if the entire universe were tiled with AIs for most of its lifetime, it may not require much information to point to one of them once you have described the universe. If the Kolmogorov complexity of the universe were really only 1000, this could easily beat cousn_it's suggestion.

Comment author: wedrifid 27 June 2011 02:45:10AM *  -1 points [-]

This is an upper bound not a lower bound. That roughly means that whenever you don't have proof to the contrary you assume the worst.

Comment author: paulfchristiano 27 June 2011 03:43:04AM 0 points [-]

What? Cousin_it gave an upper bound. I (and timtyler) are pointing out a plausible way the value could be significantly below that upper bound.

Comment author: wedrifid 27 June 2011 09:37:26AM 0 points [-]

It is true that the grandparent does not fit this context. Mistake!

Comment author: gjm 27 June 2011 02:55:27PM 2 points [-]

It seems likely that the answer depends on how rapidly and how reliably you want your "seed" to be able to turn into an actual working superintelligence. For instance:

Something along the lines of AIXI might have a very short program indeed, but not do anything useful unless you gave it more computing power and time than the entire universe can offer. (It is AIUI an open question whether that's actually finite. Let's suppose it is.)

A lots-of-humans simulator might well produce a superintelligence but its inhabitants might instead wipe themselves out in a (simulated) nuclear war, or just turn out not to be smart enough to make a working AI, or something.

Comment author: Manfred 26 June 2011 02:32:00PM 0 points [-]

If what you really want is a program that can run on a real supercomputer and reach super-human intelligence in less than a month, that's not just measuring its complexity. But I'd guess it could fit on a floppy disk.