nshepperd comments on Siren worlds and the perils of over-optimised search - Less Wrong

27 Post author: Stuart_Armstrong 07 April 2014 11:00AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (411)

You are viewing a single comment's thread. Show more comments above.

Comment author: nshepperd 15 June 2014 04:11:05AM 4 points [-]

https://github.com/moridinamael/mc-aixi

We won't get a chance to test the "planet's worth of computing power" hypothesis directly, since none of us have access to that much computing power. But, from my own experience implementing mc-aixi-ctw, I suspect that is an underestimate of the amount of compute power required.

The main problem is that the sequence prediction algorithm (CTW) makes inefficient use of sense data by "prioritizing" the most recent bits of the observation string, so only weakly makes connections between bits that are temporally separated by a lot of noise. Secondarily, plain monte carlo tree search is not well-suited to decision making in huge action spaces, because it wants to think about each action at least once. But that can most likely be addressed by reusing sequence prediction to reduce the "size" of the action space by chunking actions into functional units.

Unfortunately. both of these problems are only really technical ones, so it's always possible that some academic will figure out a better sequence predictor, lifting mc-aixi on an average laptop from "wins at pacman" to "wins at robot wars" which is about the level at which it may start posing a threat to human safety.

Comment author: V_V 16 June 2014 01:49:41PM *  1 point [-]

Unfortunately. both of these problems are only really technical ones

only?

lifting mc-aixi on an average laptop from "wins at pacman" to "wins at robot wars" which is about the level at which it may start posing a threat to human safety.

Mc-aixi is not going to win at something as open ended as robot wars just by replacing CTW or CTS with something better.
And anyway, even if it did, it wouldn't be about the level at which it may start posing a threat to human safety. Do you think that the human robot wars champions a threat to human safety? Are they even at the level of taking over the world? I don't think so.

Comment author: Cyan 16 June 2014 02:08:58PM 0 points [-]

Do you think that the human robot wars champions a threat to human safety? Are they even at the level of taking over the world?

One such champion isn't much of a threat, but only because human brains aren't copy-able.

Comment author: V_V 16 June 2014 02:15:38PM 0 points [-]

And if they were?

Comment author: Cyan 16 June 2014 02:50:30PM 0 points [-]

The question of what would happen if human brains were copy-able seems like a tangent from the discussion at hand, viz, what would happen if an there existed an AI that was capable of winning Robot Wars while running on a laptop.

Comment author: nshepperd 16 June 2014 02:29:13PM -1 points [-]

When I said a threat to human safety, I meant it literally. A robot wars champion won't take over the world (probably) but it can certainly hurt people, and will generally have no moral compunctions about doing so (only hopefully sufficient anti-harm conditioning, if its programmers thought that far ahead).

Comment author: V_V 16 June 2014 04:32:25PM *  2 points [-]

Ah yes, but in this sense, cars, trains, knives, etc., also can certainly hurt people, and will generally have no moral compunctions about doing so.
What's special about robot wars-winning AIs?

Comment author: Cyan 16 June 2014 04:58:52PM 0 points [-]

What's special about robot wars-winning AIs?

Domain-general intelligence, presumably.

Comment author: private_messaging 17 June 2014 08:15:40PM *  1 point [-]

Most basic pathfinding plus being a spinner (Hypnodisk-style) = win vs most non spinners.

Comment author: Cyan 17 June 2014 08:43:49PM -1 points [-]

I took "winning at Robot Wars" to include the task of designing the robot that competes. Perhaps nshepperd only meant piloting, though...

Comment author: private_messaging 18 June 2014 11:42:33AM *  1 point [-]

Well, we're awfully far from that. Automated programming is complete crap, automatic engineering is quite cool but its practical tools, it's not a power fantasy where you make some simple software with surprisingly little effort and then it does it all for you.

Comment author: Cyan 18 June 2014 02:21:55PM *  0 points [-]

You call it a "power fantasy" -- it's actually more of a nightmare fantasy.

Comment author: private_messaging 19 June 2014 10:23:46PM 2 points [-]

Well, historically, first, certain someone had a simple power fantasy: come up with AI somehow and then it'll just do everything. Then there was a heroic power fantasy: the others (who actually wrote some useful software and thus generally had easier time getting funding than our fantasist) are actually villains about to kill everyone, and our fantasist would save the world.

Comment author: Lumifer 16 June 2014 03:36:24PM 1 point [-]

When I said a threat to human safety, I meant it literally. A robot wars champion won't take over the world (probably) but it can certainly hurt people, and will generally have no moral compunctions about doing so

What's the difference from, say, a car assembly line robot?

Comment author: [deleted] 16 June 2014 03:55:56PM 0 points [-]

Car assembly robots have a pre-programmed routine they strictly follow. They have no learning algorithms, and usually no decision-making algorithms either. Different programs do different things!

Comment author: Lumifer 16 June 2014 04:11:25PM *  0 points [-]

Hey, look what's in the news today. I have a feeling you underappreciate the sophistication of industrial robots.

However what made me a bit confused in the grandparent post is the stress on the physical ability to harm people. As I see it, anything that can affect the physical world has the ability to harm people. So what's special about, say, robot-wars bots?

Comment author: [deleted] 16 June 2014 05:18:28PM 0 points [-]

However what made me a bit confused in the grandparent post is the stress on the physical ability to harm people. As I see it, anything that can affect the physical world has the ability to harm people. So what's special about, say, robot-wars bots?

Oh, ok. I see your point there.

Hey, look what's in the news today. I have a feeling you underappreciate the sophistication of industrial robots.

I probably do, but I still think it's worth emphasizing the particular properties of particular algorithms rather than letting people form models in their heads that say Certain Programs Are Magic And Will Do Magic Things.

Comment author: Lumifer 16 June 2014 05:35:39PM 1 point [-]

rather than letting people form models in their heads that say Certain Programs Are Magic And Will Do Magic Things.

<shrug> looks to me like a straightforward consequence of the Clarke's Third Law :-)

As an aside, I don't expect attempts to let or not let people form models in their heads to be successful :-/

Comment author: nshepperd 17 June 2014 07:26:01AM *  0 points [-]

Notice the lack of domain-general intelligence in that robot, and—on the other side—all the pre-programmed safety features it has that a mc-aixi robot would lack. Narrow AI is naturally a lot easier to reason about and build safety into. What I'm trying to stress here is the physical ability to harm people, combined with the domain-general intelligence to do it on purpose*, in the face of attempts to stop it or escape.

Different programs indeed do different things.

* (Where "purpose" includes "what the robot thought would be useful" but does not necessarily include "what the designers intended it to do".)

Comment author: TheAncientGeek 17 June 2014 09:38:14AM 0 points [-]

Nobody has bothered putting safety features into AIXI because it is so constrained by resources, but if you wanted to, it's eminently boxable.