Comment author: djm 16 July 2015 11:48:16PM 0 points [-]

This is the fundamental problem that is being researched - the top layer of abstraction would be that difficult to define one called "Be Friendly".

Instead of friendly AI maybe we should look at "dont be an asshole" AI (DBAAAI) - this may be simpler to test and monitor.

Comment author: Apteris 17 July 2015 08:45:25PM 1 point [-]

Let me clarify why I asked. I think the "multiple layers of abstraction" idea is essentially "build in a lot of 'manual' checks that the AI isn't misbehaving", and I don't think that is a desirable or even possible solution. You can write n layer of checks, but how do you know that you don't need n+1?

The idea being--as has been pointed out here on LW--that what you really want and need is a mathematical model of morality, which the AI will implement and which moral behaviour will fall out of without you having to specify it explicitly. This is what MIRI are working on with CEV & co.

Whether or not CEV or whatever emerges as the best model to use are gameable is itself a mathematical question,[1] central to the FAI problem.

[1] There are also implementation details to consider, e.g. "can I mess with the substrate" or "can I trust my substrate".

Comment author: djm 16 July 2015 02:42:20PM 4 points [-]

These are all task specific problem definition issues that occurred while fine tuning algorithms (but yes they do show how things could get out of hand)

Humans already do this very well, for example tax loopholes that are exploited but are not in the 'spirit of the law'.

The ideal (but incredibly difficult) solution would be for AI's to have multiple layers of abstraction, where each decision gets passed up and is then evaluated as "is this really what they wanted", or "am I just gaming the system".

Comment author: Apteris 16 July 2015 09:51:43PM 2 points [-]

What happens if an AI manages to game the system despite the n layers of abstraction?

Comment author: Luke_A_Somers 12 November 2014 06:06:00PM *  0 points [-]

Oh, THAT 'likely'. I thought you meant the one in the grandparent.

I stand by it, and will double down. It seems farcical that a self-improving intelligence that's at least as smart as a human (else why would it self improve rather than let us do it) would self-improve in such a way as to change its goals. That wouldn't fulfill its goals, would it, so why would it take such a 'self-improvement'? That would be a self-screwing-over instead.

If I want X, and I'm considering an improvement to my systems that would make me not want X, then I'm not going to get X if I take that improvement, so I'm going to look for some other improvement to my systems to try instead.

Eliezer's arguments for this seem pretty strong to me. Do you want to point out some flaw, or are you satisfied with saying there's no reason for it?

(ETA: I appear to be incorrect above. Eliezer was principally concerned with self-improving intelligences that are stable because those that aren't would most likely turn into those that are, eventually)

Comment author: Apteris 12 November 2014 08:22:38PM *  0 points [-]

Your argument would be stronger if you provided a citation. I've only skimmed CEV, for instance, so I'm not fully familiar with Eliezer strongest arguments in favour of goal structure tending to be preserved (though I know he did argue for that) in the course of intelligence growth. For that matter, I'm not sure what your arguments for goal stability under intelligence improvement are. Nevertheless, consider the following:

In poetic terms, our coherent extrapolated volition is our wish if we knew more, thought faster, were more the people we wished we were, had grown up farther together; where the extrapolation converges rather than diverges, where our wishes cohere rather than interfere; extrapolated as we wish that extrapolated, interpreted as we wish that interpreted.

Yudkowsky, E. (2004). Coherent Extrapolated Volition. Singularity Institute for Artificial Intelligence

(Bold mine.) See that bolded part above? Those are TODOs. They would be good to have, but they're not guaranteed. The goals of a more intelligent AI might diverge from those of its previous self; it may extrapolate differently; it may interpret differently; its desires may, at higher levels of intelligence, interfere with ours rather than cohere.

If I want X, and I'm considering an improvement to my systems that would make me not want X, then I'm not going to get X if I take that improvement, so I'm going to look for some other improvement to my systems to try instead.

A more intelligent AI might:

  • find a new way to fulfill its goals, e.g. Eliezer's example of distancing your grandmother from the fire by detonating a nuke under her;
  • discover a new thing it could do, compatible with its goal structure, that it did not see before, and that, if you're unlucky, takes priority over the other things it could be doing, e.g. you tell it "save the seals" and it starts exterminating orcas; see also Lumifer's post.
  • just decide to do things on its own. This is merely a suspicion I have, call it a mind projection, but: I think it will be challenging to design an intelligent agent with no "mind of its own", metaphorically speaking. We might succeed in that, we might not.
Comment author: KatjaGrace 04 November 2014 02:25:30AM *  3 points [-]

Bostrom summarized (p91):

We are a successful species. The reason for our success is slightly expanded mental faculties compared with other species, allowing better cultural transmission. Thus suggests that substantially greater intelligence would bring extreme power.

Our general intelligence isn't obviously the source of this improved cultural transmission. Why suppose general intelligence is the key thing, instead of improvements specific to storing and communicating information? Doesn't the observation that our cultural transmission abilities made us powerful much more strongly suggest that improved abilities to transmit culture would be very powerful? e.g. more bandwidth, better languages, better storage, better retrieval of relevant facts. It's true that AI may well have these things, but we have mostly been talking as if individual mental skills will be the important innovation.

Comment author: Apteris 04 November 2014 05:34:19PM 0 points [-]

We might be approaching a point of diminishing returns as far as improving cultural transmission is concerned. Sure, it would be useful to adopt a better language, e.g. one less ambiguous, less subject to misinterpretation, more revealing of hidden premises and assumptions. More bandwidth and better information retrieval would also help. But I don't think these constraints are what's holding AI back.

Bandwidth, storage, and retrieval can be looked at as hardware issues, and performance in these areas improves both with time and with adding more hardware. What AI requires are improvements in algorithms and in theoretical frameworks such as decision theory, morality, and systems design.

Comment author: KatjaGrace 04 November 2014 02:06:02AM 1 point [-]

Do you find the 'mail-ordered DNA scenario' plausible? (p98)

Comment author: Apteris 04 November 2014 05:15:16PM 2 points [-]

I think it will prove computationally very expensive, both to solve protein folding and to subsequently design a bootstrapping automaton. It might be difficult enough for another method of assembly to come out ahead cost-wise.

Comment author: Lumifer 28 October 2014 03:22:55PM 2 points [-]

consider the case of an AI competing with a human to devise a progressively better high-frequency trading strategy.

A more realistic example would be "competing with a human teamed up with a narrow AI".

Comment author: Apteris 28 October 2014 04:36:41PM 1 point [-]

You're right, that is more realistic. Even so, I get the feeling that the human would have less and less to do as time goes on. I quote:

“He just loaded up on value stocks,” says Mr. Fleiss, referring to the AI program. The fund gained 41% in 2009, more than doubling the Dow’s 19% gain.

As another data point, a recent chess contest between a chess grandmaster (Daniel Naroditsky) working together with an older AI (Rybka, rated ~3050) and the current best chess AI (Stockfish 5, rated 3290) ended with a 3.5 - 0.5 win for Stockfish.

Comment author: KatjaGrace 28 October 2014 01:28:10AM 4 points [-]

If someone has some money, they can invest it to get more money. Do you know what the difference is between money and intelligence that makes it plausible to expect an abrupt intelligence explosion, but reasonable to expect steady exponential growth for financial investment returns?

Comment author: Apteris 28 October 2014 12:40:32PM *  1 point [-]

While not exactly investment, consider the case of an AI competing with a human to devise a progressively better high-frequency trading strategy. An AI would probably:

  • be able to bear more things in mind at one time than the human
  • evaluate outcomes faster than the human
  • be able to iterate on its strategies faster than the human

I expect the AI's superior capacity to "drink from the fire hose" together with its faster response time to yield a higher exponent for the growth function than that resulting from the human's iterative improvement.

Comment author: KatjaGrace 21 October 2014 10:55:09PM *  1 point [-]

What empirical evidence could you look at to better predict the future winner of the Foom Debate? (for those who looked at it above)

Comment author: Apteris 27 October 2014 01:31:53PM *  0 points [-]

The effectiveness of learning hyper-heuristics for other problems, i.e. how much better algorithmically-produced algorithms perform than human-produced algorithms, and more pertinently, where the performance differential (if any) is heading.

As an example, Effective learning hyper-heuristics for the course timetabling problem says: "The dynamic scheme statistically outperforms the static counterpart, and produces competitive results when compared to the state-of-the-art, even producing a new best-known solution. Importantly, our study illustrates that algorithms with increased autonomy and generality can outperform human designed problem-specific algorithms."

Similar results can be found for other problems, bin packing, traveling salesman, and vehicle routing being just some off-the-top-of-my-head examples.

Comment author: [deleted] 13 December 2012 02:52:32AM *  6 points [-]

I don't know if you'll be able to translate to SEK, but here's my canadian dollar budget:

3000/month income after tax
-100/month food
-400/month housing
-300/month personal spending

The rest (2200) is for savings and SI (not that I've organized a monthly $1k yet or anything).

$100 for food: people are consistently amazed at this one. Oatmeal + milk + granola for breakfast. Eggs + english muffins + cheese + mayonaise + celery + peanut butter + carrots + leftovers for lunch. Cheap meat and veggies and rice and such for dinner. I shop at the local grocer for meat and veggies, and Real Canadian Superstore for everything else.

The trick is to be strict about it. Put your money in a box at the begin of the month, eat fucking beans and rice for a week if you blow the budget. You learn quick this way. Only problem is cooking. Eats up like 4 hours a week.

$400 for housing: live with roommates, and rent.

$300/mo personal: that's actually a lot of money, but you do have to be careful, you can't be buying a new jacket every month, or you won't be able to buy anything else. Again, strict budgeting.

I hope this helps people become more effective altruists!

In response to comment by [deleted] on Help Fund Lukeprog at SIAI
Comment author: Apteris 23 December 2012 02:47:22PM *  1 point [-]

Only problem is cooking. Eats up like 4 hours a week.

This article by Roger Ebert on cooking is, I suspect, highly relevant to your interests. Mine too, as a matter of fact.

Comment author: JoshuaZ 03 December 2012 01:08:27AM *  0 points [-]

I think it would be difficult to construct an ethical system where you give ''no'' consideration to cognitive capacity.

This is likely more a problem of insufficient imagination. For example, consider a system that takes seriously the idea of souls. One might very well decide that all that matters is whether an entity has a soul, completely separate from its apparent intelligence level. Similarly, a sufficiently racist individual might assign no moral weight to people of some specific racial group, regardless of their intelligence.

The comment was likely downvoted because these issues have been discussed here extensively, and there's the additional problem that I pointed out that it wouldn't even necessarily be in humanity's best interest for the entity to have such an ethical system.

Comment author: Apteris 03 December 2012 04:42:45AM 2 points [-]

For example, consider a system that takes seriously the idea of souls. One might very well decide that all that matters is whether an entity has a soul, completely separate from its apparent intelligence level. Similarly, a sufficiently racist individual might assign no moral weight to people of some specific racial group, regardless of their intelligence.

Right you are. I did not express myself well above. Let me try and restate, just for the record.

Assuming one does not assign equal rights to all autonomous agents (for instance, if we take the position that a human has more rights than a bacterium), then discriminating based on cognitive capacity (of the species, not the individual) (as one of many possible criteria) is not ipso facto wrong. It may be wrong some of the time, and it may be an approach employed by bigots, but it is not always wrong. This is my present opinion, you understand, not established fact.

there's the additional problem that I pointed out that it wouldn't even necessarily be in humanity's best interest for the entity to have such an ethical system.

Agreed. But this whole business of "we don't want the superintelligence to burn us with its magnifying glass, so we in turn won't burn ants with our magnifying glass" strikes me as rather intractable. Even though, of course, it's essential work.

I would say a few more words, but I think it's best to stop here. This subthread has cost me 66% of my Karma. :)

View more: Next