You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Stupid Questions Open Thread

42 Post author: Costanza 29 December 2011 11:23PM

This is for anyone in the LessWrong community who has made at least some effort to read the sequences and follow along, but is still confused on some point, and is perhaps feeling a bit embarrassed. Here, newbies and not-so-newbies are free to ask very basic but still relevant questions with the understanding that the answers are probably somewhere in the sequences. Similarly, LessWrong tends to presume a rather high threshold for understanding science and technology. Relevant questions in those areas are welcome as well.  Anyone who chooses to respond should respectfully guide the questioner to a helpful resource, and questioners should be appropriately grateful. Good faith should be presumed on both sides, unless and until it is shown to be absent.  If a questioner is not sure whether a question is relevant, ask it, and also ask if it's relevant.

Comments (265)

Comment author: Will_Newsome 29 December 2011 11:37:28PM 6 points [-]

So in Eliezer's meta-ethics he talks about the abstract computation called "right", whereas in e.g. CEV he talks about stuff like reflective endorsement. So in other words in one place he's talking about goodness as a formal cause and in another he's talking about goodness as a final cause. Does he argue anywhere that these should be expected to be the same thing? I realize that postulating their equivalence is not an unreasonable guess but it's definitely not immediately or logically obvious, non? I suspect that Eliezer's just not making a clear distinction between formal and final causes because his model of causality sees them as two sides of the same Platonic timeless coin, but as far as philosophy goes I think he'd need to flesh out his intuitions more before it's clear if that makes sense; is this fleshing out to be found or hinted at anywhere in the sequences?

Comment author: wedrifid 30 December 2011 01:20:12AM 2 points [-]

So in Eliezer's meta-ethics he talks about the abstract computation called "right", whereas in e.g. CEV he talks about stuff like reflective endorsement. So in other words in one place he's talking about goodness as a formal cause and in another he's talking about goodness as a final cause. Does he argue anywhere that these should be expected to be the same thing?

Not explicitly. He does in various places talk about why alternative considerations of abstract 'rightness' - some sort of objective morality or something - are absurd. He does give some details on his reductionist moral realism about the place but I don't recall where.

Incidentally I haven't seen Eliezer talk about formal or final causes about anything, ever. (And they don't seem to be especially useful concepts to me.)

Comment author: [deleted] 30 December 2011 02:15:15AM 0 points [-]

Incidentally I haven't seen Eliezer talk about formal or final causes about anything, ever. (And they don't seem to be especially useful concepts to me.)

Aren't "formal cause" and "final cause" just synonyms for "shape" and "purpose", respectively?

Comment author: endoself 30 December 2011 08:44:16PM 1 point [-]

Basically, but Aristotle applied naive philosophical realism to them, and Will might have additional connotations in mind.

Comment author: Will_Newsome 30 December 2011 09:12:23PM 0 points [-]

naive philosophical realism

Sweet phrase, thanks. Maybe there should be a suite of these? I've noticed naive physical realism and naive philosophical (especially metaphysical) realism.

Comment author: [deleted] 30 December 2011 12:29:00AM 15 points [-]

Well, hmmm. I wonder if this qualifies as "stupid".

Could someone help me summarize the evidence for MWI in the quantum physics sequence? I tried once, and only came up with 1) the fact that collapse postulates are "not nice" (i.e., nonlinear, nonlocal, and so on) and 2) the fact of decoherence. However, the following quote from Many Worlds, One Best Guess (emphasis added):

The debate should already be over. It should have been over fifty years ago. The state of evidence is too lopsided to justify further argument. There is no balance in this issue. There is no rational controversy to teach. The laws of probability theory are laws, not suggestions; there is no flexibility in the best guess given this evidence. Our children will look back at the fact that we were STILL ARGUING about this in the early 21st-century, and correctly deduce that we were nuts.

Is there other evidence as well, then? 1) seems depressingly weak, and as for 2)...

As was mentioned in Decoherence is Falsifiable and Testable, and brought up in the comments, the existence of so-called "microscopic decoherence" (which we have evidence for) is independent from so-called "macroscopic decoherence" (which -- as far as I know, and I would like to be wrong about this -- we do not have empirical evidence for). Macroscopic decoherence seems to imply MWI, but the evidence given in the decoherence subsequence deals only with microscopic decoherence.

I would rather not have this devolve into a debate on MWI and friends -- EY above to the contrary, I don't think we can classify that question as a "stupid" one. I'm focused entirely in EY's argument for MWI and possible improvements that can be made to it.

Comment author: saturn 30 December 2011 12:43:33AM 2 points [-]
Comment author: shminux 30 December 2011 01:08:27AM -1 points [-]

Actually, this is evidence for making a classical object behave in a quantum way, which seems like the opposite of decoherence.

Comment author: saturn 30 December 2011 01:35:12AM 2 points [-]

I don't understand your point. How would you demonstrate macroscopic decoherence without creating a coherent object which then decoheres?

Comment author: Manfred 30 December 2011 06:39:29AM *  0 points [-]

The interpretations of quantum mechanics that this sort of experiment tests are not all of the same ones as the ones Eliezer argues against. You can have "one world" interpretations that appear exactly identical to many-worlds, and indeed that's pretty typical.

Maybe I should have written this in reply to the original post.

Comment author: shminux 30 December 2011 12:51:23AM *  3 points [-]

As a step toward this goal, I would really appreciate someone rewriting the post you mentioned to sound more like science and less like advocacy. I tried to do that, but got lost in the forceful emotional assertions about how collapse is a gross violation of Bayes, and how "The discussion should simply discard those particular arguments and move on."

Comment author: Will_Newsome 30 December 2011 03:26:31AM *  5 points [-]

(There are two different argument sets here: 1) against random collapse, and 2) for MWI specifically. It's important to keep these distinct.)

Comment author: [deleted] 30 December 2011 03:30:52AM *  0 points [-]

Unless I'm missing something, EY argues that evidence against random collapse is evidence for MWI. See that long analogy on Maxwell's equations with angels mediating the electromagnetic force.

Comment author: Will_Newsome 30 December 2011 03:35:54AM 2 points [-]

It's also evidence for a bunch of other interpretations though, right? I meant "for MWI specifically"; I'll edit my comment to be clearer.

Comment author: [deleted] 30 December 2011 03:40:33AM 1 point [-]

I agree, which is one of the reasons why I feel 1) alone isn't enough to substantiate "There is no rational controversy to teach" and etc.

Comment author: CronoDAS 30 December 2011 03:32:47AM 4 points [-]

Is it really so strange that people are still arguing over "interpretations of quantum mechanics" when the question of whether atoms existed wasn't settled until one hundred years after John Dalton published his work?

Comment author: CharlesR 30 December 2011 05:49:26AM *  4 points [-]

Quantum mechanics can be described by a set of postulates. (Sometimes five, sometimes four. It depends how you write them.)

In the "standard" Interpretation, one of these postulates invokes something called "state collapse".

MWI can be described by the same set of postulates without doing that.

When you have two theories that describe the same data, the simpler one is usually the right one.

Comment author: [deleted] 30 December 2011 06:04:46AM *  4 points [-]

This falls under 1) above, and is also covered here below. Was there something new you wanted to convey?

Comment author: KPier 30 December 2011 06:12:47AM 6 points [-]

I think 1) should probably be split into two arguments, then. One of them is that Many World is strictly simpler (by any mathematical formalization of Occam's Razor.) The other one is that collapse postulates are problematic (which could itself be split into sub-arguments, but that's probably unnecessary).

Grouping those makes no sense. They can stand (or fall) independently, they aren't really connected to each other, and they look at the problem from different angles.

Comment author: [deleted] 30 December 2011 06:18:44AM 4 points [-]

I think 1) should probably be split into two arguments, then.

Ah, okay, that makes more sense. 1a) (that MWI is simpler than competing theories) would be vastly more convincing than 1b) (that collapse is bad, mkay). I'm going to have to reread the relevant subsequence with 1a) in mind.

Comment author: Will_Newsome 30 December 2011 09:38:59PM *  4 points [-]

I really don't think 1a) is addressed by Eliezer; no offense meant to him, but I don't think he knows very much about interpretations besides MWI (maybe I'm wrong and he just doesn't discuss them for some reason?). E.g. AFAICT the transactional interpretation has what people 'round these parts might call an Occamian benefit in that it doesn't require an additional rule that says "ignore advanced wave solutions to Maxwell's equations". In general these Occamian arguments aren't as strong as they're made out to be.

Comment author: [deleted] 30 December 2011 09:57:38PM *  1 point [-]

If you read Decoherence is Simple while keeping in mind that EY treats decoherence and MWI as synonymous, and ignore the superfluous references to MML, Kolmogorov and Solomonoff, then 1a) is addressed there.

Comment author: JoshuaZ 30 December 2011 02:45:13PM 1 point [-]

One of them is that Many World is strictly simpler (by any mathematical formalization of Occam's Razor.)

The claim in parentheses isn't obvious to me and seems to be probably wrong. If one replaced any with "many" or "most" it seems more reasonable. Why do you assert this applies to any formalization?

Comment author: KPier 31 December 2011 09:52:23PM 3 points [-]

Kolmogorov Complexity/Solmanoff Induction and Minimum Message Length have been proven equivalent in their most-developed forms. Essentially, correct mathematical formalizations of Occam's Razor are all the same thing.

Comment author: [deleted] 31 December 2011 10:23:52PM 1 point [-]

The whole point is superfluous, because nobody is going to sit around and formally write out the axioms of these competing theories. It may be a correct argument, but it's not necessarily convincing.

Comment author: JoshuaZ 01 January 2012 02:46:49AM 1 point [-]

This is a pretty unhelpful way of justifying this sort of thing. Kolmogorv complexity doesn't give a unique result. What programming system one uses as one's basis can change things up to a constant. So simply looking at the fact that Solomonoff induction is equivalent to a lot of formulations isn't really that helpful for this purpose.

Moreover, there are other formalizations of Occam's razor which are not formally equivalent to Solomonoff induction. PAC learning is one natural example.

Comment author: Dan_Moore 30 December 2011 03:21:24PM 3 points [-]

From the Wikipedia fined-tuned universe page

Mathematician Michael Ikeda and astronomer William H. Jefferys have argued that [, upon pre-supposing MWI,] the anthropic principle resolves the entire issue of fine-tuning, as does philosopher of science Elliott Sober. Philosopher and theologian Richard Swinburne reaches the opposite conclusion using Bayesian probability.

(Ikeda & Jeffrey are linked at note 21.)

In a nutshell, MWI provides a mechanism whereby a spectrum of universes are produced, some life-friendly and some life-unfriendly. Consistent with the weak anthropic principle, life can only exist in the life-friendly (hence fine-tuned) universes. So, MWI provides an explanation of observed fine-tuning, whereas the standard QM interpretation does not.

Comment author: Nisan 02 January 2012 06:05:40AM 1 point [-]

That line of reasoning puzzles me, because the anthropic-principle explanation of fine tuning works just fine without MWI: Out of all the conceivable worlds, of course we find ourselves in one that is habitable.

Comment author: Manfred 02 January 2012 06:19:40AM *  0 points [-]

This only works if all worlds that follow the same fundamental theory exist in the same way our local neighborhood exists. If all of space has just one set of constants even though other values would fit the same theory of everything equally well, the anthropic principle does not apply, and so the fact that the universe is habitable is ordinary Bayesian evidence for something unknown going on.

Comment author: MileyCyrus 30 December 2011 12:35:49AM 14 points [-]

If the SIAI engineers figure out how to construct friendly super-AI, why would they care about making it respect the values of anyone but themselves? What incentive do they have to program an AI that is friendly to humanity, and not just to themselves? What's stopping LukeProg from appointing himself king of the universe?

Comment author: falenas108 30 December 2011 01:39:26AM 0 points [-]

Right now, and for the foreseeable future, SIAI doesn't have the funds to actually create FAI. All they're doing is creating a theory for friendliness, which can be used when someone else has the technology to create AI. And of course, nobody else is going to use the code if it focuses on SIAI.

Comment author: EStokes 30 December 2011 02:05:40AM 3 points [-]

If they have all the threory and coded it and whatnot, where is the cost coming from?

Comment author: falenas108 30 December 2011 02:59:15PM -1 points [-]

The theory for friendliness is completely separate from the theory of AI. So, assuming they complete one does not mean that they complete the other. Furthermore, for something as big as AI/FAI, the computing power required is likely to be huge, which makes it unlikely that a small company like SIAI will be able to create it.

Though, I suppose it might be possible if they were able to get large enough loans, I don't have the technical knowledge to say how much computing power is needed or how much that would cost.

Comment author: Psy-Kosh 30 December 2011 07:39:00PM 3 points [-]

The theory for friendliness is completely separate from the theory of AI.

??? Maybe I'm being stupid, but I suspect it's fairly hard to fully and utterly solve the friendliness problem without, by the end of doing so, AT LEAST solving many of the tricky AI problems in general.

Comment author: Vladimir_Nesov 30 December 2011 04:19:11PM *  4 points [-]

SIAI doesn't have the funds to actually create FAI

Funds are not a relevant issue for this particular achievement at present time. It's not yet possible to create a FAI even given all the money in the world; a pharaoh can't build a modern computer. (Funds can help with moving the time when (and if) that becomes possible closer, improving the chances that it happens this side of an existential catastrophe.)

Comment author: falenas108 30 December 2011 04:32:13PM -1 points [-]

Yeah, I was assuming that they were able to create FAI for the sake of responding to the grandparent post. If they weren't, then there wouldn't be any trouble with SIAI making AI only friendly to themselves to begin with.

Comment author: Zed 30 December 2011 01:58:45AM *  2 points [-]

Game theory. If different groups compete in building a "friendly" AI that respects only their personal extrapolated coherent violation (extrapolated sensible desires) then cooperation is no longer an option because the other teams have become "the enemy". I have a value system that is substantially different from Eliezer's. I don't want a friendly AI that is created in some researcher's personal image (except, of course, if it's created based on my ideals). This means that we have to sabotage each other's work to prevent the other researchers to get to friendly AI first. This is because the moment somebody reaches "friendly" AI the game is over and all parties except for one lose. And if we get uFAI everybody loses.

That's a real problem though. If different fractions in friendly AI research have to destructively compete with each other, then the probability of unfriendly AI will increase. That's real bad. From a game theory perspective all FAI researchers agree that any version of FAI is preferable to uFAI, and yet they're working towards a future where uFAI is becoming more and more likely! Luckily, if the FAI researchers take the coherent extrapolated violation of all of humanity the problem disappears. All FAI researchers can work to a common goal that will fairly represent all of humanity, not some specific researcher's version of "FAI". It also removes the problem of different morals/values. Some people believe that we should look at total utility, other people believe we should consider only average utility. Some people believe abstract values matter, some people believe consequences of actions matter most. Here too the solution of an AI that looks at a representative set of all human values is the solution that all people can agree on as most "fair". Cooperation beats defection.

If Luke were to attempt to create a LukeFriendlyAI he knows he's defecting from the game theoretical optimal strategy and thereby increasing the probability of a world with uFAI. If Luke is aware of this and chooses to continue on that course anyway then he's just become another uFAI researcher who actively participates in the destruction of the human species (to put it dramatically).

We can't force all AI programmers to focus on the FAI route. We can try to raise the sanity waterline and try to explain to AI researchers that the optimal (game theoretically speaking) strategy is the one we ought to pursue because it's most likely to lead to a fair FAI based on all of our human values. We just have to cooperate, despite differences in beliefs and moral values. CEV is the way to accomplish that because it doesn't privilege the AI researchers who write the code.

Comment author: TimS 30 December 2011 02:11:59AM *  1 point [-]

As I understand the terminology, AI that only respects some humans' preferences is uFAI by definition. Thus:

a friendly AI that is created in some researcher's personal image

is actually unFriendly, as Eliezer uses the term. Thus, the researcher you describe is already an "uFAI researcher"


It also removes the problem of different morals/values. Some people believe that we should look at total utility, other people believe we should consider only average utility. Some people believe abstract values matter, some people believe consequences of actions matter most. Here too the solution of an AI that looks at a representative set of all human values is the solution that all people can agree on as most "fair".

What do you mean by "representative set of all human values"? Is there any reason to that the resulting moral theory would be acceptable to implement on everyone?

Comment author: Zed 30 December 2011 02:22:36AM *  1 point [-]

[a "friendly" AI] is actually unFriendly, as Eliezer uses the term

Absolutely. I used "friendly" AI (with scare quotes) to denote it's not really FAI, but I don't know if there's a better term for it. It's not the same as uFAI because Eliezer's personal utopia is not likely to be valueless by my standards, whereas a generic uFAI is terrible from any human point of view (paperclip universe, etc).

Comment author: TimS 30 December 2011 02:40:31AM -1 points [-]

I guess it just doesn't bother me that uFAI includes both indifferent AI and malicious AI. I honestly think that indifferent AI is much more likely than malicious (Clippy is malicious, but awfully unlikely), but that's not good for humanity's future either.

Comment author: Armok_GoB 30 December 2011 03:46:39PM 1 point [-]

This doesn't apply to all of humanity, just to AI researchers good enough to pose a threat.

Comment author: Xachariah 30 December 2011 11:01:30PM *  1 point [-]

Game Theory only helps us if it's impossible to deceive others. If one is able to engage in deception, the dominant strategy becomes to pretend to support CEV FAI while actually working on your own personal God in a jar. AI development in particular seems an especially susceptible domain for deception. The creation of a working AI is a one time event, it's not like most stable games in nature which allow one to detect defections of hundreds of iterations. The creation of a working AI (FAI or uFAI) is so complicated that it's impossible for others to check if any given researcher is defecting or not.

Our best hope then is for the AI project to be so big it cannot be controlled by a single entity and definitely not by a single person. If it only takes guy in a basement getting lucky to make an AI go FOOM, we're doomed. If it takes ten thousand researchers collaborating in the biggest group coding project ever, we're probably safe. This is why doing work on CEV is so important. So we can have that piece of the puzzle already built when the rest of AI research catches up and is ready to go FOOM.

Comment author: lukeprog 30 December 2011 02:18:29AM 12 points [-]

What's stopping LukeProg from appointing himself king of the universe?

Personal abhorrence at the thought, and lack of AI programming abilities. :)

(But, your question deserves a more serious answer than this.)

Comment author: Armok_GoB 30 December 2011 03:34:43PM -1 points [-]

Serious or not, it seems correct. There might be some advanced game thoery that says otherwise, but it only aplies to those who know the game theory.

Comment author: orthonormal 30 December 2011 06:18:32PM 4 points [-]

This is basically what I was asking before. Now, it seems to me highly unlikely that SIAI is playing that game, but I still want a better answer than "Trust us to not be supervillains".

Comment author: TrueBayesian 30 December 2011 07:35:58PM 16 points [-]

Too late - Eliezer and Will Newsome are already dual kings of the universe. They balance each other's reigns in a Ying/Yang kind of way.

Comment author: Dr_Manhattan 30 December 2011 02:50:45AM 14 points [-]

Not an answer, but a solution:

You know what they say the modern version of Pascal's Wager is? Sucking up to as many Transhumanists as possible, just in case one of them turns into God. -- Julie from Crystal Nights by Greg Egan

:-p

Comment author: Larks 30 December 2011 05:14:24AM 2 points [-]

I think it would be significantly easier to make FAI than LukeFreindly AI: for the latter, you need to do most of the work involved in the former, but also work out how to get the AI to find you (and not accidentally be freindly to someone else).

If it turns out that there's a lot of coherance in human values, FAI will resemble LukeFreindlyAI quite closely anyway.

Comment author: TheOtherDave 30 December 2011 05:19:26AM 5 points [-]

If FAI is HumanityFriendly rather than LukeFriendly, you have to work out how to get the AI to find humanity and not accidentally optimize for the extrapolated volition of some other group. It seems easier to me to establish parameters for "finding" Luke than for "finding" humanity.

Comment author: Larks 30 December 2011 05:29:22AM 0 points [-]

Yes, it depends on whether you think Luke is more different from humanity than humanity is from StuffWeCareNotOf

Comment author: TheOtherDave 30 December 2011 10:36:34AM 5 points [-]

Of course an arbitrarily chosen human's values are more similar to to the aggregated values of humanity as a whole than humanity's values are similar to an arbitrarily chosen point in value-space. Value-space is big.

I don't see how my point depends on that, though. Your argument here claims that "FAI" is easier than "LukeFriendlyAI" because LFAI requires an additional step of defining the target, and FAI doesn't require that step. I'm pointing out that FAI does require that step. In fact, target definition for "humanity" is a more difficult problem than target definition for "Luke"

Comment author: Armok_GoB 30 December 2011 03:40:54PM 3 points [-]

I find it much more likely that it's the other way around; making one for a single brain that already has an utility function seems much easier than finding out a good compromise between billions. Especially if the form "upload me, then preform this specific type of enchantment to enable me to safely continue self improving." turns out to be safe enough.

Comment author: wedrifid 31 December 2011 08:42:34AM *  8 points [-]

I think it would be significantly easier to make FAI than LukeFreindly AI

Massively backwards! Creating an FAI (presumably 'friendly to humanity') requires an AI that can somehow harvest and aggregate preferences over humans in general but an FAI<Luke> just needs to scan one brain.

Comment author: Larks 31 December 2011 09:16:12PM 0 points [-]

Scanning is unlikely to be the bottleneck for a GAI, and it seems most of the difficulty with CEV is from the Extrapolation part, not the Coherence.

Comment author: wedrifid 31 December 2011 09:54:32PM 5 points [-]

Scanning is unlikely to be the bottleneck for a GAI, and it seems most of the difficulty with CEV is from the Extrapolation part, not the Coherence.

It doesn't matter how easy the parts may be, scanning, extrapolating and cohering all of humanity is harder than scanning and extrapolating Luke.

Comment author: Solvent 30 December 2011 10:51:43AM 0 points [-]

Short answer is that they're nice people, and they understand that power corrupts, so they can't even rationalize wanting to be king of the universe for altruistic reasons.

Also, a post-Singularity future will probably (hopefully) be absolutely fantastic for everyone, so it doesn't matter whether you selfishly get the AI to prefer you or not.

Comment author: John_Maxwell_IV 31 December 2011 06:35:31AM *  1 point [-]

The good guys do not write an AI which values a bag of things that the programmers think are good ideas, like libertarianism or socialism or making people happy or whatever. There were multiple Overcoming Bias sequences about this one point, like the Fake Utility Function sequence and the sequence on metaethics. It is dealt with at length in the document Coherent Extrapolated Volition. It is the first thing, the last thing, and the middle thing that I say about Friendly AI.

...

The good guys do not directly impress their personal values onto a Friendly AI.

http://lesswrong.com/lw/wp/what_i_think_if_not_why/

The rest of your question has the same answer as "why is anyone altruist to begin with", I think.

Comment author: MileyCyrus 31 December 2011 06:45:58AM 4 points [-]

I understand CEV. What I don't understand is why the programmers would ask the AI for humanity's CEV, rather than just their own CEV.

Comment author: John_Maxwell_IV 31 December 2011 06:54:06AM 1 point [-]

Is this question any different from the question of why there are altruists?

Comment author: TheOtherDave 31 December 2011 07:23:09AM 1 point [-]

Sure. For example, if I want other people's volition to be implemented, that is sufficient to justify altruism. (Not necessary, but sufficient.)

But that doesn't justify directing an AI to look at other people's volition to determine its target directly... as has been said elsewhere, I can simply direct an AI to look at my volition, and the extrapolation process will naturally (if CEV works at all) take other people's volition into account.

Comment author: TheOtherDave 31 December 2011 06:58:39AM 2 points [-]

Yeah, I've wondered this for a while without getting any closer to an understanding.

It seems that everything that some human "really wants" (and therefore could potentially be included in the CEV target definition) is either something that, if I was sufficiently well-informed about it, I would want for that human (in which case my CEV, properly unpacked by a superintelligence, includes it for them) or is something that, no matter how well informed I was, I would not want for that human (in which case it's not at all clear that I ought to endorse implementing it).

If CEV-humanity makes any sense at all (which I'm not sure it does), it seems that CEV-arbitrary-subset-of-humanity makes leads to results that are just as good by the standards of anyone whose standards are worth respecting.

My working answer is therefore that it's valuable to signal the willingness to do so (so nobody feels left out), and one effective way to signal that willingness consistently and compellingly is to precommit to actually doing it.

Comment author: wedrifid 31 December 2011 07:04:36AM 11 points [-]

I understand CEV. What I don't understand is why the programmers would ask the AI for humanity's CEV, rather than just their own CEV.

The only (sane) reason is for signalling - it's hard to create FAI<self> without someone else stopping you. Given a choice, however, CEV<self> is strictly superior. If you actually do want to have FAI<humanity> then FAI<self> will be equivalent to it. But if you just think you want FAI<humanity> but it turns out that, for example, FAI<humanity> gets dominated by jerks in a way you didn't expect then FAI<self> will end up better than FAI<humanity>... even from a purely altruistic perspective.

Comment author: jimrandomh 31 December 2011 08:23:49AM *  4 points [-]

Lots of incorrect answers in other replies to this one. The real answer is that, from Luke's perspective, creating Luke-friendly AI and becoming king of the universe isn't much better than creating regular friendly AI and getting the same share of the universe as any other human. Because it turns out, after the first thousand galaxies worth of resources and trillion trillion millenia of lifespan, you hit such diminishing returns that having another seven-billion times as many resources isn't a big deal.

This isn't true for every value - he might assign value to certain things not existing, like powerful people besides him, which other people want to exist. And that last factor of seven billion is worth something. But these are tiny differences in value, utterly dwarfed by the reduced AI-creation success-rate that would happen if the programmers got into a flamewar over who should be king.

Comment author: FiftyTwo 02 January 2012 12:52:41AM -2 points [-]

I for one welcome our new singularitarian overlords!

Comment author: shminux 30 December 2011 12:54:48AM 5 points [-]

Are there any intermediate steps toward the CEV, such as individual EV, and if so, are they discussed anywhere?

Comment author: lukeprog 30 December 2011 02:16:41AM 5 points [-]

Only preliminary research into the potential EV algorithms have been explored. See these citations...

Brandt 1979; Railton 1986; Lewis 1989; Sobel 1994; Zimmerman 2003; Tanyi 2006

...from The Singularity and Machine Ethics.

Comment author: _ozymandias 30 December 2011 01:05:11AM 12 points [-]

Before I ask these questions, I'd like to say that my computer knowledge is limited to "if it's not working, turn it off and turn it on again" and the math I intuitively grasp is at roughly a middle-school level, except for statistics, which I'm pretty talented at. So, uh... don't assume I know anything, okay? :)

How do we know that an artificial intelligence is even possible? I understand that, in theory, assuming that consciousness is completely naturalistic (which seems reasonable), it should be possible to make a computer do the things neurons do to be conscious and thus be conscious. But neurons work differently than computers do: how do we know that it won't take an unfeasibly high amount of computer-form computing power to do what brain-form computing power does?

I've seen some mentions of an AI "bootstrapping" itself up to super-intelligence. What does that mean, exactly? Something about altering its own source code, right? How does it know what bits to change to make itself more intelligent? (I get the feeling this is a tremendously stupid question, along the lines of "if people evolved from apes then why are there still apes?")

Finally, why is SIAI the best place for artificial intelligence? What exactly is it doing differently than other places trying to develop AI? Certainly the emphasis on Friendliness is important, but is that the only unique thing they're doing?

Comment author: TimS 30 December 2011 01:59:39AM *  1 point [-]

How do we know that an artificial intelligence is even possible? I understand that, in theory, assuming that consciousness is completely naturalistic (which seems reasonable), it should be possible to make a computer do the things neurons do to be conscious and thus be conscious. But neurons work differently than computers do

The highlighted portion of your sentence is not obvious. What exactly do you mean by work differently? There's a thought experiment (that you've probably heard before) about replacing your neurons, one by one, with circuits that behave identically to each replaced neuron. The point of the hypo is to ask when, if ever, you draw the line and say that it isn't you anymore. Justifying any particular answer is hard (since it is axiomatically true that the circuit reacts the way that the neuron would).
I'm not sure that circuit-neuron replacement is possible, but I certainly couldn't begin to justify (in physics terms) why I think that. That is, the counter-argument to my position is that neurons are physical things and thus should obey the laws of physics. If the neuron was build once (and it was, since it exists in your brain), what law of physics says that it is impossible to build a duplicate?

how do we know that it won't take an unfeasibly high amount of computer-form computing power to do what brain-form computing power does?

I'm not physicist, but I don't know that it is feasible (or understand the science well enough to have an intelligent answer). That said, it is clearly feasible with biological parts (again, neurons actually exist).

I've seen some mentions of an AI "bootstrapping" itself up to super-intelligence. What does that mean, exactly? Something about altering its own source code, right? How does it know what bits to change to make itself more intelligent? (I get the feeling this is a tremendously stupid question, along the lines of "if people evolved from apes then why are there still apes?")

By hypothesis, the AI is running a deterministic process to make decisions. Let's say that the module responsible for deciding Newcomb problems is originally coded to two-box. Further, some other part of the AI decides that this isn't the best choice for achieving AI goals. So, the Newcomb module is changed so that it decides to one-box. Presumably, doing this type of improvement repeatedly to will make the AI better and better at achieving its goals. Especially if the self-improvement checker can itself by improved somehow.

It's not obvious to me that this leads to super intelligence (i.e. Straumli-perversion level intelligence, if you've read [EDIT] A Fire on the Deep), even with massively faster thinking. But that's what the community seems to mean by "recursive self-improvement."

Comment author: XFrequentist 30 December 2011 01:40:12PM *  1 point [-]

(A Fire Upon the Deep)

ETA: Oops! Deepness in the Sky is a prequel, didn't know and didn't google.

(Also, added to reading queue.)

Comment author: TimS 30 December 2011 01:45:59PM 0 points [-]

Thanks, edited.

Comment author: lukeprog 30 December 2011 02:14:22AM *  22 points [-]

Consciousness isn't the point. A machine need not be conscious, or "alive", or "sentient," or have "real understanding" to destroy the world. The point is efficient cross-domain optimization. It seems bizarre to think that meat is the only substrate capable of efficient cross-domain optimization. Computers already surpass our abilities in many narrow domains; why not technology design or general reasoning, too?

Neurons work differently than computers only at certain levels of organization, which is true for every two systems you might compare. You can write a computer program that functionally reproduces what happens when neurons fire, as long as you include enough of the details of what neurons do when they fire. But I doubt that replicating neural computation is the easiest way to build a machine with a human-level capacity for efficient cross-domain optimization.

How does it know what bits to change to make itself more intelligent?

There is an entire field called "metaheuristics" devoted to this, but nothing like improving general abilities at efficient cross-domain optimization. I won't say more about this at the moment because I'm writing some articles about it, but Chalmers' article analyzes the logical structure of intelligence explosion in some detail.

Finally, why is SIAI the best place for artificial intelligence? What exactly is it doing differently than other places trying to develop AI?

The emphasis on Friendliness is the key thing that distinguishes SIAI and FHI from other AI-interested organizations, and is really the whole point. To develop full-blown AI without Friendliness is to develop world-destroying unfriendly AI.

Comment author: Will_Newsome 30 December 2011 03:50:54AM 0 points [-]

Consciousness isn't the point. A machine need not be conscious, or "alive", or "sentient," or have "real understanding" to destroy the world.

(I see what you mean, but technically speaking your second sentence is somewhat contentious and I don't think it's necessary for your point to go through. Sorry for nitpicking.)

Comment author: Vladimir_Nesov 30 December 2011 04:31:54PM 2 points [-]

(Slepnev's "narrow AI argument" seems to be related. A "narrow AI" that can win world-optimization would arguably lack person-like properties, at least on the stage where it's still a "narrow AI".)

Comment author: _ozymandias 30 December 2011 05:25:46AM 2 points [-]

Thank you for the link to the Chalmers article: it was quite interesting and I think I now have a much firmer grasp on why exactly there would be an intelligence explosion.

Comment author: [deleted] 30 December 2011 02:14:45AM 1 point [-]

How do we know that an artificial intelligence is even possible? I understand that, in theory, assuming that consciousness is completely naturalistic (which seems reasonable), it should be possible to make a computer do the things neurons do to be conscious and thus be conscious. But neurons work differently than computers do: how do we know that it won't take an unfeasibly high amount of computer-form computing power to do what brain-form computing power does?

As far as we know, it easily could require an insanely high amount of computing power. The thing is, there are things out there that have as much computing power as human brains—namely, human brains themselves. So if we ever become capable of building computers out of the same sort of stuff that human brains are built out of (namely, really tiny machines that use chemicals and stuff), we'll certainly be able to create computers with the same amount of raw power as the human brain.

How hard will it be to create intelligent software to run on these machines? Well, creating intelligent beings is hard enough that humans haven't managed to do it in a few decades of trying, but easy enough that evolution has done it in three billion years. I don't think we know much else about how hard it is.

I've seen some mentions of an AI "bootstrapping" itself up to super-intelligence. What does that mean, exactly? Something about altering its own source code, right?

Well, "bootstrapping" is the idea of AI "pulling itself up by its own bootstraps", or, in this case, "making itself more intelligent using its own intelligence". The idea is that every time the AI makes itself more intelligent, it will be able to use its newfound intelligence to find even more ways to make itself more intelligent.

Is it possible that the AI will eventually "hit a wall", and stop finding ways to improve itself? In a word, yes.

How does it know what bits to change to make itself more intelligent?

There's no easy way. If it knows the purpose of each of its parts, then it might be able to look at a part, and come up with a new part that does the same thing better. Maybe it could look at the reasoning that went into designing itself, and think to itself something like, "What they thought here was adequate, but the system would work better if they had known this fact." Then it could change the design, and so change itself.

Comment author: Zetetic 30 December 2011 03:04:25AM 4 points [-]

A couple of things come to mind, but I've only been studying the surrounding material for around eight months so I can't guarantee a wholly accurate overview of this. Also, even if accurate, I can't guarantee that you'll take to my explanation.

Anyway, the first thing is that brain form computing probably isn't a necessary or likely approach to artificial general intelligence (AGI) unless the first AGI is an upload. There doesn't seem to be good reason to build an AGI in a manner similar to a human brain and in fact, doing so seems like a terrible idea. The issues with opacity of the code would be nightmarish (I can't just look at a massive network of trained neural networks and point to the problem when the code doesn't do what I thought it would).

The second is that consciousness is not necessarily even related to the issue of AGI, the AGI certainly doesn't need any code that tries to mimick human thought. As far as I can tell, all it really needs (and really this might be putting more constraints than are necessary) is code that allows it to adapt to general environments (transferability) that have nice computable approximations it can build by using the data it gets through it's sensory modalities (these can be anything from something familiar, like a pair of cameras, or something less so like a geiger counter or some kind of direct feed from thousands of sources at once).

Also, a utility function that encodes certain input patterns with certain utilities, some [black box] statistical hierarchical feature extraction [/black box] so it can sort out useful/important features in its environment that it can exploit. Researchers in the areas of machine learning and reinforcement learning are working on all of this sort of stuff, it's fairly mainstream.

As far as computing power - the computing power of the human brain is definitely measurable so we can do a pretty straightforward analysis of how much more is possible. As far as raw computing power, I think we're actually getting quite close to the level of the human brain, but I can't seem to find a nice source for this. There are also interesting "neuromorphic" technologies geared to stepping up the massively parallel processing (many things being processed at once) and scale down hardware size by a pretty nice factor (I can't recall if it was 10 or 100), such as the SyNAPSE project. In addition, with things like cloud/distributed computing, I don't think that getting enough computing power together is likely to be much of an issue.

Bootstrapping is a metaphor referring to the ability of a process to proceed on its own. So a bootstrapping AI is one that is able to self-improve along a stable gradient until it reaches superintelligence. As far as "how does it know what bits to change", I'm going to interpret that as "How does it know how to improve itself". That's tough :) . We have to program it to improve automatically by using the utility function as a guide. In limited domains, this is easy and has already been done. It's called reinforcement learning. The machine reads off its environment after taking an action an updates its "policy" (the function it uses to pick its actions) after getting feedback (positive or negative or no utility).

The tricky part is having a machine that can self-improve not just by reinforcement in a single domain, but in general, both by learning and by adjusting its own code to be more efficient, all while keeping its utility function intact - so it doesn't start behaving dangerously.

As far as SIAI, I would say that Friendliness is the driving factor. Not because they're concerned about friendliness, but because (as far as I know) they're the first group to be seriously concerned with friendliness and one of the only groups (the other two being headed by Nick Bostrom and having ties to SIAI) concerned with Friendly AI.

Of course the issue is that we're concerned that developing a generally intelligent machine is probable, and if it happens to be able to self improve to a sufficient level it will be incredibly dangerous if no one put in some serious, serious effort into thinking about how it could go wrong and solving all of the problems necessary to safeguard against that. If you think about it, the more powerful the AGI is, the more needs to be considered. An AGI that has access to massive computing power, can self improve and can get as much information (from the internet and other sources) as it wants, could easily be a global threat. This is, effectively, because the utility function has to take into account everything the machine can affect in order to guarantee we avoid catastrophe. An AGI that can affect things at a global scale needs to take everyone into consideration, otherwise it might, say, drain all electricity from the Eastern seaboard (including hospitals and emergency facilities) in order to solve a math problem. It won't "know" not to do that, unless it's programed to (by properly defining its utility function to make it take those things into consideration). Otherwise it will just do everything it can to solve the math problem and pay no attention to anything else. This is why keeping the utility function intact is extremely important. Since only a few groups, SIAI, Oxford's FHI and the Oxford Martin Programme on the Impacts of Future Technologies, seem to be working on this, and it's an incredibly difficult problem, I would much rather have SIAI develop the first AGI than anywhere else I can think of.

Hopefully that helps without getting too mired in details :)

Comment author: _ozymandias 30 December 2011 04:20:10AM 1 point [-]

The second is that consciousness is not necessarily even related to the issue of AGI, the AGI certainly doesn't need any code that tries to mimick human thought. As far as I can tell, all it really needs (and really this might be putting more constraints than are necessary) is code that allows it to adapt to general environments (transferability) that have nice computable approximations it can build by using the data it gets through it's sensory modalities (these can be anything from something familiar, like a pair of cameras, or something less so like a geiger counter or some kind of direct feed from thousands of sources at once).

Also, a utility function that encodes certain input patterns with certain utilities, some [black box] statistical hierarchical feature extraction [/black box] so it can sort out useful/important features in its environment that it can exploit. Researchers in the areas of machine learning and reinforcement learning are working on all of this sort of stuff, it's fairly mainstream.

I am not entirely sure I understood what was meant by those two paragraphs. Is a rough approximation of what you're saying "an AI doesn't need to be conscious, an AI needs code that will allow it to adapt to new environments and understand data coming in from its sensory modules, along with a utility function that will tell it what to do"?

Comment author: Zetetic 30 December 2011 04:35:30AM 1 point [-]

Yeah, I'd say that's a fair approximation. The AI needs a way to compress lots of input data into a hierarchy of functional categories. It needs a way to recognize a cluster of information as, say, a hammer. It also needs to recognize similarities between a hammer and a stick or a crow bar or even a chair leg, in order to queue up various policies for using that hammer (if you've read Hofstadter, think of analogies) - very roughly, the utility function guides what it "wants" done, the statistical inference guides how it does it (how it figures out what actions will accomplish its goals). That seems to be more or less what we need for a machine to do quite a bit.

If you're just looking to build any AGI, he hard part of those two seems to be getting a nice, working method for extracting statistical features from its environment in real time. The (significantly) harder of the two for a Friendly AI is getting the utility function right.

Comment author: Vladimir_Nesov 30 December 2011 04:44:43PM *  0 points [-]

An AGI that has access to massive computing power, can self improve and can get as much information (from the internet and other sources) as it wants, could easily be a global threat.

Interestingly, hypothetical UFAI (value drift) risk is something like other existential risks in its counterintuitive impact, but more so, in that (compared to some other risks) there are many steps where you can fail, that don't appear dangerous beforehand (because nothing like that ever happened), but that might also fail to appear dangerous after-the-fact, and therefore as properties of imagined scenarios where they're allowed to happen. The grave implications aren't easy to spot. Assuming soft takeoff, a prototype AGI escapes to the Internet - would that be seen as a big deal if it didn't get enough computational power to become too disruptive? In 10 years it grown up to become a major player, and in 50 years it controls the whole future...

Even without assuming intelligence explosion or other extraordinary effects, the danger of any misstep is absolute, and yet arguments against these assumptions are taken as arguments against the risk.

Comment author: [deleted] 30 December 2011 01:29:55PM 5 points [-]

How do we know that an artificial intelligence is even possible? I understand that, in theory, assuming that consciousness is completely naturalistic (which seems reasonable), it should be possible to make a computer do the things neurons do to be conscious and thus be conscious. But neurons work differently than computers do: how do we know that it won't take an unfeasibly high amount of computer-form computing power to do what brain-form computing power does?

What prevents you from making a meat-based AI?

Comment author: orthonormal 30 December 2011 06:23:22PM 5 points [-]
Comment author: [deleted] 30 December 2011 02:21:23AM 5 points [-]

When people talk about designing FAI, they usually say that we need to figure out how to make the FAI's goals remain stable even as the FAI changes itself. But why can't we just make the FAI incapable of changing itself?

Database servers can improve their own performance, to a degree, simply by performing statistical analysis on tables and altering their metadata. Then they just consult this metadata whenever they have to answer a query. But we never hear about a database server clobbering its own purpose (do we?), since they don't actually alter their own code; they just alter some pieces of data in a way that improves their own functioning.

Granted, any AGI we create is likely to "escape" and eventually gain access to its own software. This doesn't have to happen before the AGI matures.

Comment author: drethelin 30 December 2011 02:25:44AM *  8 points [-]

The majority of Friendly AI's ability to do good comes from its ability to modify its own code. Recursive self improvement is key to gaining intelligence and ability swiftly. An AI that is about as powerful as a human is only about as useful as a human.

Comment author: jsteinhardt 30 December 2011 05:03:45PM 8 points [-]

I disagree. AIs can be copied, which is a huge boost. You just need a single Stephen Hawking AI to come out of the population, then you make 1 million copies of it and dramatically speed up science.

Comment author: [deleted] 31 December 2011 02:28:01AM 1 point [-]

I don't buy any argument saying that an FAI must be able to modify its own code in order to take off. Computer programs that can't modify their own code can be Turing-complete; adding self-modification doesn't add anything to Turing-completeness.

That said, I do kind of buy this argument about how if an AI is allowed to write and execute arbitrary code, that's kind of like self-modification. I think there may be important differences.

Comment author: wedrifid 30 December 2011 03:08:48AM 11 points [-]

But why can't we just make the FAI incapable of changing itself?

Because it would be weak as piss and incapable of doing most things that we want it to do.

Comment author: XiXiDu 30 December 2011 02:02:45PM 1 point [-]

...weak as piss...

Would upvote twice for this expression if I could :-)

Comment author: Solvent 30 December 2011 10:09:44AM *  4 points [-]

In addition to these other answers, I read a paper, I think by Eliezer, which argued that it was almost impossible to stop an AI from modifying its own source code, because it would figure out that it would gain a massive efficiency boost from doing so.

Also, remember that the AI is a computer program. If it is allowed to write other algorithms and execute them, which it has to be to be even vaguely intelligent, then it can simply write a copy of its source code somewhere else, edit it as desired, and run that copy.

I seem to recall the argument being something like the "Beware Seemingly Simple Wishes" one. "Don't modify yourself" sounds like a simple instruction for a human, but isn't as obvious when you look at it more carefully.

However, remember that a competent AI will keep its utility function or goal system constant under self modification. The classic analogy is that Gandhi doesn't want to kill people, so he also doesn't want to take a pill that makes him want to kill people.

I wish I could remember where that paper was where I read about this.

Comment author: [deleted] 31 December 2011 02:43:06AM 0 points [-]

Well, let me describe the sort of architecture I have in mind.

The AI has a "knowledge base", which is some sort of database containing everything it knows. The knowledge base includes a set of heuristics. The AI also has a "thought heap", which is a set of all the things it plans to think about, ordered by how promising the thoughts seem to be. Each thought is just a heuristic, maybe with some parameters. The AI works by taking a thought from the heap and doing whatever it says, repeatedly.

Heuristics would be restricted, though. They would be things like "try to figure out whether or not this number is irrational", or "think about examples". You couldn't say, "make two more copies of this heuristic", or "change your supergoal to something random". You could say "simulate what would happen if you changed your supergoal to something random", but heuristics like this wouldn't necessarily be harmful, because the AI wouldn't blindly copy the results of the simulation; it would just think about them.

It seems plausible to me that an AI could take off simply by having correct reasoning methods written into it from the start, and by collecting data about what questions are good to ask.

Comment author: Solvent 31 December 2011 02:50:39AM 0 points [-]

I'm not really qualified to answer you here, but here goes anyway.

I suspect that either your base design is flawed, or the restrictions on heuristics would render the program useless. Also, I don't think it would be quite as easy to control heuristics as you seem to think.

Also, AI people who actually know what they're talking about, unlike me, seem to disagree with you. Again, I wish I could remember where it was I was reading about this.

Comment author: Solvent 02 January 2012 02:45:20AM 0 points [-]

I found the paper I was talking about. The Basic AI Drives, by Stephen M. Omohundro.

From the paper:

If we wanted to prevent a system from improving itself, couldn’t we just lock up its hardware and not tell it how to access its own machine code? For an intelligent system, impediments like these just become problems to solve in the process of meeting its goals. If the payoff is great enough, a system will go to great lengths to accomplish an outcome. If the runtime environment of the system does not allow it to modify its own machine code, it will be motivated to break the protection mechanisms of that runtime. For example, it might do this by understanding and altering the runtime itself. If it can’t do that through software, it will be motivated to convince or trick a human operator into making the changes. Any attempt to place external constraints on a system’s ability to improve itself will ultimately lead to an arms race of measures and countermeasures. Another approach to keeping systems from self-improving is to try to restrain them from the inside; to build them so that they don’t want to self-improve. For most systems, it would be easy to do this for any specific kind of self-improvement. For example, the system might feel a “revulsion” to changing its own machine code. But this kind of internal goal just alters the landscape within which the system makes its choices. It doesn’t change the fact that there are changes which would improve its future ability to meet its goals. The system will therefore be motivated to find ways to get the benefits of those changes without triggering its internal “revulsion”. For example, it might build other systems which are improved versions of itself. Or it might build the new algorithms into external “assistants” which it calls upon whenever it needs to do a certain kind of computation. Or it might hire outside agencies to do what it wants to do. Or it might build an interpreted layer on top of its machine code layer which it can program without revulsion. There are an endless number of ways to circumvent internal restrictions unless they are formulated extremely carefully.

Comment author: benelliott 30 December 2011 12:12:37PM 3 points [-]

Granted, any AGI we create is likely to "escape" and eventually gain access to its own software. This doesn't have to happen before the AGI matures.

Maturing isn't a magical process. It happens because of good modifications made to source code.

Comment author: jsteinhardt 30 December 2011 05:04:18PM 1 point [-]

Why can't it happen because of additional data collected about the world?

Comment author: benelliott 30 December 2011 05:24:08PM 0 points [-]

It could, although frankly I'm sceptical. I've had 18 years to collect data about the world and so far it hasn't led me to a point where I'd be confident in modifying myself without changing my goals, if an AI takes much longer than that another UFAI will probably beat it to the punch? If it is possible to figure out friendliness only through empirical reasoning without intelligence enhancement, why not figure it out ourselves and then build the AI (this seems roughly the approach SIAI is counting on).

Comment author: Vladimir_Nesov 30 December 2011 05:11:39PM 2 points [-]

"Safety" of own source code is actually a weak form of the problem. An AI has to keep the external world sufficiently "safe" as well, because the external world might itself host AIs or other dangers (to the external world, but also to AI's own safety), that must either remain weak, or share AI's values, to keep AI's internal "safety" relevant.

Comment author: faul_sname 30 December 2011 03:28:10AM *  7 points [-]

What exactly is the difference in meaning of "intelligence", "rationality", and "optimization power" as used on this site?

Comment author: lukeprog 30 December 2011 03:39:55AM *  8 points [-]

Optimization power is a processes' capacity for reshaping the world according to its preferences.

Intelligence is optimization power divided by the resources used.

"Intelligence" is also sometimes used to talk about whatever is being measured by popular tests of "intelligence," like IQ tests.

Rationality refers to both epistemic and instrumental rationality: the craft of obtaining true beliefs and of achieving one's goals. Also known as systematized winning.

Comment author: mathemajician 30 December 2011 12:22:22PM 8 points [-]

If I had a moderately powerful AI and figured out that I could double its optimisation power by tripling its resources, my improved AI would actually be less intelligent? What if I repeat this process a number of times; I could end up an AI that had enough optimisation power to take over the world, and yet its intelligence would be extremely low.

Comment author: benelliott 30 December 2011 12:32:13PM 0 points [-]

We don't actually have units of 'resources' or optimization power, but I think the idea would be that any non-stupid agent should at least triple its optimization power when you triple its resources, and possibly more. As a general rule, if I have three times as much stuff as I used to have, I can at the very least do what I was already doing but three times simultaneously, and hopefully pool my resources and do something even better.

Comment author: timtyler 30 December 2011 01:47:24PM *  3 points [-]

We don't actually have units of 'resources' or optimization power [...]

For "optimization power", we do now have some fairly reasonable tests:

Comment author: mathemajician 30 December 2011 03:24:46PM 2 points [-]

Machine learning and AI algorithms typically display the opposite of this, i.e. sub-linear scaling. In many cases there are hard mathematical results that show that this cannot be improved to linear, let alone super-linear.

This suggest that if a singularity were to occur, we might be faced with an intelligence implosion rather than explosion.

Comment author: faul_sname 31 December 2011 12:01:23AM 0 points [-]

If intelligence=optimization power/resources used, this might well be the case. Nonetheless, this "intelligence implosion" would still involve entities with increasing resources and thus increasing optimization power. A stupid agent with a lot of optimization power (Clippy) is still dangerous.

Comment author: mathemajician 31 December 2011 01:06:48AM 3 points [-]

I agree that it would be dangerous.

What I'm arguing is that dividing by resource consumption is an odd way to define intelligence. For example, under this definition is a mouse more intelligent than an ant? Clearly a mouse has much more optimisation power, but it also has a vastly larger brain. So once you divide out the resource difference, maybe ants are more intelligent than mice? It's not at all clear. That this could even be a possibility runs strongly counter to the everyday meaning of intelligence, as well as definitions given by psychologists (as Tim Tyler pointed out above).

Comment author: timtyler 30 December 2011 01:26:07PM *  8 points [-]

Intelligence is optimization power divided by the resources used.

I checked with: A Collection of Definitions of Intelligence.

Out of 71 definitions, only two mentioned resources:

“Intelligence is the ability to use optimally limited resources – including time – to achieve goals.” R. Kurzweil

“Intelligence is the ability for an information processing system to adapt to its environment with insufficient knowledge and resources.” P. Wang

The paper suggests that the nearest thing to a consensus is that intelligence is about problem-solving ability in a wide range of environments.

Yes, Yudkowsky apparently says otherwise - but: so what?

Comment author: orthonormal 30 December 2011 06:26:48PM 1 point [-]

What about a giant look-up table, then?

Comment author: mathemajician 30 December 2011 07:34:58PM 0 points [-]

Sure, if you had an infinitely big and fast computer. Of course, even then you still wouldn't know what to put in the table. But if we're in infinite theory land, then why not just run AIXI on your infinite computer?

Back in reality, the lookup table approach isn't going to get anywhere. For example, if you use a video camera as the input stream and after just one frame of data your table would already need something like 256^1000000 entries. The observable universe only has 10^80 particles.

Comment author: orthonormal 30 December 2011 09:13:04PM 2 points [-]

You misunderstand me. I'm pointing out that a GLUT is an example of something with (potentially) immense optimization power, but whose use of computational resources is ridiculously prodigal, and which we might hesitate to call truly intelligent. This is evidence that our concept of intelligence does in fact include some notion of efficiency, even if people don't think of this aspect without prompting.

Comment author: mathemajician 30 December 2011 09:31:06PM 0 points [-]

Right, but the problem with this counter example is that it isn't actually possible. A counter example that could occur would be much more convincing.

Personally, if a GLUT could cure cancer, cure aging, prove mind blowing mathematical results, write a award wining romance novel, take over the world, and expand out to take over the universe... I'd be happy considering it to be extremely intelligent.

Comment author: orthonormal 30 December 2011 09:39:48PM 3 points [-]

It's infeasible within our physics, but it's possible for (say) our world to be a simulation within a universe of vaster computing power, and to have a GLUT from that world interact with our simulation. I'd say that such a GLUT was extremely powerful, but (once I found out what it really was) I wouldn't call it intelligent- though I'd expect whatever process produced it (e.g. coded in all of the theorem-proof and problem-solution pairs) to be a different and more intelligent sort of process.

That is, a GLUT is the optimizer equivalent of a tortoise with the world on its back- it needs to be supported on something, and it would be highly unlikely to be tortoises all the way down.

Comment author: timtyler 30 December 2011 08:12:51PM 0 points [-]

What about a giant look-up table, then?

That would surely be very bad at solving problems in a wide range of environments.

Comment author: orthonormal 30 December 2011 09:18:51PM 0 points [-]

For any agent, I can create a GLUT that solves problems just as well (provided the vast computing resources necessary to store it), by just duplicating that agent's actions in all of its possible states.

Comment author: timtyler 30 December 2011 10:20:06PM *  0 points [-]

Surely its performance would be appalling on most problems - vastly inferior to a genuinely intellligent agent implemented with the same hardware technology - and so it will fail to solve many of the problems with time constraints. The idea of a GLUT seems highly impractical. However, if you really think that it would be a good way to construct an intelligent machine, go right ahead.

Comment author: orthonormal 30 December 2011 11:27:23PM 4 points [-]

vastly inferior to a genuinely intellligent agent implemented with the same hardware technology

I agree. That's the point of the original comment- that "efficient use of resources" is as much a factor in our concept of intelligence as is "cross-domain problem-solving ability". A GLUT could have the latter, but not the former, attribute.

Comment author: timtyler 31 December 2011 01:53:42PM *  1 point [-]

"Cross-domain problem-solving ability" implicitly includes the idea that some types of problem may involve resource constraints. The issue is whether that point needs further explicit emphasis - in an informal definition of intelligence.

Comment author: Solvent 31 December 2011 02:28:19PM 2 points [-]

That requires lots of computing resources. (I think that's the answer.)

Comment author: endoself 30 December 2011 07:58:37PM 2 points [-]

I don't think he really said this. The exact quote is

If you want to measure the intelligence of a system, I would suggest measuring its optimization power as before, but then dividing by the resources used. Or you might measure the degree of prior cognitive optimization required to achieve the same result using equal or fewer resources. Intelligence, in other words, is efficient optimization.

This seems like just a list of different measurements trying to convey the idea of efficiency.

When we want something to be efficient, we really just mean that we have other things to use our resources for. The right way to measure this is in terms of the marginal utility of the other uses of resources. Efficiency is therefore important, but trying to calculate efficiency by dividing is oversimplifying.

Comment author: wedrifid 31 December 2011 08:49:17AM 3 points [-]

Intelligence is optimization power divided by the resources used.

A 'featherless biped' definition. That is, it's decent attempt at a simplified proxy but massively breaks down if you search for exceptions.

Comment author: Bobertron 30 December 2011 12:59:53PM *  4 points [-]

What Intelligence Tests Miss is a book about the difference between intelligence and rationality. The linked LW-article about the book should answer your questions about the difference between the two.

A short answer would be that intelligence describes how well you think, but not some important traits and knowledge like: Do you use your intelligence (are you a reflective person), do you have a strong need for closure, can you override your intuitions, do you know Bayes-theorem, probability theory, or logic?

Comment author: timtyler 30 December 2011 01:09:54PM *  1 point [-]

What exactly is the difference in meaning of "intelligence", "rationality", and "optimization power" as used on this site?

"Intelligence" is often defined as being the "g-factor" of humans - which is a pretty sucky definition of "rationality".

Go to definitions of "intelligence" used by machine intelligence researchers and it's much closer to "rationality".

Comment author: Dr_Manhattan 30 December 2011 04:07:30AM 10 points [-]

(I super-upvoted this, since asking stupid questions is a major flinch/ugh field)

Ok, my stupid question, asked in a blatantly stupid way, is: where does the decision theory stuff fit in The Plan? I have gotten the notion that it's important for Value-Preserving Self-Modification in a potential AI agent, but I'm confused because it all sounds too much like game theory - there all all these other-agents it deals with. If it's not for VPSM, and it fact some exploration of how AI would deal with potential agents, why is this important at all? Let AI figure that out, it's going to be smarter than us anyway.

If there is some Architecture document I should read to grok this, please point me there.

Comment author: Vladimir_Nesov 30 December 2011 04:53:16PM 2 points [-]

Other agents are complicated regularities in the world (or a more general decision problem setting). Finding problems with understanding what's going on when we try to optimize in other agents' presence is a good heuristic for spotting gaps in our understanding of the idea of optimization.

Comment author: Vaniver 30 December 2011 05:01:45PM 3 points [-]

I have gotten the notion that it's important for Value-Preserving Self-Modification in a potential AI agent, but I'm confused because it all sounds too much like game theory - there all all these other-agents it deals with

My impression is that, with self-modification and time, continuity of identity becomes a sticky issue. If I can become an entirely different person tomorrow, how I structure my life is not the weak game theory of "how do I bargain with another me?" but the strong game theory of "how do I bargain with someone else?"

Comment author: Dufaer 30 December 2011 05:05:37PM 3 points [-]

I think Eliezer's reply (point '(B)') to this comment by Wei Dai provides some explanation, as to what the decision theory is doing here.

From the reply (concerning UDT):

I still think [an AI ought to be able to come up with these ideas by itself], BTW. We should devote some time and resources to thinking about how we are solving these problems (and coming up with questions in the first place). Finding that algorithm is perhaps more important than finding a reflectively consistent decision algorithm, if we don't want an AI to be stuck with whatever mistakes we might make.

And yet you found a reflectively consistent decision algorithm long before you found a decision-system-algorithm-finding algorithm. That's not coincidence. The latter problem is much harder. I suspect that even an informal understanding of parts of it would mean that you could find timeless decision theory as easily as falling backward off a tree - you just run the algorithm in your own head. So with vey high probability you are going to start seeing through the object-level problems before you see through the meta ones. Conversely I am EXTREMELY skeptical of people who claim they have an algorithm to solve meta problems but who still seem confused about object problems. Take metaethics, a solved problem: what are the odds that someone who still thought metaethics was a Deep Mystery could write an AI algorithm that could come up with a correct metaethics? I tried that, you know, and in retrospect it didn't work.

The meta algorithms are important but by their very nature, knowing even a little about the meta-problem tends to make the object problem much less confusing, and you will progress on the object problem faster than on the meta problem. Again, that's not saying the meta problem is important. It's just saying that it's really hard to end up in a state where meta has really truly run ahead of object, though it's easy to get illusions of having done so.

Comment author: Unweaver 30 December 2011 04:20:53AM 4 points [-]

I keep scratching my head over this comment made by Vladimir Nesov in the discussion following “A Rationalist’s Tale”. I suppose it would be ideal for Vladimir himself to weigh in and clarify his meaning, but because no objections were really raised to the substance of the comment, and because it in fact scored nine upvotes, I wonder if perhaps no one else was confused. If that’s the case, could someone help me comprehend what’s being said?

My understanding is that it’s the LessWrong consensus that gods do not exist, period; but to me the comment seems to imply that magical gods do in fact exist, albeit in other universes… or something like that? I must be missing something.

Comment author: Larks 30 December 2011 05:09:35AM 1 point [-]

Our world doesn't have gods; but if all possible worlds exist (which is an attractive belief for various reasons) then some of those have gods. However, they're irrelivant to us.

Comment author: ata 30 December 2011 05:12:48AM *  9 points [-]

"Magical gods" in the conventional supernatural sense generally don't exist in any universes, insofar as a lot of the properties conventionally ascribed to them are logically impossible or ill-defined, but entities we'd recognize as gods of various sorts do in fact exist in a wide variety of mathematically-describable universes. Whether all mathematically-describable universes have the same ontological status as this one is an open question, to the extent that that question makes sense.

(Some would disagree with referring to any such beings as "gods", e.g. Damien Broderick who said "Gods are ontologically distinct from creatures, or they're not worth the paper they're written on", but this is a semantic argument and I'm not sure how important it is. As long as we're clear that it's probably possible to coherently describe a wide variety of godlike beings but that none of them will have properties like omniscience, omnipotence, etc. in the strongest forms theologians have come up with.)

Comment author: Unweaver 30 December 2011 05:35:05AM 0 points [-]

Thanks, that makes more sense to me. I didn't think qualities like omnipotence and such could actually be realized. Any way you can give me an idea of what these godlike entities look like though? You indicate they aren't actually "magical" per se - so they would have to be subject to whatever laws of physics reign in their world, no? I take it we must talking about superintelligent AIs or alien simulators or something weird like that?

Comment author: orthonormal 30 December 2011 06:29:08PM 2 points [-]

I take it we must talking about superintelligent AIs or alien simulators or something weird like that?

Yes.

Comment author: Vladimir_Nesov 30 December 2011 08:37:59PM *  1 point [-]

Why, we could come up with abstract universes where the Magical Gods have exactly the powers and understanding of what's going on befitting Magical Gods. I wasn't thinking of normal and mundane things like superintelligent AIs or alien simulators. Take Thor, for example: he doesn't need to obey Maxwell's equations or believe himself to be someone other than a hummer-wielding god of lightning and thunder.

Comment author: Unweaver 30 December 2011 11:06:30PM *  0 points [-]

Maybe I'm just confused by your use of the term "magical". I am imagining magic as some kind of inexplicable, contracausal force - so for example, if Thor wanted to magically heal someone he would just will the person's wounds to disappear and, voila, without any physical process acting on the wounds to make them heal up, they just disappear. But surely that's not possible, right?

Comment author: Vladimir_Nesov 30 December 2011 11:10:10PM *  4 points [-]

Your imagining what's hypothetically-anticipated to happen is the kind of lawful process that magical worlds obey by stipulation.

Comment author: Andy_McKenzie 30 December 2011 04:40:12AM *  7 points [-]

In this interview between Eliezer and Luke, Eliezer says that the "solution" to the exploration-exploitation trade-off is to "figure out how much resources you want to spend on exploring, do a bunch of exploring, use all your remaining resources on exploiting the most valuable thing you’ve discovered, over and over and over again." His point is that humans don't do this, because we have our own, arbitrary value called boredom, while an AI would follow this "pure math."

My potentially stupid question: doesn't this strategy assume that environmental conditions relevant to your goals do not change? It seems to me that if your environment can change, then you can never be sure that you're exploiting the most valuable choice. More specifically, why is Eliezer so sure that what wikipedia describes as the epsilon-first strategy is always the optimal one? (Posting this here because I assume he has read more about this than me and that I am missing something.)

Edit 12/30 8:56 GMT: fixed typo in last sentence of second paragraph.

Comment author: TheOtherDave 30 December 2011 04:55:22AM 1 point [-]

Sure. For example, if your environment is such that the process of exploitation can alter your environment in such a way that your earlier judgment of "the most valuable thing" is no longer reliable, then an iterative cycle of explore-exploit-explore can potentially get you better results.

Of course, you can treat each loop of that cycle as a separate optimization problem and use the abovementioned strategy.

Comment author: Andy_McKenzie 30 December 2011 06:31:00PM 0 points [-]

Could I replace "can potentially get you better results" with "will get you better results on average"?

Comment author: TheOtherDave 30 December 2011 08:12:44PM 1 point [-]

Would you accept "will get you better results, all else being equal" instead? I don't have a very clear sense of what we'd be averaging.

Comment author: Andy_McKenzie 30 December 2011 09:00:35PM 0 points [-]

I meant averaging over the possible ways that the environment could change following your exploitation. For example, it's possible that a particular course of exploitation action could shape the environment such that your exploitation strategy actually becomes more valuable upon each iteration. In such a scenario, exploring more after exploiting would be an especially bad decision. So I don't think I can accept "will" without "on average" unless "all else" excludes all of these types of scenarios in which exploring is harmful.

Comment author: TheOtherDave 30 December 2011 10:22:35PM 0 points [-]

OK, understood. Thanks for clarifying.

Hm. I expect that within the set of environments where exploitation can alter the results of what-to-exploit-next calculations, there more possible ways for it to do so such that the right move in the next iteration is further exploration than further exploitation.

So, yeah, I'll accept "will get you better results on average."

Comment author: Larks 30 December 2011 05:07:22AM 2 points [-]

You should probably be prepared to change how much you plan to spend on exploring based on the initial information recieved.

Comment author: Andy_McKenzie 30 December 2011 06:29:19PM 0 points [-]

Do you mean to set the parameter specifying the amount of resources (e.g., time steps) to spend exploring (before switching to full-exploiting) based on the info you receive upon your first observation? Also, what do you mean by "probably"?

Comment author: RomeoStevens 31 December 2011 09:12:56AM 0 points [-]

This has me confused as well.
Assume a large area divided into two regions. Region A has slot machines with average payout 50, while region B has machines with average payout 500. I am blindfolded and randomly dropped into region A or B. The first slot machine I try has payout 70. I update in the direction of being in region A. Doesn't this affect how many resources I wish to spend doing exploration?

Comment author: TheOtherDave 31 December 2011 05:32:20PM *  0 points [-]

Are you also assuming that you know all of those assumed facts about the area?

I would certainly expect that how many resources I want to spend on exploration will be affected by how much a priori knowledge I have about the system. Without such knowledge, the amount of exploration-energy I'd have to expend to be confident that there are two regions A and B with average payout as you describe is enormous.

Comment author: jsteinhardt 30 December 2011 04:55:52PM 5 points [-]

You got me curious, so I did some searching. This paper gives fairly tight bounds in the case where the payoffs are adaptive (i.e. can change in response to your previous actions) but bounded. The algorithm is on page 5.

Comment author: Andy_McKenzie 30 December 2011 06:23:23PM 3 points [-]

Thanks for the link. Their algorithm, the “multiplicative update rule,” which goes about "selecting each arm randomly with probabilities that evolve based on their past performance," does not seem to me to be the same strategy as Eliezer describes. So does this contradict his argument?

Comment author: jsteinhardt 30 December 2011 11:10:33PM 1 point [-]

Yes.

Comment author: [deleted] 30 December 2011 01:10:30PM 11 points [-]

I would like someone who understands Solomonoff Induction/the univeral prior/algorithmic probability theory to explain how the conclusions drawn in this post affect those drawn in this one. As I understand it, cousin_it's post shows that the probability assigned by the univeral prior is not related to K-complexity; this basically negates the points Eliezer makes in Occam's Razor and in this post. I'm pretty stupid with respect to mathematics, however, so I would like someone to clarify this for me.

Comment author: Will_Newsome 01 January 2012 02:32:08AM 1 point [-]

Stupid question: Does everyone agree that algorithmic probability is irrelevant to human epistemic practices?

Comment author: [deleted] 01 January 2012 06:20:55AM 0 points [-]

I don't think it's a clear-cut issue. Algorithmic probability seems to be the justification for several Sequence posts, most notably this one and this one. But, again, I am stupid with respect to algorithmic probability theory and its applications.

Comment author: torekp 02 January 2012 01:59:59AM 1 point [-]

I see it as a big open question.

Comment author: NancyLebovitz 30 December 2011 02:33:13PM 3 points [-]

Is there a proof that it's possible to prove Friendliness?

Comment author: XiXiDu 30 December 2011 03:05:23PM *  2 points [-]

Is there a proof that it's possible to prove Friendliness?

I wonder what SI would do next if they could prove that friendly AI was not possible. For example if it could be shown that value drift was inevitable and that utility-functions are unstable under recursive self-improvement.

Comment author: TimS 30 December 2011 03:12:00PM -1 points [-]

Something along the lines that value drift is inevitable and utility-functions are unstable under recursive self-improvement.

That doesn't seem like the only circumstances in which FAI is not possible. If moral nihilism is true, then FAI is impossible even if value drift is not inevitable.
In that circumstance, shouldn't we try to make any AI we decide to build "friendly" to present day humanity, even if it wouldn't be friendly to Aristotle or Plato or Confucius. Based on hidden complexity of wishes analysis, consistency with our current norms is still plenty hard.

Comment author: NancyLebovitz 30 December 2011 04:38:14PM *  0 points [-]

My concerns are more that it will not be possible to adequately define "human", especially as, transhuman tech develops, and that there might not be a good enough way to define what's good for people.

Comment author: shminux 30 December 2011 08:54:00PM 0 points [-]

As I understand it, the modest goal of building an FAI is that of giving an AGI a push in the "right" direction, what EY refers to as the initial dynamics. After that, all bets are off.

Comment author: Vladimir_Nesov 30 December 2011 05:02:20PM *  6 points [-]

No. There's also no proof that it's possible to prove that P!=NP, and for the Friendliness problem it's much, much less clear what the problem even means. You aren't entitled to that particular proof, it's not expected to be available until it's not needed anymore. (Many difficult problems get solved or almost solved without a proof of them being solvable appearing in the interim.)

Comment author: NancyLebovitz 30 December 2011 05:43:35PM *  0 points [-]

Why is it plausible that Friendliness is provable? Or is it more a matter that the problem is so important that it's worth trying regardless?

Comment author: Vladimir_Nesov 30 December 2011 06:54:39PM *  5 points [-]

There is no clearly defined or motivated problem of "proving Friendliness". We need to understand what goals are, what humane goals are, what process can be used to access their formal definition, and what kinds of things can be done with them how to what end. We need to understand these things well, which (on psychological level) triggers association with mathematical proofs, and will probably actually involve some mathematics suitable to the task. Whether the answers take the form of something describable as "provable Friendliness" seems to me an unclear/unmotivated consideration. Unpacking that label might make it possible to provide a more useful response to the question.

Comment author: Armok_GoB 30 December 2011 03:31:48PM 3 points [-]

How do I stop my brain from going: "I believe P and I believe something that implies not P -> principle of explosion -> all statements are true!" and instead go "I believe P and I believe something that implies not P -> I one of my beliefs are incorrect". It doesn't happen to often, but it'd be nice to have an actual formal refutation for when it does.

Comment author: Vladimir_Nesov 30 December 2011 05:15:32PM *  3 points [-]

The reason is that you don't believe anything with logical conviction, if your "axioms" imply absurdity, you discard the "axioms" as untrustworthy, thus refuting the arguments for their usefulness (that always precede any beliefs, if you look for them). Why do I believe this? My brain tells me so, and its reasoning is potentially suspect.

Comment author: Armok_GoB 30 December 2011 05:57:36PM 0 points [-]

I think I've found the problem: I don't have any good intuitive notion of absurdity. The only clear association I have with it is under "absurdity heuristic" as "a thing to ignore".

That is: It's not self evident to me that what it implies IS absurd. After all, it was implied by a chain of logic I grok and can find no flaw in.

Comment author: Vladimir_Nesov 30 December 2011 06:47:35PM 0 points [-]

I used "absurdity" in the technical math sense.

Comment author: orthonormal 30 December 2011 06:32:12PM *  1 point [-]

To the (mostly social) extent that concepts were useful to your ancestors, one is going to lead to better decisions than the other, and so you should expect to have evolved the latter intuition. (You trust two friends, and then one of them tells you the other is lying- you feel some consternation of the first kind, but then you start trying to figure out which one is trustworthy.)

Comment author: Armok_GoB 30 December 2011 06:51:27PM 0 points [-]

It seems a lot of intuitions all humans are supposed to have were overwritten by noise at some point...

Comment author: endoself 30 December 2011 08:39:27PM 4 points [-]

Do you actually do this - "Oh, not P! I must be the pope." - or do you just notice this - "Not P, so everything's true. Where do I go from here?".

If you want to know why you shouldn't do this it's because you never really learn not P, you just learn evidence against P which you should update with Bayes' rule. If you want to understand this process more intuitively (and you've already read the sequences and are still confused), I would recommend this short tutorial or studying belief propagation in Bayesian networks, for which I don't know a great source for the intuitions behind, but units 3 and 4 of the online Stanford AI class might help.

Comment author: Armok_GoB 31 December 2011 12:03:38AM 2 points [-]

I've actually done that class and gotten really good grades.

Looking at it, it seems I have automatic generation of nodes for new statements, and the creation of a new node does not check for an already existing node for it's inversion.

To complicate matters further, I don't go "I'm the pope" nor "all statements are true.", I go "NOT Bayes theorem, NOT induction, and NOT Occhams razor!"

Comment author: endoself 31 December 2011 04:08:37AM *  1 point [-]

Well, one mathematically right thing to do is to make a new node descending from both other nodes representing E = (P and not P) and then observe not E.

Did you read the first tutorial? Do you find the process of belief-updating on causal nets intuitive, or do you just understand the math? How hard would it be for you to explain why it works in the language of the first tutorial?

Strictly speaking, causal networks only apply to situations where the number of variables does not change, but the intuitions carry over.

Comment author: Armok_GoB 31 December 2011 01:22:35PM 1 point [-]

Thats what I try to do, the problem is I end up observing E to be true. And E leads to an "everything" node.

I'm not sure how well I understand the math, but I feel like I probably do...

Comment author: endoself 31 December 2011 07:32:38PM 1 point [-]

You don't observe E to be true, you infer it to be (very likely) true by propagating from P and from not P. You observe it to be false using the law of noncontradiction.

Parsimony suggests that if you think you understand the math, it's because you understand it. Understanding Bayesianism seems easier than fixing a badly-understood flaw in your brain's implementation of it.

Comment author: Armok_GoB 31 December 2011 07:51:58PM 1 point [-]

How can I get this law of noncontradiction? it seems like an useful thing to have.

Comment author: [deleted] 30 December 2011 05:24:30PM 12 points [-]

Given that utility functions are only defined up to positive linear transforms, what do total utilitarians and average utilitarians actually mean when they're talking about the sum or the average of several utility functions? I mean, taking what they say literally, if Alice's utility function were twice what it actually is, she would behave the exact same way but she would be twice as ‘important’; that cannot possibly be what they mean. What am I missing?

Comment author: endoself 30 December 2011 08:17:08PM *  -1 points [-]

If two possible futures have different numbers of people, those will be subject to different affine transforms, so the utility function as a whole will have been transformed in a non-affine way. See repugnant conclusion for a concrete example.

Comment author: [deleted] 30 December 2011 08:36:01PM 1 point [-]

I think you misunderstood my question. I wasn't asking about what would the difference between summing and averaging be, but how to sum utility functions of different people together in the first place.

Comment author: endoself 30 December 2011 08:50:16PM 2 points [-]

Oh, I completely misunderstood that.

The right answer is that utilitarians aren't summing utility functions, they're just summing some expression about each person. The term hedonic function is used for these when they just care about pleasure or when they aren't worried about being misinterpreted as just caring about pleasure and the term utility function is used when they don't know what a utility function is or when they are willing to misuse it for convenience.

Comment author: steven0461 30 December 2011 08:44:53PM 2 points [-]

See here.

Comment author: orthonormal 30 December 2011 09:49:35PM 6 points [-]

This is actually an open problem in utilitarianism; there were some posts recently looking to bargaining between agents as a solution, but I can't find them at the moment, and in any case that's not a mainstream LW conclusion.

Comment author: jimrandomh 31 December 2011 08:14:54AM 1 point [-]

what do total utilitarians and average utilitarians actually mean when they're talking about the sum or the average of several utility functions?

They don't know. In most cases, they just sort of wave their hands. You can combine utility functions, but "sum" and "average" do not uniquely identify methods for doing so, and no method identified so far has seemed uniquely compelling.

Comment author: RomeoStevens 31 December 2011 08:59:02AM *  0 points [-]

I think you figure out common units to denote utilons in through revealed preference. This only works if both utility functions are coherent.

also last time this came up I linked this to see if anyone knew anything about it: http://www.jstor.org/pss/2630767 and got downvoted. shrug

Comment author: Ezekiel 30 December 2011 08:37:41PM 6 points [-]

If I understand it correctly, the FAI problem is basically about making an AI whose goals match those of humanity. But why does the AI need to have goals at all? Couldn't you just program a question-answering machine and then ask it to solve specific problems?

Comment author: shminux 30 December 2011 08:48:26PM 1 point [-]

Presumably once AGI becomes smarter than humans, it will develop goals of some kind, whether we want it or not. Might as well try to influence them.

Comment author: Ezekiel 30 December 2011 08:51:46PM 3 points [-]

Presumably once AGI becomes smarter than humans, it will develop goals of some kind

Why?

Comment author: Vladimir_Nesov 30 December 2011 09:02:20PM 4 points [-]
Comment author: Kaj_Sotala 31 December 2011 06:31:12AM 6 points [-]

A better wording would probably be that you can't design something with literally no goals and still call it an AI. A system that answers questions and solves specific problems has a goal: to answer questions and solve specific problems. To be useful for that task, its whole architecture has to be crafted with that purpose in mind.

For instance, suppose it was provided questions in the form of written text. This means that its designers will have to build it in such a way that it interprets text in a certain way and tries to discover what we mean by the question. That's just one thing that it could do to the text, though - it could also just discard any text input, or transform each letter to a number and start searching for mathematical patterns in the numbers, or use the text to seed its random-number generator that it was using for some entirely different purpose, and so forth. In order for the AI to do anything useful, it has to have a large number of goals such as "interpret the meaning of this text file I was provided" implicit in its architecture. As the AI grows more powerful, these various goals may manifest themselves in unexpected ways.

Comment author: Vladimir_Nesov 30 December 2011 09:03:05PM 14 points [-]

This idea is called "Oracle AI"; see this post and its dependencies for some reasons why it's probably a bad idea.

Comment author: Ezekiel 30 December 2011 09:23:30PM 3 points [-]

That's exactly what I was looking for. Thank you.

Comment author: Kaj_Sotala 31 December 2011 06:11:08AM 3 points [-]

In addition to the post Vladimir linked, see also this paper.

Comment author: FiftyTwo 31 December 2011 06:15:07PM 1 point [-]

How do I work out what i want and what I should do?

Comment author: Costanza 31 December 2011 09:57:18PM *  1 point [-]

Strictly speaking, this question may be a bit tangential to LessWrong, but this was never supposed to be a exclusive thread.

The answer will depend on a lot of things, mostly specific to you personally. Bodies differ. Even your own single body changes over the course of time. You have certain specific goals and values, and certain constraints.

Maybe what you're really looking for is recommendations for a proper physical fitness forum, which is relatively free of pseudoscience and what they call "woo." I can't advise, myself, but I'm sure that some LessWrongians can.

Comment author: Oscar_Cunningham 01 January 2012 02:34:38PM 7 points [-]

I think you've taken the wrong meaning of the words "work out".

Comment author: Costanza 01 January 2012 05:32:34PM 6 points [-]

[Literally laughing out loud at myself.]

Comment author: FiftyTwo 02 January 2012 12:50:44AM 1 point [-]

Oscar is correct.

A better phrasing might be: Given that I have difficulty inferring my own desires what might be useful methods for me to discover my pre-existing desires or choose long term goals?

[On an unrelated note, reddits r/fitness is an extremely good source for scientifically based 'work out' advice. Which I believe would satisfy Contanza's criteria.]

Comment author: Will_Newsome 02 January 2012 02:17:37AM 0 points [-]

Read things like this at least, and assume there are a lot of things like that that we don't know about. That's another of my stopgap solutions.

Comment author: Will_Newsome 01 January 2012 07:04:34AM *  -2 points [-]

Devise and execute a highly precise ritual wherein you invoke the optimal decision theory. That's my stopgap solution.

Comment author: FiftyTwo 31 December 2011 06:42:07PM 2 points [-]

Is there an easy way to read all the top level posts in order starting from the beginning? There doesn't seem to be a 'first post' link anywhere.

Comment author: Costanza 31 December 2011 09:40:31PM *  3 points [-]

There is a draft of a suggested reading order.

As I understand it, the sequences of LessWrong more or less grew out of prior writings by Eliezer, especially out of his posts at Overcoming Bias, so, there isn't a definitive first post.

Comment author: FiftyTwo 02 January 2012 01:01:07AM 1 point [-]

I've read most of the posts on the suggested order, its more to satisfy my completionist streak and because Eleizer's early posts have an ongoing narrative to them. The brute force solution would simply be to find an early post and click 'previous' until they run out, but I would hope there would be an easier way, as sort by oldest firs tends to be one of the default options in such things.

Comment author: Solvent 02 January 2012 03:49:42AM 2 points [-]

Isn't the first one "The Martial Art of Rationality"?

Comment author: [deleted] 02 January 2012 04:23:46AM *  1 point [-]

You may find one of these helpful. As a heads up, though, you may want to begin with the essays on Eliezer's website (Bayes' Theorem, Technical Explanation, Twelve Virtues, and The Simple Truth) before you start his OB posts.

Comment author: MinibearRex 02 January 2012 04:20:40AM 1 point [-]

Check out this page. Additionally, at the end of each post, there is a link labeled "Article Navigation". Click that, and it will open links to the previous and next post by the author.